// Copyright 2016 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. package gensupport import ( "context" "encoding/base64" "encoding/binary" "errors" "fmt" "io" "net/http" "strings" "sync" "time" "github.com/google/uuid" "google.golang.org/api/internal" ) const ( crc32cPrefix = "crc32c" hashHeaderKey = "X-Goog-Hash" ) // ResumableUpload is used by the generated APIs to provide resumable uploads. // It is not used by developers directly. type ResumableUpload struct { Client *http.Client // URI is the resumable resource destination provided by the server after specifying "&uploadType=resumable". URI string UserAgent string // User-Agent for header of the request // Media is the object being uploaded. Media *MediaBuffer // MediaType defines the media type, e.g. "image/jpeg". MediaType string mu sync.Mutex // guards progress progress int64 // number of bytes uploaded so far // Callback is an optional function that will be periodically called with the cumulative number of bytes uploaded. Callback func(int64) // Retry optionally configures retries for requests made against the upload. Retry *RetryConfig // ChunkRetryDeadline configures the per-chunk deadline after which no further // retries should happen. ChunkRetryDeadline time.Duration // ChunkTransferTimeout configures the per-chunk transfer timeout. If a chunk upload stalls for longer than // this duration, the upload will be retried. ChunkTransferTimeout time.Duration // Track current request invocation ID and attempt count for retry metrics // and idempotency headers. invocationID string attempts int } // Progress returns the number of bytes uploaded at this point. func (rx *ResumableUpload) Progress() int64 { rx.mu.Lock() defer rx.mu.Unlock() return rx.progress } // doUploadRequest performs a single HTTP request to upload data. // off specifies the offset in rx.Media from which data is drawn. // size is the number of bytes in data. // final specifies whether data is the final chunk to be uploaded. func (rx *ResumableUpload) doUploadRequest(ctx context.Context, data io.Reader, off, size int64, final bool) (*http.Response, error) { req, err := http.NewRequest("POST", rx.URI, data) if err != nil { return nil, err } req.ContentLength = size var contentRange string if final { if size == 0 { contentRange = fmt.Sprintf("bytes */%v", off) } else { contentRange = fmt.Sprintf("bytes %v-%v/%v", off, off+size-1, off+size) } } else { contentRange = fmt.Sprintf("bytes %v-%v/*", off, off+size-1) } req.Header.Set("Content-Range", contentRange) req.Header.Set("Content-Type", rx.MediaType) req.Header.Set("User-Agent", rx.UserAgent) // TODO(b/274504690): Consider dropping gccl-invocation-id key since it // duplicates the X-Goog-Gcs-Idempotency-Token header (added in v0.115.0). baseXGoogHeader := "gl-go/" + GoVersion() + " gdcl/" + internal.Version invocationHeader := fmt.Sprintf("gccl-invocation-id/%s gccl-attempt-count/%d", rx.invocationID, rx.attempts) req.Header.Set("X-Goog-Api-Client", strings.Join([]string{baseXGoogHeader, invocationHeader}, " ")) // Set idempotency token header which is used by GCS uploads. req.Header.Set("X-Goog-Gcs-Idempotency-Token", rx.invocationID) // Google's upload endpoint uses status code 308 for a // different purpose than the "308 Permanent Redirect" // since-standardized in RFC 7238. Because of the conflict in // semantics, Google added this new request header which // causes it to not use "308" and instead reply with 200 OK // and sets the upload-specific "X-HTTP-Status-Code-Override: // 308" response header. req.Header.Set("X-GUploader-No-308", "yes") // Server accepts checksum only on final request through header. if final && rx.Media.enableAutoChecksum { req.Header.Set(hashHeaderKey, fmt.Sprintf("%v=%v", crc32cPrefix, encodeUint32(rx.Media.fullObjectChecksum))) } return SendRequest(ctx, rx.Client, req) } func statusResumeIncomplete(resp *http.Response) bool { // This is how the server signals "status resume incomplete" // when X-GUploader-No-308 is set to "yes": return resp != nil && resp.Header.Get("X-Http-Status-Code-Override") == "308" } // reportProgress calls a user-supplied callback to report upload progress. // If old==updated, the callback is not called. func (rx *ResumableUpload) reportProgress(old, updated int64) { if updated-old == 0 { return } rx.mu.Lock() rx.progress = updated rx.mu.Unlock() if rx.Callback != nil { rx.Callback(updated) } } // transferChunk performs a single HTTP request to upload a single chunk. // It uses a goroutine to perform the upload and a timer to enforce ChunkTransferTimeout. func (rx *ResumableUpload) transferChunk(ctx context.Context, chunk io.Reader, off, size int64, done bool) (*http.Response, error) { // If no timeout is specified, perform the request synchronously without a timer. if rx.ChunkTransferTimeout == 0 { res, err := rx.doUploadRequest(ctx, chunk, off, size, done) if err != nil { return res, err } return res, nil } // Start a timer for the ChunkTransferTimeout duration. timer := time.NewTimer(rx.ChunkTransferTimeout) // A struct to hold the result from the goroutine. type uploadResult struct { res *http.Response err error } // A buffered channel to receive the result of the upload. resultCh := make(chan uploadResult, 1) // Create a cancellable context for the upload request. This allows us to // abort the request if the timer fires first. rCtx, cancel := context.WithCancel(ctx) // NOTE: We do NOT use `defer cancel()` here. The context must remain valid // for the caller to read the response body of a successful request. // Cancellation is handled manually on timeout paths. // Starting the chunk upload in parallel. go func() { res, err := rx.doUploadRequest(rCtx, chunk, off, size, done) resultCh <- uploadResult{res: res, err: err} }() // Wait for timer to fire or result channel to have the uploadResult or ctx to be cancelled. select { // Note: Calling cancel() will guarantee that the goroutine finishes, // so these two cases will never block forever on draining the resultCh. case <-ctx.Done(): // Context is cancelled for the overall upload. cancel() // Drain resultCh. <-resultCh return nil, ctx.Err() case <-timer.C: // Chunk Transfer timer fired before resultCh so we return context.DeadlineExceeded. cancel() // Drain resultCh. <-resultCh return nil, context.DeadlineExceeded case result := <-resultCh: // Handle the result from the upload. if result.err != nil { return result.res, result.err } return result.res, nil } } // uploadChunkWithRetries attempts to upload a single chunk, with retries // within ChunkRetryDeadline if ChunkTransferTimeout is non-zero. func (rx *ResumableUpload) uploadChunkWithRetries(ctx context.Context, chunk io.Reader, off, size int64, done bool) (*http.Response, error) { // Configure error retryable criteria. shouldRetry := rx.Retry.errorFunc() // Configure single chunk retry deadline. chunkRetryDeadline := defaultRetryDeadline if rx.ChunkRetryDeadline != 0 { chunkRetryDeadline = rx.ChunkRetryDeadline } // Each chunk gets its own initialized-at-zero backoff and invocation ID. bo := rx.Retry.backoff() quitAfterTimer := time.NewTimer(chunkRetryDeadline) defer quitAfterTimer.Stop() rx.attempts = 1 rx.invocationID = uuid.New().String() var pause time.Duration var resp *http.Response var err error // Retry loop for a single chunk. for { // Wait for the backoff period, unless the context is canceled or the // retry deadline is hit. backoffPauseTimer := time.NewTimer(pause) select { case <-ctx.Done(): backoffPauseTimer.Stop() if err == nil { err = ctx.Err() } return resp, err case <-backoffPauseTimer.C: case <-quitAfterTimer.C: backoffPauseTimer.Stop() return resp, err } backoffPauseTimer.Stop() // Check for context cancellation or timeout once more. If more than one // case in the select statement above was satisfied at the same time, Go // will choose one arbitrarily. // That can cause an operation to go through even if the context was // canceled before or the timeout was reached. select { case <-ctx.Done(): if err == nil { err = ctx.Err() } return resp, err case <-quitAfterTimer.C: return resp, err default: } // We close the response's body here, since we definitely will not // return `resp` now. If we close it before the select case above, a // timer may fire and cause us to return a response with a closed body // (in which case, the caller will not get the error message in the body). if resp != nil && resp.Body != nil { // Read the body to EOF - if the Body is not both read to EOF and closed, // the Client's underlying RoundTripper may not be able to re-use the // persistent TCP connection to the server for a subsequent "keep-alive" request. // See https://pkg.go.dev/net/http#Client.Do io.Copy(io.Discard, resp.Body) resp.Body.Close() } resp, err = rx.transferChunk(ctx, chunk, off, size, done) status := 0 if resp != nil { status = resp.StatusCode } // We sent "X-GUploader-No-308: yes" (see comment elsewhere in // this file), so we don't expect to get a 308. if status == 308 { return nil, errors.New("unexpected 308 response status code") } // Chunk upload should be retried if the ChunkTransferTimeout is non-zero and err is context deadline exceeded // or we encounter a retryable error. if (rx.ChunkTransferTimeout != 0 && errors.Is(err, context.DeadlineExceeded)) || shouldRetry(status, err) { rx.attempts++ pause = bo.Pause() chunk, _, _, _ = rx.Media.Chunk() continue } return resp, err } } // Upload starts the process of a resumable upload with a cancellable context. // It is called from the auto-generated API code and is not visible to the user. // Before sending an HTTP request, Upload calls any registered hook functions, // and calls the returned functions after the request returns (see send.go). // rx is private to the auto-generated API code. // Exactly one of resp or err will be nil. If resp is non-nil, the caller must call resp.Body.Close. // Upload does not parse the response into the error on a non 200 response; // it is the caller's responsibility to call resp.Body.Close. func (rx *ResumableUpload) Upload(ctx context.Context) (*http.Response, error) { for { chunk, off, size, err := rx.Media.Chunk() done := err == io.EOF if !done && err != nil { return nil, err } resp, err := rx.uploadChunkWithRetries(ctx, chunk, off, int64(size), done) // There are a couple of cases where it's possible for err and resp to both // be non-nil. However, we expose a simpler contract to our callers: exactly // one of resp and err will be non-nil. This means that any response body // must be closed here before returning a non-nil error. if err != nil { if resp != nil && resp.Body != nil { resp.Body.Close() } // If there were retries, indicate this in the error message and wrap the final error. if rx.attempts > 1 { return nil, fmt.Errorf("chunk upload failed after %d attempts, final error: %w", rx.attempts, err) } return nil, err } // This case is very unlikely but possible only if rx.ChunkRetryDeadline is // set to a very small value, in which case no requests will be sent before // the deadline. Return an error to avoid causing a panic. if resp == nil { return nil, fmt.Errorf("upload request to %v not sent, choose larger value for ChunkRetryDeadline", rx.URI) } if resp.StatusCode == http.StatusOK { rx.reportProgress(off, off+int64(size)) } if statusResumeIncomplete(resp) { // The upload is not yet complete, but the server has acknowledged this chunk. // We don't have anything to do with the response body. if resp.Body != nil { io.Copy(io.Discard, resp.Body) resp.Body.Close() } rx.Media.Next() continue } return resp, nil } } // Encode a uint32 as Base64 in big-endian byte order. func encodeUint32(u uint32) string { b := make([]byte, 4) binary.BigEndian.PutUint32(b, u) return base64.StdEncoding.EncodeToString(b) }