github.com/grailbio/base@v0.0.11/file/s3file/file_chunk_read.go (about) 1 package s3file 2 3 import ( 4 "context" 5 "fmt" 6 "io" 7 "path/filepath" 8 "sync" 9 "sync/atomic" 10 11 "github.com/aws/aws-sdk-go/aws" 12 "github.com/aws/aws-sdk-go/service/s3" 13 "github.com/aws/aws-sdk-go/service/s3/s3iface" 14 "github.com/grailbio/base/errors" 15 "github.com/grailbio/base/file/internal/s3bufpool" 16 "github.com/grailbio/base/file/s3file/internal/autolog" 17 "github.com/grailbio/base/log" 18 "github.com/grailbio/base/traverse" 19 ) 20 21 type ( 22 // chunkReaderAt is similar to ioctx.ReaderAt except it is not concurrency-safe. 23 // It's currently used to implement S3-recommended read parallelism for large reads, though 24 // clients of s3file still only see the non-parallel io.Reader API. 25 // TODO: Expose concurrency-safe ReaderAt API to clients. 26 chunkReaderAt struct { 27 // name is redundant with (bucket, key). 28 name, bucket, key, versionID string 29 // newRetryPolicy creates retry policies. It must be concurrency- and goroutine-safe. 30 newRetryPolicy func() retryPolicy 31 32 // previousR is a body reader open from a previous ReadAt. It's an optimization for 33 // clients that do many small reads. It may be nil (before first read, after errors, etc.). 34 previousR *posReader 35 // chunks is used locally within ReadAt. It's stored here only to reduce allocations. 36 chunks []readChunk 37 } 38 readChunk struct { 39 // s3Offset is the position of this *chunk* in the coordinates of the S3 object. 40 // That is, dst[0] will eventually contain s3Object[s3Offset]. 41 s3Offset int64 42 // dst contains the chunk's data after read. After read, dstN < len(dst) iff there was an 43 // error or EOF. 44 dst []byte 45 // dstN tracks how much of dst is already filled. 46 dstN int 47 // r is the current reader for this chunk. It may be nil or at the wrong position for 48 // this chunk's state; then we'd need a new reader. 49 r *posReader 50 } 51 52 // posReader wraps the S3 SDK's reader with retries and remembers its offset in the S3 object. 53 posReader struct { 54 rc io.ReadCloser 55 offset int64 56 // ids is set when posReader is opened. 57 ids s3RequestIDs 58 // info is set when posReader is opened, unless there's an error or EOF. 59 info s3Info 60 } 61 ) 62 63 // ReadChunkBytes is the size for individual S3 API read operations, guided by S3 docs: 64 // As a general rule, when you download large objects within a Region from Amazon S3 to 65 // Amazon EC2, we suggest making concurrent requests for byte ranges of an object at the 66 // granularity of 8–16 MB. 67 // https://web.archive.org/web/20220325121400/https://docs.aws.amazon.com/AmazonS3/latest/userguide/optimizing-performance-design-patterns.html 68 func ReadChunkBytes() int { return s3bufpool.BufBytes } 69 70 // ReadAt is not concurrency-safe. 71 // s3Info may be empty if no object metadata is fetched (zero-sized request, error). 72 func (r *chunkReaderAt) ReadAt(ctx context.Context, dst []byte, offset int64) (int, s3Info, error) { 73 if len(dst) == 0 { 74 return 0, s3Info{}, nil 75 } 76 r.chunks = r.chunks[:0] 77 for buf, bufOff := dst, offset; len(buf) > 0; { 78 size := len(buf) 79 if size > s3bufpool.BufBytes { 80 size = s3bufpool.BufBytes 81 } 82 r.chunks = append(r.chunks, readChunk{ 83 s3Offset: bufOff, 84 dst: buf[:size:size], 85 }) 86 bufOff += int64(size) 87 buf = buf[size:] 88 } 89 90 // The first chunk gets to try to use a previously-opened reader (best-effort). 91 // Note: If len(r.chunks) == 1 we're both reusing a saved reader and saving it again. 92 r.chunks[0].r, r.previousR = r.previousR, nil 93 defer func() { 94 r.previousR = r.chunks[len(r.chunks)-1].r 95 }() 96 97 var ( 98 infoMu sync.Mutex 99 info s3Info 100 ) 101 // TODO: traverse (or other common lib) support for exiting on first error to reduce latency. 102 err := traverse.Each(len(r.chunks), func(chunkIdx int) (err error) { 103 chunk := &r.chunks[chunkIdx] 104 policy := r.newRetryPolicy() 105 106 defer func() { 107 if err != nil { 108 err = annotate(err, chunk.r.maybeIDs(), &policy) 109 } 110 }() 111 // Leave the last chunk's reader open for future reuse. 112 if chunkIdx < len(r.chunks)-1 { 113 defer func() { chunk.r.Close(); chunk.r = nil }() 114 } 115 116 metric := metrics.Op("read").Start() 117 defer metric.Done() 118 119 attemptLoop: 120 for attempt := 0; ; attempt++ { 121 switch err { 122 case nil: // Initial attempt. 123 case io.EOF, io.ErrUnexpectedEOF: 124 // In rare cases the S3 SDK returns EOF for chunks that are not actually at EOF. 125 // To work around this, we ignore EOF errors, and keep reading as long as the 126 // object metadata size field says we're not done. See BXDS-2220 for details. 127 // See also: https://github.com/aws/aws-sdk-go/issues/4510 128 default: 129 if !policy.shouldRetry(ctx, err, r.name) { 130 break attemptLoop 131 } 132 } 133 err = nil 134 remainingBuf := chunk.dst[chunk.dstN:] 135 if len(remainingBuf) == 0 { 136 break 137 } 138 139 if attempt > 0 { 140 metric.Retry() 141 } 142 143 rangeStart := chunk.s3Offset + int64(chunk.dstN) 144 switch { 145 case chunk.r != nil && chunk.r.offset == rangeStart: 146 // We're ready to read. 147 case chunk.r != nil: 148 chunk.r.Close() 149 fallthrough 150 default: 151 chunk.r, err = newPosReader(ctx, policy.client(), r.name, r.bucket, r.key, r.versionID, rangeStart) 152 if err == io.EOF { 153 // rangeStart is at or past EOF, so this chunk is done. 154 err = nil 155 break attemptLoop 156 } 157 if err != nil { 158 continue 159 } 160 } 161 162 var size int64 163 infoMu.Lock() 164 if info == (s3Info{}) { 165 info = chunk.r.info 166 } else if info.etag != chunk.r.info.etag { 167 err = eTagChangedError(r.name, info.etag, chunk.r.info.etag) 168 } 169 size = info.size 170 infoMu.Unlock() 171 if err != nil { 172 continue 173 } 174 175 bytesUntilEOF := size - chunk.s3Offset - int64(chunk.dstN) 176 if bytesUntilEOF <= 0 { 177 break 178 } 179 if bytesUntilEOF < int64(len(remainingBuf)) { 180 remainingBuf = remainingBuf[:bytesUntilEOF] 181 } 182 var n int 183 n, err = io.ReadFull(chunk.r, remainingBuf) 184 chunk.dstN += n 185 if err == nil { 186 break 187 } 188 // Discard our reader after an error. This error is often due to throttling 189 // (especially connection reset), so we want to retry with a new HTTP request which 190 // may go to a new host. 191 chunk.r.Close() 192 chunk.r = nil 193 } 194 metric.Bytes(chunk.dstN) 195 return err 196 }) 197 198 var nBytes int 199 for _, chunk := range r.chunks { 200 nBytes += chunk.dstN 201 if chunk.dstN < len(chunk.dst) { 202 if err == nil { 203 err = io.EOF 204 } 205 break 206 } 207 } 208 return nBytes, info, err 209 } 210 211 func eTagChangedError(name, oldETag, newETag string) error { 212 return errors.E(errors.Precondition, fmt.Sprintf( 213 "read %v: ETag changed from %v to %v", name, oldETag, newETag)) 214 } 215 216 func (r *chunkReaderAt) Close() { r.previousR.Close() } 217 218 var ( 219 nOpenPos int32 220 nOpenPosOnce sync.Once 221 ) 222 223 func newPosReader( 224 ctx context.Context, 225 client s3iface.S3API, 226 name, bucket, key, versionID string, 227 offset int64, 228 ) (*posReader, error) { 229 nOpenPosOnce.Do(func() { 230 autolog.Register(func() { 231 log.Printf("s3file open posReader: %d", atomic.LoadInt32(&nOpenPos)) 232 }) 233 }) 234 r := posReader{offset: offset} 235 input := s3.GetObjectInput{ 236 Bucket: aws.String(bucket), 237 Key: aws.String(key), 238 Range: aws.String(fmt.Sprintf("bytes=%d-", r.offset)), 239 } 240 if versionID != "" { 241 input.VersionId = aws.String(versionID) 242 } 243 output, err := client.GetObjectWithContext(ctx, &input, r.ids.captureOption()) 244 if err != nil { 245 if output.Body != nil { 246 if errClose := output.Body.Close(); errClose != nil { 247 log.Printf("s3file.newPosReader: ignoring body close error: %v", err) 248 } 249 } 250 if awsErr, ok := getAWSError(err); ok && awsErr.Code() == "InvalidRange" { 251 // Since we're reading many chunks in parallel, some can be past the end of 252 // the object, resulting in range errors. Treat these as EOF. 253 err = io.EOF 254 } 255 return nil, err 256 } 257 _ = atomic.AddInt32(&nOpenPos, 1) 258 if output.ContentLength == nil || output.ETag == nil || output.LastModified == nil { 259 return nil, errors.E("s3file.newPosReader: object missing metadata (ContentLength, ETag, LastModified)") 260 } 261 if *output.ContentLength < 0 { 262 // We do not expect AWS to return negative ContentLength, but we are 263 // defensive, as things may otherwise break very confusingly for 264 // callers. 265 return nil, io.EOF 266 } 267 r.info = s3Info{ 268 name: filepath.Base(name), 269 size: offset + *output.ContentLength, 270 modTime: *output.LastModified, 271 etag: *output.ETag, 272 } 273 r.rc = output.Body 274 return &r, nil 275 } 276 277 // Read usually delegates to the underlying reader, except: (&posReader{}).Read is valid and 278 // always at EOF; nil.Read panics. 279 func (p *posReader) Read(dst []byte) (int, error) { 280 if p.rc == nil { 281 return 0, io.EOF 282 } 283 n, err := p.rc.Read(dst) 284 p.offset += int64(n) 285 return n, err 286 } 287 288 // Close usually delegates to the underlying reader, except: (&posReader{}).Close 289 // and nil.Close do nothing. 290 func (p *posReader) Close() { 291 if p == nil || p.rc == nil { 292 return 293 } 294 _ = atomic.AddInt32(&nOpenPos, -1) 295 if err := p.rc.Close(); err != nil { 296 // Note: Since the caller is already done reading from p.rc, we don't expect this error to 297 // indicate a problem with the correctness of past Reads, instead signaling some resource 298 // leakage (network connection, buffers, etc.). We can't retry the resource release: 299 // * io.Closer does not define behavior for multiple Close calls and 300 // s3.GetObjectOutput.Body doesn't say anything implementation-specific. 301 // * Body may be a net/http.Response.Body [1] but the standard library doesn't say 302 // anything about multiple Close either (and even if it did, we shouldn't rely on the 303 // AWS SDK's implementation details in all cases or in the future). 304 // Without a retry opportunity, it seems like callers could either ignore the potential 305 // leak, or exit the OS process. We assume, for now, that callers won't want to do the 306 // latter, so we hide the error. (This could eventually lead to OS process exit due to 307 // resource exhaustion, so arguably this hiding doesn't add much harm, though of course it 308 // may be confusing.) We could consider changing this in the future, especially if we notice 309 // such resource leaks in real programs. 310 // 311 // [1] https://github.com/aws/aws-sdk-go/blob/e842504a6323096540dc3defdc7cb357d8749893/private/protocol/rest/unmarshal.go#L89-L90 312 log.Printf("s3file.posReader.Close: ignoring body close error: %v", err) 313 } 314 } 315 316 // maybeIDs returns ids if available, otherwise zero. p == nil is allowed. 317 func (p *posReader) maybeIDs() s3RequestIDs { 318 if p == nil { 319 return s3RequestIDs{} 320 } 321 return p.ids 322 }