github.com/Heebron/moby@v0.0.0-20221111184709-6eab4f55faf7/distribution/xfer/transfer.go (about) 1 package xfer // import "github.com/docker/docker/distribution/xfer" 2 3 import ( 4 "context" 5 "runtime" 6 "sync" 7 8 "github.com/docker/docker/pkg/progress" 9 "github.com/pkg/errors" 10 ) 11 12 // DoNotRetry is an error wrapper indicating that the error cannot be resolved 13 // with a retry. 14 type DoNotRetry struct { 15 Err error 16 } 17 18 // Error returns the stringified representation of the encapsulated error. 19 func (e DoNotRetry) Error() string { 20 return e.Err.Error() 21 } 22 23 // IsDoNotRetryError returns true if the error is caused by DoNotRetry error, 24 // and the transfer should not be retried. 25 func IsDoNotRetryError(err error) bool { 26 var dnr DoNotRetry 27 return errors.As(err, &dnr) 28 } 29 30 // watcher is returned by Watch and can be passed to Release to stop watching. 31 type watcher struct { 32 // signalChan is used to signal to the watcher goroutine that 33 // new progress information is available, or that the transfer 34 // has finished. 35 signalChan chan struct{} 36 // releaseChan signals to the watcher goroutine that the watcher 37 // should be detached. 38 releaseChan chan struct{} 39 // running remains open as long as the watcher is watching the 40 // transfer. It gets closed if the transfer finishes or the 41 // watcher is detached. 42 running chan struct{} 43 } 44 45 // transfer represents an in-progress transfer. 46 type transfer interface { 47 watch(progressOutput progress.Output) *watcher 48 release(*watcher) 49 context() context.Context 50 close() 51 done() <-chan struct{} 52 released() <-chan struct{} 53 broadcast(mainProgressChan <-chan progress.Progress) 54 } 55 56 type xfer struct { 57 mu sync.Mutex 58 59 ctx context.Context 60 cancel context.CancelFunc 61 62 // watchers keeps track of the goroutines monitoring progress output, 63 // indexed by the channels that release them. 64 watchers map[chan struct{}]*watcher 65 66 // lastProgress is the most recently received progress event. 67 lastProgress progress.Progress 68 // hasLastProgress is true when lastProgress has been set. 69 hasLastProgress bool 70 71 // running remains open as long as the transfer is in progress. 72 running chan struct{} 73 // releasedChan stays open until all watchers release the transfer and 74 // the transfer is no longer tracked by the transferManager. 75 releasedChan chan struct{} 76 77 // broadcastDone is true if the main progress channel has closed. 78 broadcastDone bool 79 // closed is true if Close has been called 80 closed bool 81 // broadcastSyncChan allows watchers to "ping" the broadcasting 82 // goroutine to wait for it for deplete its input channel. This ensures 83 // a detaching watcher won't miss an event that was sent before it 84 // started detaching. 85 broadcastSyncChan chan struct{} 86 } 87 88 // newTransfer creates a new transfer. 89 func newTransfer() transfer { 90 t := &xfer{ 91 watchers: make(map[chan struct{}]*watcher), 92 running: make(chan struct{}), 93 releasedChan: make(chan struct{}), 94 broadcastSyncChan: make(chan struct{}), 95 } 96 97 // This uses context.Background instead of a caller-supplied context 98 // so that a transfer won't be cancelled automatically if the client 99 // which requested it is ^C'd (there could be other viewers). 100 t.ctx, t.cancel = context.WithCancel(context.Background()) 101 102 return t 103 } 104 105 // Broadcast copies the progress and error output to all viewers. 106 func (t *xfer) broadcast(mainProgressChan <-chan progress.Progress) { 107 for { 108 var ( 109 p progress.Progress 110 ok bool 111 ) 112 select { 113 case p, ok = <-mainProgressChan: 114 default: 115 // We've depleted the channel, so now we can handle 116 // reads on broadcastSyncChan to let detaching watchers 117 // know we're caught up. 118 select { 119 case <-t.broadcastSyncChan: 120 continue 121 case p, ok = <-mainProgressChan: 122 } 123 } 124 125 t.mu.Lock() 126 if ok { 127 t.lastProgress = p 128 t.hasLastProgress = true 129 for _, w := range t.watchers { 130 select { 131 case w.signalChan <- struct{}{}: 132 default: 133 } 134 } 135 } else { 136 t.broadcastDone = true 137 } 138 t.mu.Unlock() 139 if !ok { 140 close(t.running) 141 return 142 } 143 } 144 } 145 146 // Watch adds a watcher to the transfer. The supplied channel gets progress 147 // updates and is closed when the transfer finishes. 148 func (t *xfer) watch(progressOutput progress.Output) *watcher { 149 t.mu.Lock() 150 defer t.mu.Unlock() 151 152 w := &watcher{ 153 releaseChan: make(chan struct{}), 154 signalChan: make(chan struct{}), 155 running: make(chan struct{}), 156 } 157 158 t.watchers[w.releaseChan] = w 159 160 if t.broadcastDone { 161 close(w.running) 162 return w 163 } 164 165 go func() { 166 defer func() { 167 close(w.running) 168 }() 169 var ( 170 done bool 171 lastWritten progress.Progress 172 hasLastWritten bool 173 ) 174 for { 175 t.mu.Lock() 176 hasLastProgress := t.hasLastProgress 177 lastProgress := t.lastProgress 178 t.mu.Unlock() 179 180 // Make sure we don't write the last progress item 181 // twice. 182 if hasLastProgress && (!done || !hasLastWritten || lastProgress != lastWritten) { 183 progressOutput.WriteProgress(lastProgress) 184 lastWritten = lastProgress 185 hasLastWritten = true 186 } 187 188 if done { 189 return 190 } 191 192 select { 193 case <-w.signalChan: 194 case <-w.releaseChan: 195 done = true 196 // Since the watcher is going to detach, make 197 // sure the broadcaster is caught up so we 198 // don't miss anything. 199 select { 200 case t.broadcastSyncChan <- struct{}{}: 201 case <-t.running: 202 } 203 case <-t.running: 204 done = true 205 } 206 } 207 }() 208 209 return w 210 } 211 212 // Release is the inverse of Watch; indicating that the watcher no longer wants 213 // to be notified about the progress of the transfer. All calls to Watch must 214 // be paired with later calls to Release so that the lifecycle of the transfer 215 // is properly managed. 216 func (t *xfer) release(watcher *watcher) { 217 t.mu.Lock() 218 delete(t.watchers, watcher.releaseChan) 219 220 if len(t.watchers) == 0 { 221 if t.closed { 222 // released may have been closed already if all 223 // watchers were released, then another one was added 224 // while waiting for a previous watcher goroutine to 225 // finish. 226 select { 227 case <-t.releasedChan: 228 default: 229 close(t.releasedChan) 230 } 231 } else { 232 t.cancel() 233 } 234 } 235 t.mu.Unlock() 236 237 close(watcher.releaseChan) 238 // Block until the watcher goroutine completes 239 <-watcher.running 240 } 241 242 // Done returns a channel which is closed if the transfer completes or is 243 // cancelled. Note that having 0 watchers causes a transfer to be cancelled. 244 func (t *xfer) done() <-chan struct{} { 245 // Note that this doesn't return t.ctx.Done() because that channel will 246 // be closed the moment Cancel is called, and we need to return a 247 // channel that blocks until a cancellation is actually acknowledged by 248 // the transfer function. 249 return t.running 250 } 251 252 // Released returns a channel which is closed once all watchers release the 253 // transfer AND the transfer is no longer tracked by the transferManager. 254 func (t *xfer) released() <-chan struct{} { 255 return t.releasedChan 256 } 257 258 // Context returns the context associated with the transfer. 259 func (t *xfer) context() context.Context { 260 return t.ctx 261 } 262 263 // Close is called by the transferManager when the transfer is no longer 264 // being tracked. 265 func (t *xfer) close() { 266 t.mu.Lock() 267 t.closed = true 268 if len(t.watchers) == 0 { 269 close(t.releasedChan) 270 } 271 t.mu.Unlock() 272 } 273 274 // doFunc is a function called by the transferManager to actually perform 275 // a transfer. It should be non-blocking. It should wait until the start channel 276 // is closed before transferring any data. If the function closes inactive, that 277 // signals to the transferManager that the job is no longer actively moving 278 // data - for example, it may be waiting for a dependent transfer to finish. 279 // This prevents it from taking up a slot. 280 type doFunc func(progressChan chan<- progress.Progress, start <-chan struct{}, inactive chan<- struct{}) transfer 281 282 // transferManager is used by LayerDownloadManager and LayerUploadManager to 283 // schedule and deduplicate transfers. It is up to the transferManager 284 // to make the scheduling and concurrency decisions. 285 type transferManager struct { 286 mu sync.Mutex 287 288 concurrencyLimit int 289 activeTransfers int 290 transfers map[string]transfer 291 waitingTransfers []chan struct{} 292 } 293 294 // newTransferManager returns a new transferManager. 295 func newTransferManager(concurrencyLimit int) *transferManager { 296 return &transferManager{ 297 concurrencyLimit: concurrencyLimit, 298 transfers: make(map[string]transfer), 299 } 300 } 301 302 // setConcurrency sets the concurrencyLimit 303 func (tm *transferManager) setConcurrency(concurrency int) { 304 tm.mu.Lock() 305 tm.concurrencyLimit = concurrency 306 tm.mu.Unlock() 307 } 308 309 // transfer checks if a transfer matching the given key is in progress. If not, 310 // it starts one by calling xferFunc. The caller supplies a channel which 311 // receives progress output from the transfer. 312 func (tm *transferManager) transfer(key string, xferFunc doFunc, progressOutput progress.Output) (transfer, *watcher) { 313 tm.mu.Lock() 314 defer tm.mu.Unlock() 315 316 for { 317 xfer, present := tm.transfers[key] 318 if !present { 319 break 320 } 321 // transfer is already in progress. 322 watcher := xfer.watch(progressOutput) 323 324 select { 325 case <-xfer.context().Done(): 326 // We don't want to watch a transfer that has been cancelled. 327 // Wait for it to be removed from the map and try again. 328 xfer.release(watcher) 329 tm.mu.Unlock() 330 // The goroutine that removes this transfer from the 331 // map is also waiting for xfer.Done(), so yield to it. 332 // This could be avoided by adding a Closed method 333 // to transfer to allow explicitly waiting for it to be 334 // removed the map, but forcing a scheduling round in 335 // this very rare case seems better than bloating the 336 // interface definition. 337 runtime.Gosched() 338 <-xfer.done() 339 tm.mu.Lock() 340 default: 341 return xfer, watcher 342 } 343 } 344 345 start := make(chan struct{}) 346 inactive := make(chan struct{}) 347 348 if tm.concurrencyLimit == 0 || tm.activeTransfers < tm.concurrencyLimit { 349 close(start) 350 tm.activeTransfers++ 351 } else { 352 tm.waitingTransfers = append(tm.waitingTransfers, start) 353 } 354 355 mainProgressChan := make(chan progress.Progress) 356 xfer := xferFunc(mainProgressChan, start, inactive) 357 watcher := xfer.watch(progressOutput) 358 go xfer.broadcast(mainProgressChan) 359 tm.transfers[key] = xfer 360 361 // When the transfer is finished, remove from the map. 362 go func() { 363 for { 364 select { 365 case <-inactive: 366 tm.mu.Lock() 367 tm.inactivate(start) 368 tm.mu.Unlock() 369 inactive = nil 370 case <-xfer.done(): 371 tm.mu.Lock() 372 if inactive != nil { 373 tm.inactivate(start) 374 } 375 delete(tm.transfers, key) 376 tm.mu.Unlock() 377 xfer.close() 378 return 379 } 380 } 381 }() 382 383 return xfer, watcher 384 } 385 386 func (tm *transferManager) inactivate(start chan struct{}) { 387 // If the transfer was started, remove it from the activeTransfers 388 // count. 389 select { 390 case <-start: 391 // Start next transfer if any are waiting 392 if len(tm.waitingTransfers) != 0 { 393 close(tm.waitingTransfers[0]) 394 tm.waitingTransfers = tm.waitingTransfers[1:] 395 } else { 396 tm.activeTransfers-- 397 } 398 default: 399 } 400 }