-
Notifications
You must be signed in to change notification settings - Fork 3.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Parallel Container Layer Unpacking #8881
Comments
/cc @elfinhe |
/assign @ike-ma |
I'm not quite back from leave, but some context here: The sequential layer unpack is a consequence of snapshot creation requiring a committed parent, and an individual snapshot not being committed until all of its content has been written. For the overlay snapshotter, the backing filesystem does not have a concept of a "committed" lowerdir, and containerd unpacks without an active overlay mount anyway (opting to write whiteout markers explicitly). Because the overlay filesystem does not have a dependency on a committed parent-child relationship, we can implement a further optimization of writing out the actual snapshot in parallel. For some storage devices (such as the PD-SSD device tested above), there are performance gains from concurrent IO over sequential IO and this approach leads to faster overall image pull times. |
We have also discussed in the past of having a "rebase" function on snapshotter. Such a function would be very lightweight in the overlay snapshotter since it would just be updating the parent field. The rebase could possibly be performed on commit so that unpacks could just occur without parents and then commit the snapshots in order with the appropriate parent. Ideally content store would not gain new functionality dealing with any processed content. Snapshotters might have more room for optimization functionality specific to a single snapshotter, as we already have that today on unpack. |
What is the problem you're trying to solve
The current ContainerD fetches layers in parallel, but unpacks them in a single thread layer by layer sequentially.
Describe the solution you'd like
Proposal:
We propose a config below in ContainerD (/etc/containerd/config.toml) to support unpacking images in parallel.
Key Changes: ContainerD Handlers + Content
This option reuses the existing FetchHandler and Content to support pre-decompression during the fetch phase of each layer. The actual decompression happens directly following the fetch process as the graph below. Then the unpack or snapshot handling becomes a light operation by just renaming the pre-decompression folder to the desired overlay path.
Fetch
Unpack and Snapshot processing
Additional context
Who should enable this feature?
Who should NOT enable this feature?
HDD with high seek time for random read and write.
Potential Benefit
If user can improve disk performance, the image pull latency by this proposal can be reduced significantly. For example, Tao was able to achieve 3X faster image pull (120 seconds -> 40 seconds) for pulling a popular container: gcr.io/deeplearning-platform-release/base-cu113:m106 (5.4 GB) with a common setup of a deep learning node (2500 GB PD-SSD, 32 vCPUs).
The text was updated successfully, but these errors were encountered: