package layer // import "github.com/docker/docker/layer" import ( "context" "errors" "fmt" "io" "os" "path/filepath" "sync" "github.com/containerd/log" "github.com/docker/distribution" "github.com/docker/docker/daemon/graphdriver" "github.com/docker/docker/pkg/idtools" "github.com/docker/docker/pkg/plugingetter" "github.com/docker/docker/pkg/stringid" "github.com/moby/locker" "github.com/opencontainers/go-digest" "github.com/vbatts/tar-split/tar/asm" "github.com/vbatts/tar-split/tar/storage" ) // maxLayerDepth represents the maximum number of // layers which can be chained together. 125 was // chosen to account for the 127 max in some // graphdrivers plus the 2 additional layers // used to create a rwlayer. const maxLayerDepth = 125 type layerStore struct { store *fileMetadataStore driver graphdriver.Driver useTarSplit bool layerMap map[ChainID]*roLayer layerL sync.Mutex mounts map[string]*mountedLayer mountL sync.Mutex // protect *RWLayer() methods from operating on the same name/id locker *locker.Locker } // StoreOptions are the options used to create a new Store instance type StoreOptions struct { Root string MetadataStorePathTemplate string GraphDriver string GraphDriverOptions []string IDMapping idtools.IdentityMapping PluginGetter plugingetter.PluginGetter ExperimentalEnabled bool } // NewStoreFromOptions creates a new Store instance func NewStoreFromOptions(options StoreOptions) (Store, error) { driver, err := graphdriver.New(options.GraphDriver, options.PluginGetter, graphdriver.Options{ Root: options.Root, DriverOptions: options.GraphDriverOptions, IDMap: options.IDMapping, ExperimentalEnabled: options.ExperimentalEnabled, }) if err != nil { if options.GraphDriver != "" { return nil, fmt.Errorf("error initializing graphdriver: %v: %s", err, options.GraphDriver) } return nil, fmt.Errorf("error initializing graphdriver: %v", err) } log.G(context.TODO()).Debugf("Initialized graph driver %s", driver) root := fmt.Sprintf(options.MetadataStorePathTemplate, driver) return newStoreFromGraphDriver(root, driver) } // newStoreFromGraphDriver creates a new Store instance using the provided // metadata store and graph driver. The metadata store will be used to restore // the Store. func newStoreFromGraphDriver(root string, driver graphdriver.Driver) (Store, error) { caps := graphdriver.Capabilities{} if capDriver, ok := driver.(graphdriver.CapabilityDriver); ok { caps = capDriver.Capabilities() } ms, err := newFSMetadataStore(root) if err != nil { return nil, err } ls := &layerStore{ store: ms, driver: driver, layerMap: map[ChainID]*roLayer{}, mounts: map[string]*mountedLayer{}, locker: locker.New(), useTarSplit: !caps.ReproducesExactDiffs, } ids, mounts, err := ms.List() if err != nil { return nil, err } for _, id := range ids { l, err := ls.loadLayer(id) if err != nil { log.G(context.TODO()).Debugf("Failed to load layer %s: %s", id, err) continue } if l.parent != nil { l.parent.referenceCount++ } } for _, mount := range mounts { if err := ls.loadMount(mount); err != nil { log.G(context.TODO()).Debugf("Failed to load mount %s: %s", mount, err) } } return ls, nil } func (ls *layerStore) Driver() graphdriver.Driver { return ls.driver } func (ls *layerStore) loadLayer(layer ChainID) (*roLayer, error) { cl, ok := ls.layerMap[layer] if ok { return cl, nil } diff, err := ls.store.GetDiffID(layer) if err != nil { return nil, fmt.Errorf("failed to get diff id for %s: %s", layer, err) } size, err := ls.store.GetSize(layer) if err != nil { return nil, fmt.Errorf("failed to get size for %s: %s", layer, err) } cacheID, err := ls.store.GetCacheID(layer) if err != nil { return nil, fmt.Errorf("failed to get cache id for %s: %s", layer, err) } parent, err := ls.store.GetParent(layer) if err != nil { return nil, fmt.Errorf("failed to get parent for %s: %s", layer, err) } descriptor, err := ls.store.GetDescriptor(layer) if err != nil { return nil, fmt.Errorf("failed to get descriptor for %s: %s", layer, err) } cl = &roLayer{ chainID: layer, diffID: diff, size: size, cacheID: cacheID, layerStore: ls, references: map[Layer]struct{}{}, descriptor: descriptor, } if parent != "" { p, err := ls.loadLayer(parent) if err != nil { return nil, err } cl.parent = p } ls.layerMap[cl.chainID] = cl return cl, nil } func (ls *layerStore) loadMount(mount string) error { ls.mountL.Lock() defer ls.mountL.Unlock() if _, ok := ls.mounts[mount]; ok { return nil } mountID, err := ls.store.GetMountID(mount) if err != nil { return err } initID, err := ls.store.GetInitID(mount) if err != nil { return err } parent, err := ls.store.GetMountParent(mount) if err != nil { return err } ml := &mountedLayer{ name: mount, mountID: mountID, initID: initID, layerStore: ls, references: map[RWLayer]*referencedRWLayer{}, } if parent != "" { p, err := ls.loadLayer(parent) if err != nil { return err } ml.parent = p p.referenceCount++ } ls.mounts[ml.name] = ml return nil } func (ls *layerStore) applyTar(tx *fileMetadataTransaction, ts io.Reader, parent string, layer *roLayer) error { digester := digest.Canonical.Digester() tr := io.TeeReader(ts, digester.Hash()) rdr := tr if ls.useTarSplit { tsw, err := tx.TarSplitWriter(true) if err != nil { return err } metaPacker := storage.NewJSONPacker(tsw) defer tsw.Close() // we're passing nil here for the file putter, because the ApplyDiff will // handle the extraction of the archive rdr, err = asm.NewInputTarStream(tr, metaPacker, nil) if err != nil { return err } } applySize, err := ls.driver.ApplyDiff(layer.cacheID, parent, rdr) // discard trailing data but ensure metadata is picked up to reconstruct stream // unconditionally call io.Copy here before checking err to ensure the resources // allocated by NewInputTarStream above are always released io.Copy(io.Discard, rdr) // ignore error as reader may be closed if err != nil { return err } layer.size = applySize layer.diffID = DiffID(digester.Digest()) log.G(context.TODO()).Debugf("Applied tar %s to %s, size: %d", layer.diffID, layer.cacheID, applySize) return nil } func (ls *layerStore) Register(ts io.Reader, parent ChainID) (Layer, error) { return ls.registerWithDescriptor(ts, parent, distribution.Descriptor{}) } func (ls *layerStore) registerWithDescriptor(ts io.Reader, parent ChainID, descriptor distribution.Descriptor) (Layer, error) { // cErr is used to hold the error which will always trigger // cleanup of creates sources but may not be an error returned // to the caller (already exists). var cErr error var pid string var p *roLayer if string(parent) != "" { ls.layerL.Lock() p = ls.get(parent) ls.layerL.Unlock() if p == nil { return nil, ErrLayerDoesNotExist } pid = p.cacheID // Release parent chain if error defer func() { if cErr != nil { ls.layerL.Lock() ls.releaseLayer(p) ls.layerL.Unlock() } }() if p.depth() >= maxLayerDepth { cErr = ErrMaxDepthExceeded return nil, cErr } } // Create new roLayer layer := &roLayer{ parent: p, cacheID: stringid.GenerateRandomID(), referenceCount: 1, layerStore: ls, references: map[Layer]struct{}{}, descriptor: descriptor, } if cErr = ls.driver.Create(layer.cacheID, pid, nil); cErr != nil { return nil, cErr } tx, cErr := ls.store.StartTransaction() if cErr != nil { return nil, cErr } defer func() { if cErr != nil { log.G(context.TODO()).Debugf("Cleaning up layer %s: %v", layer.cacheID, cErr) if err := ls.driver.Remove(layer.cacheID); err != nil { log.G(context.TODO()).Errorf("Error cleaning up cache layer %s: %v", layer.cacheID, err) } if err := tx.Cancel(); err != nil { log.G(context.TODO()).Errorf("Error canceling metadata transaction %q: %s", tx.String(), err) } } }() if cErr = ls.applyTar(tx, ts, pid, layer); cErr != nil { return nil, cErr } if layer.parent == nil { layer.chainID = ChainID(layer.diffID) } else { layer.chainID = createChainIDFromParent(layer.parent.chainID, layer.diffID) } if cErr = storeLayer(tx, layer); cErr != nil { return nil, cErr } ls.layerL.Lock() defer ls.layerL.Unlock() if existingLayer := ls.get(layer.chainID); existingLayer != nil { // Set error for cleanup, but do not return the error cErr = errors.New("layer already exists") return existingLayer.getReference(), nil } if cErr = tx.Commit(layer.chainID); cErr != nil { return nil, cErr } ls.layerMap[layer.chainID] = layer return layer.getReference(), nil } func (ls *layerStore) get(layer ChainID) *roLayer { l, ok := ls.layerMap[layer] if !ok { return nil } l.referenceCount++ return l } func (ls *layerStore) Get(l ChainID) (Layer, error) { ls.layerL.Lock() defer ls.layerL.Unlock() layer := ls.get(l) if layer == nil { return nil, ErrLayerDoesNotExist } return layer.getReference(), nil } func (ls *layerStore) Map() map[ChainID]Layer { ls.layerL.Lock() defer ls.layerL.Unlock() layers := map[ChainID]Layer{} for k, v := range ls.layerMap { layers[k] = v } return layers } func (ls *layerStore) deleteLayer(layer *roLayer, metadata *Metadata) error { // Rename layer digest folder first so we detect orphan layer(s) // if ls.driver.Remove fails var dir string for { dgst := digest.Digest(layer.chainID) tmpID := fmt.Sprintf("%s-%s-removing", dgst.Encoded(), stringid.GenerateRandomID()) dir = filepath.Join(ls.store.root, string(dgst.Algorithm()), tmpID) err := os.Rename(ls.store.getLayerDirectory(layer.chainID), dir) if os.IsExist(err) { continue } break } err := ls.driver.Remove(layer.cacheID) if err != nil { return err } err = os.RemoveAll(dir) if err != nil { return err } metadata.DiffID = layer.diffID metadata.ChainID = layer.chainID metadata.Size = layer.Size() if err != nil { return err } metadata.DiffSize = layer.size return nil } func (ls *layerStore) releaseLayer(l *roLayer) ([]Metadata, error) { depth := 0 removed := []Metadata{} for { if l.referenceCount == 0 { panic("layer not retained") } l.referenceCount-- if l.referenceCount != 0 { return removed, nil } if len(removed) == 0 && depth > 0 { panic("cannot remove layer with child") } if l.hasReferences() { panic("cannot delete referenced layer") } // Remove layer from layer map first so it is not considered to exist // when if ls.deleteLayer fails. delete(ls.layerMap, l.chainID) var metadata Metadata if err := ls.deleteLayer(l, &metadata); err != nil { return nil, err } removed = append(removed, metadata) if l.parent == nil { return removed, nil } depth++ l = l.parent } } func (ls *layerStore) Release(l Layer) ([]Metadata, error) { ls.layerL.Lock() defer ls.layerL.Unlock() layer, ok := ls.layerMap[l.ChainID()] if !ok { return []Metadata{}, nil } if !layer.hasReference(l) { return nil, ErrLayerNotRetained } layer.deleteReference(l) return ls.releaseLayer(layer) } func (ls *layerStore) CreateRWLayer(name string, parent ChainID, opts *CreateRWLayerOpts) (_ RWLayer, err error) { var ( storageOpt map[string]string initFunc MountInit mountLabel string ) if opts != nil { mountLabel = opts.MountLabel storageOpt = opts.StorageOpt initFunc = opts.InitFunc } ls.locker.Lock(name) defer ls.locker.Unlock(name) ls.mountL.Lock() _, ok := ls.mounts[name] ls.mountL.Unlock() if ok { return nil, ErrMountNameConflict } var pid string var p *roLayer if string(parent) != "" { ls.layerL.Lock() p = ls.get(parent) ls.layerL.Unlock() if p == nil { return nil, ErrLayerDoesNotExist } pid = p.cacheID // Release parent chain if error defer func() { if err != nil { ls.layerL.Lock() ls.releaseLayer(p) ls.layerL.Unlock() } }() } m := &mountedLayer{ name: name, parent: p, mountID: ls.mountID(name), layerStore: ls, references: map[RWLayer]*referencedRWLayer{}, } if initFunc != nil { pid, err = ls.initMount(m.mountID, pid, mountLabel, initFunc, storageOpt) if err != nil { return } m.initID = pid } createOpts := &graphdriver.CreateOpts{ StorageOpt: storageOpt, } if err = ls.driver.CreateReadWrite(m.mountID, pid, createOpts); err != nil { return } if err = ls.saveMount(m); err != nil { return } return m.getReference(), nil } func (ls *layerStore) GetRWLayer(id string) (RWLayer, error) { ls.locker.Lock(id) defer ls.locker.Unlock(id) ls.mountL.Lock() mount := ls.mounts[id] ls.mountL.Unlock() if mount == nil { return nil, ErrMountDoesNotExist } return mount.getReference(), nil } func (ls *layerStore) GetMountID(id string) (string, error) { ls.mountL.Lock() mount := ls.mounts[id] ls.mountL.Unlock() if mount == nil { return "", ErrMountDoesNotExist } log.G(context.TODO()).Debugf("GetMountID id: %s -> mountID: %s", id, mount.mountID) return mount.mountID, nil } func (ls *layerStore) ReleaseRWLayer(l RWLayer) ([]Metadata, error) { name := l.Name() ls.locker.Lock(name) defer ls.locker.Unlock(name) ls.mountL.Lock() m := ls.mounts[name] ls.mountL.Unlock() if m == nil { return []Metadata{}, nil } if err := m.deleteReference(l); err != nil { return nil, err } if m.hasReferences() { return []Metadata{}, nil } if err := ls.driver.Remove(m.mountID); err != nil { log.G(context.TODO()).Errorf("Error removing mounted layer %s: %s", m.name, err) m.retakeReference(l) return nil, err } if m.initID != "" { if err := ls.driver.Remove(m.initID); err != nil { log.G(context.TODO()).Errorf("Error removing init layer %s: %s", m.name, err) m.retakeReference(l) return nil, err } } if err := ls.store.RemoveMount(m.name); err != nil { log.G(context.TODO()).Errorf("Error removing mount metadata: %s: %s", m.name, err) m.retakeReference(l) return nil, err } ls.mountL.Lock() delete(ls.mounts, name) ls.mountL.Unlock() ls.layerL.Lock() defer ls.layerL.Unlock() if m.parent != nil { return ls.releaseLayer(m.parent) } return []Metadata{}, nil } func (ls *layerStore) saveMount(mount *mountedLayer) error { if err := ls.store.SetMountID(mount.name, mount.mountID); err != nil { return err } if mount.initID != "" { if err := ls.store.SetInitID(mount.name, mount.initID); err != nil { return err } } if mount.parent != nil { if err := ls.store.SetMountParent(mount.name, mount.parent.chainID); err != nil { return err } } ls.mountL.Lock() ls.mounts[mount.name] = mount ls.mountL.Unlock() return nil } func (ls *layerStore) initMount(graphID, parent, mountLabel string, initFunc MountInit, storageOpt map[string]string) (string, error) { // Use "-init" to maintain compatibility with graph drivers // which are expecting this layer with this special name. If all // graph drivers can be updated to not rely on knowing about this layer // then the initID should be randomly generated. initID := fmt.Sprintf("%s-init", graphID) createOpts := &graphdriver.CreateOpts{ MountLabel: mountLabel, StorageOpt: storageOpt, } if err := ls.driver.CreateReadWrite(initID, parent, createOpts); err != nil { return "", err } p, err := ls.driver.Get(initID, "") if err != nil { return "", err } if err := initFunc(p); err != nil { ls.driver.Put(initID) return "", err } if err := ls.driver.Put(initID); err != nil { return "", err } return initID, nil } func (ls *layerStore) getTarStream(rl *roLayer) (io.ReadCloser, error) { if !ls.useTarSplit { var parentCacheID string if rl.parent != nil { parentCacheID = rl.parent.cacheID } return ls.driver.Diff(rl.cacheID, parentCacheID) } r, err := ls.store.TarSplitReader(rl.chainID) if err != nil { return nil, err } pr, pw := io.Pipe() go func() { err := ls.assembleTarTo(rl.cacheID, r, nil, pw) if err != nil { pw.CloseWithError(err) } else { pw.Close() } }() return pr, nil } func (ls *layerStore) assembleTarTo(graphID string, metadata io.ReadCloser, size *int64, w io.Writer) error { diffDriver, ok := ls.driver.(graphdriver.DiffGetterDriver) if !ok { diffDriver = &naiveDiffPathDriver{ls.driver} } defer metadata.Close() // get our relative path to the container fileGetCloser, err := diffDriver.DiffGetter(graphID) if err != nil { return err } defer fileGetCloser.Close() metaUnpacker := storage.NewJSONUnpacker(metadata) upackerCounter := &unpackSizeCounter{metaUnpacker, size} log.G(context.TODO()).Debugf("Assembling tar data for %s", graphID) return asm.WriteOutputTarStream(fileGetCloser, upackerCounter, w) } func (ls *layerStore) Cleanup() error { orphanLayers, err := ls.store.getOrphan() if err != nil { log.G(context.TODO()).WithError(err).Error("cannot get orphan layers") } if len(orphanLayers) > 0 { log.G(context.TODO()).Debugf("found %v orphan layers", len(orphanLayers)) } for _, orphan := range orphanLayers { log.G(context.TODO()).WithField("cache-id", orphan.cacheID).Debugf("removing orphan layer, chain ID: %v", orphan.chainID) err = ls.driver.Remove(orphan.cacheID) if err != nil && !os.IsNotExist(err) { log.G(context.TODO()).WithError(err).WithField("cache-id", orphan.cacheID).Error("cannot remove orphan layer") continue } err = ls.store.Remove(orphan.chainID, orphan.cacheID) if err != nil { log.G(context.TODO()).WithError(err).WithField("chain-id", orphan.chainID).Error("cannot remove orphan layer metadata") } } return ls.driver.Cleanup() } func (ls *layerStore) DriverStatus() [][2]string { return ls.driver.Status() } func (ls *layerStore) DriverName() string { return ls.driver.String() } type naiveDiffPathDriver struct { graphdriver.Driver } type fileGetPutter struct { storage.FileGetter driver graphdriver.Driver id string } func (w *fileGetPutter) Close() error { return w.driver.Put(w.id) } func (n *naiveDiffPathDriver) DiffGetter(id string) (graphdriver.FileGetCloser, error) { p, err := n.Driver.Get(id, "") if err != nil { return nil, err } return &fileGetPutter{storage.NewPathFileGetter(p), n.Driver, id}, nil }