pull.go 17 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561
  1. package graph
  2. import (
  3. "fmt"
  4. "io"
  5. "io/ioutil"
  6. "net"
  7. "net/url"
  8. "os"
  9. "strings"
  10. "time"
  11. log "github.com/Sirupsen/logrus"
  12. "github.com/docker/docker/engine"
  13. "github.com/docker/docker/image"
  14. "github.com/docker/docker/pkg/tarsum"
  15. "github.com/docker/docker/registry"
  16. "github.com/docker/docker/utils"
  17. )
  18. func (s *TagStore) CmdPull(job *engine.Job) engine.Status {
  19. if n := len(job.Args); n != 1 && n != 2 {
  20. return job.Errorf("Usage: %s IMAGE [TAG]", job.Name)
  21. }
  22. var (
  23. localName = job.Args[0]
  24. tag string
  25. sf = utils.NewStreamFormatter(job.GetenvBool("json"))
  26. authConfig = &registry.AuthConfig{}
  27. metaHeaders map[string][]string
  28. )
  29. // Resolve the Repository name from fqn to RepositoryInfo
  30. repoInfo, err := registry.ResolveRepositoryInfo(job, localName)
  31. if err != nil {
  32. return job.Error(err)
  33. }
  34. if len(job.Args) > 1 {
  35. tag = job.Args[1]
  36. }
  37. job.GetenvJson("authConfig", authConfig)
  38. job.GetenvJson("metaHeaders", &metaHeaders)
  39. c, err := s.poolAdd("pull", repoInfo.LocalName+":"+tag)
  40. if err != nil {
  41. if c != nil {
  42. // Another pull of the same repository is already taking place; just wait for it to finish
  43. job.Stdout.Write(sf.FormatStatus("", "Repository %s already being pulled by another client. Waiting.", repoInfo.LocalName))
  44. <-c
  45. return engine.StatusOK
  46. }
  47. return job.Error(err)
  48. }
  49. defer s.poolRemove("pull", repoInfo.LocalName+":"+tag)
  50. log.Debugf("pulling image from host %q with remote name %q", repoInfo.Index.Name, repoInfo.RemoteName)
  51. endpoint, err := repoInfo.GetEndpoint()
  52. if err != nil {
  53. return job.Error(err)
  54. }
  55. r, err := registry.NewSession(authConfig, registry.HTTPRequestFactory(metaHeaders), endpoint, true)
  56. if err != nil {
  57. return job.Error(err)
  58. }
  59. logName := repoInfo.LocalName
  60. if tag != "" {
  61. logName += ":" + tag
  62. }
  63. if len(repoInfo.Index.Mirrors) == 0 && (repoInfo.Index.Official || endpoint.Version == registry.APIVersion2) {
  64. j := job.Eng.Job("trust_update_base")
  65. if err = j.Run(); err != nil {
  66. log.Errorf("error updating trust base graph: %s", err)
  67. }
  68. log.Debugf("pulling v2 repository with local name %q", repoInfo.LocalName)
  69. if err := s.pullV2Repository(job.Eng, r, job.Stdout, repoInfo, tag, sf, job.GetenvBool("parallel")); err == nil {
  70. if err = job.Eng.Job("log", "pull", logName, "").Run(); err != nil {
  71. log.Errorf("Error logging event 'pull' for %s: %s", logName, err)
  72. }
  73. return engine.StatusOK
  74. } else if err != registry.ErrDoesNotExist {
  75. log.Errorf("Error from V2 registry: %s", err)
  76. }
  77. log.Debug("image does not exist on v2 registry, falling back to v1")
  78. }
  79. log.Debugf("pulling v1 repository with local name %q", repoInfo.LocalName)
  80. if err = s.pullRepository(r, job.Stdout, repoInfo, tag, sf, job.GetenvBool("parallel")); err != nil {
  81. return job.Error(err)
  82. }
  83. if err = job.Eng.Job("log", "pull", logName, "").Run(); err != nil {
  84. log.Errorf("Error logging event 'pull' for %s: %s", logName, err)
  85. }
  86. return engine.StatusOK
  87. }
  88. func (s *TagStore) pullRepository(r *registry.Session, out io.Writer, repoInfo *registry.RepositoryInfo, askedTag string, sf *utils.StreamFormatter, parallel bool) error {
  89. out.Write(sf.FormatStatus("", "Pulling repository %s", repoInfo.CanonicalName))
  90. repoData, err := r.GetRepositoryData(repoInfo.RemoteName)
  91. if err != nil {
  92. if strings.Contains(err.Error(), "HTTP code: 404") {
  93. return fmt.Errorf("Error: image %s:%s not found", repoInfo.RemoteName, askedTag)
  94. }
  95. // Unexpected HTTP error
  96. return err
  97. }
  98. log.Debugf("Retrieving the tag list")
  99. tagsList, err := r.GetRemoteTags(repoData.Endpoints, repoInfo.RemoteName, repoData.Tokens)
  100. if err != nil {
  101. log.Errorf("unable to get remote tags: %s", err)
  102. return err
  103. }
  104. for tag, id := range tagsList {
  105. repoData.ImgList[id] = &registry.ImgData{
  106. ID: id,
  107. Tag: tag,
  108. Checksum: "",
  109. }
  110. }
  111. log.Debugf("Registering tags")
  112. // If no tag has been specified, pull them all
  113. var imageId string
  114. if askedTag == "" {
  115. for tag, id := range tagsList {
  116. repoData.ImgList[id].Tag = tag
  117. }
  118. } else {
  119. // Otherwise, check that the tag exists and use only that one
  120. id, exists := tagsList[askedTag]
  121. if !exists {
  122. return fmt.Errorf("Tag %s not found in repository %s", askedTag, repoInfo.CanonicalName)
  123. }
  124. imageId = id
  125. repoData.ImgList[id].Tag = askedTag
  126. }
  127. errors := make(chan error)
  128. layers_downloaded := false
  129. for _, image := range repoData.ImgList {
  130. downloadImage := func(img *registry.ImgData) {
  131. if askedTag != "" && img.Tag != askedTag {
  132. log.Debugf("(%s) does not match %s (id: %s), skipping", img.Tag, askedTag, img.ID)
  133. if parallel {
  134. errors <- nil
  135. }
  136. return
  137. }
  138. if img.Tag == "" {
  139. log.Debugf("Image (id: %s) present in this repository but untagged, skipping", img.ID)
  140. if parallel {
  141. errors <- nil
  142. }
  143. return
  144. }
  145. // ensure no two downloads of the same image happen at the same time
  146. if c, err := s.poolAdd("pull", "img:"+img.ID); err != nil {
  147. if c != nil {
  148. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), "Layer already being pulled by another client. Waiting.", nil))
  149. <-c
  150. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), "Download complete", nil))
  151. } else {
  152. log.Debugf("Image (id: %s) pull is already running, skipping: %v", img.ID, err)
  153. }
  154. if parallel {
  155. errors <- nil
  156. }
  157. return
  158. }
  159. defer s.poolRemove("pull", "img:"+img.ID)
  160. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), fmt.Sprintf("Pulling image (%s) from %s", img.Tag, repoInfo.CanonicalName), nil))
  161. success := false
  162. var lastErr, err error
  163. var is_downloaded bool
  164. for _, ep := range repoInfo.Index.Mirrors {
  165. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), fmt.Sprintf("Pulling image (%s) from %s, mirror: %s", img.Tag, repoInfo.CanonicalName, ep), nil))
  166. if is_downloaded, err = s.pullImage(r, out, img.ID, ep, repoData.Tokens, sf); err != nil {
  167. // Don't report errors when pulling from mirrors.
  168. log.Debugf("Error pulling image (%s) from %s, mirror: %s, %s", img.Tag, repoInfo.CanonicalName, ep, err)
  169. continue
  170. }
  171. layers_downloaded = layers_downloaded || is_downloaded
  172. success = true
  173. break
  174. }
  175. if !success {
  176. for _, ep := range repoData.Endpoints {
  177. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), fmt.Sprintf("Pulling image (%s) from %s, endpoint: %s", img.Tag, repoInfo.CanonicalName, ep), nil))
  178. if is_downloaded, err = s.pullImage(r, out, img.ID, ep, repoData.Tokens, sf); err != nil {
  179. // It's not ideal that only the last error is returned, it would be better to concatenate the errors.
  180. // As the error is also given to the output stream the user will see the error.
  181. lastErr = err
  182. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), fmt.Sprintf("Error pulling image (%s) from %s, endpoint: %s, %s", img.Tag, repoInfo.CanonicalName, ep, err), nil))
  183. continue
  184. }
  185. layers_downloaded = layers_downloaded || is_downloaded
  186. success = true
  187. break
  188. }
  189. }
  190. if !success {
  191. err := fmt.Errorf("Error pulling image (%s) from %s, %v", img.Tag, repoInfo.CanonicalName, lastErr)
  192. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), err.Error(), nil))
  193. if parallel {
  194. errors <- err
  195. return
  196. }
  197. }
  198. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), "Download complete", nil))
  199. if parallel {
  200. errors <- nil
  201. }
  202. }
  203. if parallel {
  204. go downloadImage(image)
  205. } else {
  206. downloadImage(image)
  207. }
  208. }
  209. if parallel {
  210. var lastError error
  211. for i := 0; i < len(repoData.ImgList); i++ {
  212. if err := <-errors; err != nil {
  213. lastError = err
  214. }
  215. }
  216. if lastError != nil {
  217. return lastError
  218. }
  219. }
  220. for tag, id := range tagsList {
  221. if askedTag != "" && id != imageId {
  222. continue
  223. }
  224. if err := s.Set(repoInfo.LocalName, tag, id, true); err != nil {
  225. return err
  226. }
  227. }
  228. requestedTag := repoInfo.CanonicalName
  229. if len(askedTag) > 0 {
  230. requestedTag = repoInfo.CanonicalName + ":" + askedTag
  231. }
  232. WriteStatus(requestedTag, out, sf, layers_downloaded)
  233. return nil
  234. }
  235. func (s *TagStore) pullImage(r *registry.Session, out io.Writer, imgID, endpoint string, token []string, sf *utils.StreamFormatter) (bool, error) {
  236. history, err := r.GetRemoteHistory(imgID, endpoint, token)
  237. if err != nil {
  238. return false, err
  239. }
  240. out.Write(sf.FormatProgress(utils.TruncateID(imgID), "Pulling dependent layers", nil))
  241. // FIXME: Try to stream the images?
  242. // FIXME: Launch the getRemoteImage() in goroutines
  243. layers_downloaded := false
  244. for i := len(history) - 1; i >= 0; i-- {
  245. id := history[i]
  246. // ensure no two downloads of the same layer happen at the same time
  247. if c, err := s.poolAdd("pull", "layer:"+id); err != nil {
  248. log.Debugf("Image (id: %s) pull is already running, skipping: %v", id, err)
  249. <-c
  250. }
  251. defer s.poolRemove("pull", "layer:"+id)
  252. if !s.graph.Exists(id) {
  253. out.Write(sf.FormatProgress(utils.TruncateID(id), "Pulling metadata", nil))
  254. var (
  255. imgJSON []byte
  256. imgSize int
  257. err error
  258. img *image.Image
  259. )
  260. retries := 5
  261. for j := 1; j <= retries; j++ {
  262. imgJSON, imgSize, err = r.GetRemoteImageJSON(id, endpoint, token)
  263. if err != nil && j == retries {
  264. out.Write(sf.FormatProgress(utils.TruncateID(id), "Error pulling dependent layers", nil))
  265. return layers_downloaded, err
  266. } else if err != nil {
  267. time.Sleep(time.Duration(j) * 500 * time.Millisecond)
  268. continue
  269. }
  270. img, err = image.NewImgJSON(imgJSON)
  271. layers_downloaded = true
  272. if err != nil && j == retries {
  273. out.Write(sf.FormatProgress(utils.TruncateID(id), "Error pulling dependent layers", nil))
  274. return layers_downloaded, fmt.Errorf("Failed to parse json: %s", err)
  275. } else if err != nil {
  276. time.Sleep(time.Duration(j) * 500 * time.Millisecond)
  277. continue
  278. } else {
  279. break
  280. }
  281. }
  282. for j := 1; j <= retries; j++ {
  283. // Get the layer
  284. status := "Pulling fs layer"
  285. if j > 1 {
  286. status = fmt.Sprintf("Pulling fs layer [retries: %d]", j)
  287. }
  288. out.Write(sf.FormatProgress(utils.TruncateID(id), status, nil))
  289. layer, err := r.GetRemoteImageLayer(img.ID, endpoint, token, int64(imgSize))
  290. if uerr, ok := err.(*url.Error); ok {
  291. err = uerr.Err
  292. }
  293. if terr, ok := err.(net.Error); ok && terr.Timeout() && j < retries {
  294. time.Sleep(time.Duration(j) * 500 * time.Millisecond)
  295. continue
  296. } else if err != nil {
  297. out.Write(sf.FormatProgress(utils.TruncateID(id), "Error pulling dependent layers", nil))
  298. return layers_downloaded, err
  299. }
  300. layers_downloaded = true
  301. defer layer.Close()
  302. err = s.graph.Register(img,
  303. utils.ProgressReader(layer, imgSize, out, sf, false, utils.TruncateID(id), "Downloading"))
  304. if terr, ok := err.(net.Error); ok && terr.Timeout() && j < retries {
  305. time.Sleep(time.Duration(j) * 500 * time.Millisecond)
  306. continue
  307. } else if err != nil {
  308. out.Write(sf.FormatProgress(utils.TruncateID(id), "Error downloading dependent layers", nil))
  309. return layers_downloaded, err
  310. } else {
  311. break
  312. }
  313. }
  314. }
  315. out.Write(sf.FormatProgress(utils.TruncateID(id), "Download complete", nil))
  316. }
  317. return layers_downloaded, nil
  318. }
  319. func WriteStatus(requestedTag string, out io.Writer, sf *utils.StreamFormatter, layers_downloaded bool) {
  320. if layers_downloaded {
  321. out.Write(sf.FormatStatus("", "Status: Downloaded newer image for %s", requestedTag))
  322. } else {
  323. out.Write(sf.FormatStatus("", "Status: Image is up to date for %s", requestedTag))
  324. }
  325. }
  326. // downloadInfo is used to pass information from download to extractor
  327. type downloadInfo struct {
  328. imgJSON []byte
  329. img *image.Image
  330. tmpFile *os.File
  331. length int64
  332. downloaded bool
  333. err chan error
  334. }
  335. func (s *TagStore) pullV2Repository(eng *engine.Engine, r *registry.Session, out io.Writer, repoInfo *registry.RepositoryInfo, tag string, sf *utils.StreamFormatter, parallel bool) error {
  336. endpoint, err := r.V2RegistryEndpoint(repoInfo.Index)
  337. if err != nil {
  338. return fmt.Errorf("error getting registry endpoint: %s", err)
  339. }
  340. auth, err := r.GetV2Authorization(endpoint, repoInfo.RemoteName, true)
  341. if err != nil {
  342. return fmt.Errorf("error getting authorization: %s", err)
  343. }
  344. var layersDownloaded bool
  345. if tag == "" {
  346. log.Debugf("Pulling tag list from V2 registry for %s", repoInfo.CanonicalName)
  347. tags, err := r.GetV2RemoteTags(endpoint, repoInfo.RemoteName, auth)
  348. if err != nil {
  349. return err
  350. }
  351. for _, t := range tags {
  352. if downloaded, err := s.pullV2Tag(eng, r, out, endpoint, repoInfo, t, sf, parallel, auth); err != nil {
  353. return err
  354. } else if downloaded {
  355. layersDownloaded = true
  356. }
  357. }
  358. } else {
  359. if downloaded, err := s.pullV2Tag(eng, r, out, endpoint, repoInfo, tag, sf, parallel, auth); err != nil {
  360. return err
  361. } else if downloaded {
  362. layersDownloaded = true
  363. }
  364. }
  365. requestedTag := repoInfo.CanonicalName
  366. if len(tag) > 0 {
  367. requestedTag = repoInfo.CanonicalName + ":" + tag
  368. }
  369. WriteStatus(requestedTag, out, sf, layersDownloaded)
  370. return nil
  371. }
  372. func (s *TagStore) pullV2Tag(eng *engine.Engine, r *registry.Session, out io.Writer, endpoint *registry.Endpoint, repoInfo *registry.RepositoryInfo, tag string, sf *utils.StreamFormatter, parallel bool, auth *registry.RequestAuthorization) (bool, error) {
  373. log.Debugf("Pulling tag from V2 registry: %q", tag)
  374. manifestBytes, err := r.GetV2ImageManifest(endpoint, repoInfo.RemoteName, tag, auth)
  375. if err != nil {
  376. return false, err
  377. }
  378. manifest, verified, err := s.verifyManifest(eng, manifestBytes)
  379. if err != nil {
  380. return false, fmt.Errorf("error verifying manifest: %s", err)
  381. }
  382. if err := checkValidManifest(manifest); err != nil {
  383. return false, err
  384. }
  385. if verified {
  386. out.Write(sf.FormatStatus(repoInfo.CanonicalName+":"+tag, "The image you are pulling has been verified"))
  387. } else {
  388. out.Write(sf.FormatStatus(tag, "Pulling from %s", repoInfo.CanonicalName))
  389. }
  390. downloads := make([]downloadInfo, len(manifest.FSLayers))
  391. for i := len(manifest.FSLayers) - 1; i >= 0; i-- {
  392. var (
  393. sumStr = manifest.FSLayers[i].BlobSum
  394. imgJSON = []byte(manifest.History[i].V1Compatibility)
  395. )
  396. img, err := image.NewImgJSON(imgJSON)
  397. if err != nil {
  398. return false, fmt.Errorf("failed to parse json: %s", err)
  399. }
  400. downloads[i].img = img
  401. // Check if exists
  402. if s.graph.Exists(img.ID) {
  403. log.Debugf("Image already exists: %s", img.ID)
  404. continue
  405. }
  406. chunks := strings.SplitN(sumStr, ":", 2)
  407. if len(chunks) < 2 {
  408. return false, fmt.Errorf("expected 2 parts in the sumStr, got %#v", chunks)
  409. }
  410. sumType, checksum := chunks[0], chunks[1]
  411. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), "Pulling fs layer", nil))
  412. downloadFunc := func(di *downloadInfo) error {
  413. log.Debugf("pulling blob %q to V1 img %s", sumStr, img.ID)
  414. if c, err := s.poolAdd("pull", "img:"+img.ID); err != nil {
  415. if c != nil {
  416. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), "Layer already being pulled by another client. Waiting.", nil))
  417. <-c
  418. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), "Download complete", nil))
  419. } else {
  420. log.Debugf("Image (id: %s) pull is already running, skipping: %v", img.ID, err)
  421. }
  422. } else {
  423. defer s.poolRemove("pull", "img:"+img.ID)
  424. tmpFile, err := ioutil.TempFile("", "GetV2ImageBlob")
  425. if err != nil {
  426. return err
  427. }
  428. r, l, err := r.GetV2ImageBlobReader(endpoint, repoInfo.RemoteName, sumType, checksum, auth)
  429. if err != nil {
  430. return err
  431. }
  432. defer r.Close()
  433. // Wrap the reader with the appropriate TarSum reader.
  434. tarSumReader, err := tarsum.NewTarSumForLabel(r, true, sumType)
  435. if err != nil {
  436. return fmt.Errorf("unable to wrap image blob reader with TarSum: %s", err)
  437. }
  438. io.Copy(tmpFile, utils.ProgressReader(ioutil.NopCloser(tarSumReader), int(l), out, sf, false, utils.TruncateID(img.ID), "Downloading"))
  439. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), "Verifying Checksum", nil))
  440. if finalChecksum := tarSumReader.Sum(nil); !strings.EqualFold(finalChecksum, sumStr) {
  441. return fmt.Errorf("image verification failed: checksum mismatch - expected %q but got %q", sumStr, finalChecksum)
  442. }
  443. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), "Download complete", nil))
  444. log.Debugf("Downloaded %s to tempfile %s", img.ID, tmpFile.Name())
  445. di.tmpFile = tmpFile
  446. di.length = l
  447. di.downloaded = true
  448. }
  449. di.imgJSON = imgJSON
  450. return nil
  451. }
  452. if parallel {
  453. downloads[i].err = make(chan error)
  454. go func(di *downloadInfo) {
  455. di.err <- downloadFunc(di)
  456. }(&downloads[i])
  457. } else {
  458. err := downloadFunc(&downloads[i])
  459. if err != nil {
  460. return false, err
  461. }
  462. }
  463. }
  464. var layersDownloaded bool
  465. for i := len(downloads) - 1; i >= 0; i-- {
  466. d := &downloads[i]
  467. if d.err != nil {
  468. err := <-d.err
  469. if err != nil {
  470. return false, err
  471. }
  472. }
  473. if d.downloaded {
  474. // if tmpFile is empty assume download and extracted elsewhere
  475. defer os.Remove(d.tmpFile.Name())
  476. defer d.tmpFile.Close()
  477. d.tmpFile.Seek(0, 0)
  478. if d.tmpFile != nil {
  479. err = s.graph.Register(d.img,
  480. utils.ProgressReader(d.tmpFile, int(d.length), out, sf, false, utils.TruncateID(d.img.ID), "Extracting"))
  481. if err != nil {
  482. return false, err
  483. }
  484. // FIXME: Pool release here for parallel tag pull (ensures any downloads block until fully extracted)
  485. }
  486. out.Write(sf.FormatProgress(utils.TruncateID(d.img.ID), "Pull complete", nil))
  487. layersDownloaded = true
  488. } else {
  489. out.Write(sf.FormatProgress(utils.TruncateID(d.img.ID), "Already exists", nil))
  490. }
  491. }
  492. if err = s.Set(repoInfo.LocalName, tag, downloads[0].img.ID, true); err != nil {
  493. return false, err
  494. }
  495. return layersDownloaded, nil
  496. }