pull.go 17 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588
  1. package graph
  2. import (
  3. "bytes"
  4. "encoding/json"
  5. "fmt"
  6. "io"
  7. "io/ioutil"
  8. "net"
  9. "net/url"
  10. "os"
  11. "strings"
  12. "time"
  13. "github.com/docker/docker/engine"
  14. "github.com/docker/docker/image"
  15. "github.com/docker/docker/pkg/log"
  16. "github.com/docker/docker/registry"
  17. "github.com/docker/docker/utils"
  18. "github.com/docker/libtrust"
  19. )
  20. func (s *TagStore) verifyManifest(eng *engine.Engine, manifestBytes []byte) (*registry.ManifestData, bool, error) {
  21. sig, err := libtrust.ParsePrettySignature(manifestBytes, "signatures")
  22. if err != nil {
  23. return nil, false, fmt.Errorf("error parsing payload: %s", err)
  24. }
  25. keys, err := sig.Verify()
  26. if err != nil {
  27. return nil, false, fmt.Errorf("error verifying payload: %s", err)
  28. }
  29. payload, err := sig.Payload()
  30. if err != nil {
  31. return nil, false, fmt.Errorf("error retrieving payload: %s", err)
  32. }
  33. var manifest registry.ManifestData
  34. if err := json.Unmarshal(payload, &manifest); err != nil {
  35. return nil, false, fmt.Errorf("error unmarshalling manifest: %s", err)
  36. }
  37. if manifest.SchemaVersion != 1 {
  38. return nil, false, fmt.Errorf("unsupported schema version: %d", manifest.SchemaVersion)
  39. }
  40. var verified bool
  41. for _, key := range keys {
  42. job := eng.Job("trust_key_check")
  43. b, err := key.MarshalJSON()
  44. if err != nil {
  45. return nil, false, fmt.Errorf("error marshalling public key: %s", err)
  46. }
  47. namespace := manifest.Name
  48. if namespace[0] != '/' {
  49. namespace = "/" + namespace
  50. }
  51. stdoutBuffer := bytes.NewBuffer(nil)
  52. job.Args = append(job.Args, namespace)
  53. job.Setenv("PublicKey", string(b))
  54. // Check key has read/write permission (0x03)
  55. job.SetenvInt("Permission", 0x03)
  56. job.Stdout.Add(stdoutBuffer)
  57. if err = job.Run(); err != nil {
  58. return nil, false, fmt.Errorf("error running key check: %s", err)
  59. }
  60. result := engine.Tail(stdoutBuffer, 1)
  61. log.Debugf("Key check result: %q", result)
  62. if result == "verified" {
  63. verified = true
  64. }
  65. }
  66. return &manifest, verified, nil
  67. }
  68. func (s *TagStore) CmdPull(job *engine.Job) engine.Status {
  69. if n := len(job.Args); n != 1 && n != 2 {
  70. return job.Errorf("Usage: %s IMAGE [TAG]", job.Name)
  71. }
  72. var (
  73. localName = job.Args[0]
  74. tag string
  75. sf = utils.NewStreamFormatter(job.GetenvBool("json"))
  76. authConfig = &registry.AuthConfig{}
  77. metaHeaders map[string][]string
  78. mirrors []string
  79. )
  80. if len(job.Args) > 1 {
  81. tag = job.Args[1]
  82. }
  83. job.GetenvJson("authConfig", authConfig)
  84. job.GetenvJson("metaHeaders", &metaHeaders)
  85. c, err := s.poolAdd("pull", localName+":"+tag)
  86. if err != nil {
  87. if c != nil {
  88. // Another pull of the same repository is already taking place; just wait for it to finish
  89. job.Stdout.Write(sf.FormatStatus("", "Repository %s already being pulled by another client. Waiting.", localName))
  90. <-c
  91. return engine.StatusOK
  92. }
  93. return job.Error(err)
  94. }
  95. defer s.poolRemove("pull", localName+":"+tag)
  96. // Resolve the Repository name from fqn to endpoint + name
  97. hostname, remoteName, err := registry.ResolveRepositoryName(localName)
  98. if err != nil {
  99. return job.Error(err)
  100. }
  101. endpoint, err := registry.NewEndpoint(hostname)
  102. if err != nil {
  103. return job.Error(err)
  104. }
  105. r, err := registry.NewSession(authConfig, registry.HTTPRequestFactory(metaHeaders), endpoint, true)
  106. if err != nil {
  107. return job.Error(err)
  108. }
  109. var isOfficial bool
  110. if endpoint.VersionString(1) == registry.IndexServerAddress() {
  111. // If pull "index.docker.io/foo/bar", it's stored locally under "foo/bar"
  112. localName = remoteName
  113. isOfficial = isOfficialName(remoteName)
  114. if isOfficial && strings.IndexRune(remoteName, '/') == -1 {
  115. remoteName = "library/" + remoteName
  116. }
  117. // Use provided mirrors, if any
  118. mirrors = s.mirrors
  119. }
  120. if err = s.pullRepository(r, job.Stdout, localName, remoteName, tag, sf, job.GetenvBool("parallel"), mirrors); err != nil {
  121. return job.Error(err)
  122. }
  123. return engine.StatusOK
  124. }
  125. func (s *TagStore) pullRepository(r *registry.Session, out io.Writer, localName, remoteName, askedTag string, sf *utils.StreamFormatter, parallel bool, mirrors []string) error {
  126. out.Write(sf.FormatStatus("", "Pulling repository %s", localName))
  127. repoData, err := r.GetRepositoryData(remoteName)
  128. if err != nil {
  129. if strings.Contains(err.Error(), "HTTP code: 404") {
  130. return fmt.Errorf("Error: image %s not found", remoteName)
  131. }
  132. // Unexpected HTTP error
  133. return err
  134. }
  135. log.Debugf("Retrieving the tag list")
  136. tagsList, err := r.GetRemoteTags(repoData.Endpoints, remoteName, repoData.Tokens)
  137. if err != nil {
  138. log.Errorf("%v", err)
  139. return err
  140. }
  141. for tag, id := range tagsList {
  142. repoData.ImgList[id] = &registry.ImgData{
  143. ID: id,
  144. Tag: tag,
  145. Checksum: "",
  146. }
  147. }
  148. log.Debugf("Registering tags")
  149. // If no tag has been specified, pull them all
  150. var imageId string
  151. if askedTag == "" {
  152. for tag, id := range tagsList {
  153. repoData.ImgList[id].Tag = tag
  154. }
  155. } else {
  156. // Otherwise, check that the tag exists and use only that one
  157. id, exists := tagsList[askedTag]
  158. if !exists {
  159. return fmt.Errorf("Tag %s not found in repository %s", askedTag, localName)
  160. }
  161. imageId = id
  162. repoData.ImgList[id].Tag = askedTag
  163. }
  164. errors := make(chan error)
  165. layers_downloaded := false
  166. for _, image := range repoData.ImgList {
  167. downloadImage := func(img *registry.ImgData) {
  168. if askedTag != "" && img.Tag != askedTag {
  169. log.Debugf("(%s) does not match %s (id: %s), skipping", img.Tag, askedTag, img.ID)
  170. if parallel {
  171. errors <- nil
  172. }
  173. return
  174. }
  175. if img.Tag == "" {
  176. log.Debugf("Image (id: %s) present in this repository but untagged, skipping", img.ID)
  177. if parallel {
  178. errors <- nil
  179. }
  180. return
  181. }
  182. // ensure no two downloads of the same image happen at the same time
  183. if c, err := s.poolAdd("pull", "img:"+img.ID); err != nil {
  184. if c != nil {
  185. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), "Layer already being pulled by another client. Waiting.", nil))
  186. <-c
  187. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), "Download complete", nil))
  188. } else {
  189. log.Debugf("Image (id: %s) pull is already running, skipping: %v", img.ID, err)
  190. }
  191. if parallel {
  192. errors <- nil
  193. }
  194. return
  195. }
  196. defer s.poolRemove("pull", "img:"+img.ID)
  197. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), fmt.Sprintf("Pulling image (%s) from %s", img.Tag, localName), nil))
  198. success := false
  199. var lastErr, err error
  200. var is_downloaded bool
  201. if mirrors != nil {
  202. for _, ep := range mirrors {
  203. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), fmt.Sprintf("Pulling image (%s) from %s, mirror: %s", img.Tag, localName, ep), nil))
  204. if is_downloaded, err = s.pullImage(r, out, img.ID, ep, repoData.Tokens, sf); err != nil {
  205. // Don't report errors when pulling from mirrors.
  206. log.Debugf("Error pulling image (%s) from %s, mirror: %s, %s", img.Tag, localName, ep, err)
  207. continue
  208. }
  209. layers_downloaded = layers_downloaded || is_downloaded
  210. success = true
  211. break
  212. }
  213. }
  214. if !success {
  215. for _, ep := range repoData.Endpoints {
  216. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), fmt.Sprintf("Pulling image (%s) from %s, endpoint: %s", img.Tag, localName, ep), nil))
  217. if is_downloaded, err = s.pullImage(r, out, img.ID, ep, repoData.Tokens, sf); err != nil {
  218. // It's not ideal that only the last error is returned, it would be better to concatenate the errors.
  219. // As the error is also given to the output stream the user will see the error.
  220. lastErr = err
  221. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), fmt.Sprintf("Error pulling image (%s) from %s, endpoint: %s, %s", img.Tag, localName, ep, err), nil))
  222. continue
  223. }
  224. layers_downloaded = layers_downloaded || is_downloaded
  225. success = true
  226. break
  227. }
  228. }
  229. if !success {
  230. err := fmt.Errorf("Error pulling image (%s) from %s, %v", img.Tag, localName, lastErr)
  231. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), err.Error(), nil))
  232. if parallel {
  233. errors <- err
  234. return
  235. }
  236. }
  237. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), "Download complete", nil))
  238. if parallel {
  239. errors <- nil
  240. }
  241. }
  242. if parallel {
  243. go downloadImage(image)
  244. } else {
  245. downloadImage(image)
  246. }
  247. }
  248. if parallel {
  249. var lastError error
  250. for i := 0; i < len(repoData.ImgList); i++ {
  251. if err := <-errors; err != nil {
  252. lastError = err
  253. }
  254. }
  255. if lastError != nil {
  256. return lastError
  257. }
  258. }
  259. for tag, id := range tagsList {
  260. if askedTag != "" && id != imageId {
  261. continue
  262. }
  263. if err := s.Set(localName, tag, id, true); err != nil {
  264. return err
  265. }
  266. }
  267. requestedTag := localName
  268. if len(askedTag) > 0 {
  269. requestedTag = localName + ":" + askedTag
  270. }
  271. WriteStatus(requestedTag, out, sf, layers_downloaded)
  272. return nil
  273. }
  274. func (s *TagStore) pullImage(r *registry.Session, out io.Writer, imgID, endpoint string, token []string, sf *utils.StreamFormatter) (bool, error) {
  275. history, err := r.GetRemoteHistory(imgID, endpoint, token)
  276. if err != nil {
  277. return false, err
  278. }
  279. out.Write(sf.FormatProgress(utils.TruncateID(imgID), "Pulling dependent layers", nil))
  280. // FIXME: Try to stream the images?
  281. // FIXME: Launch the getRemoteImage() in goroutines
  282. layers_downloaded := false
  283. for i := len(history) - 1; i >= 0; i-- {
  284. id := history[i]
  285. // ensure no two downloads of the same layer happen at the same time
  286. if c, err := s.poolAdd("pull", "layer:"+id); err != nil {
  287. log.Debugf("Image (id: %s) pull is already running, skipping: %v", id, err)
  288. <-c
  289. }
  290. defer s.poolRemove("pull", "layer:"+id)
  291. if !s.graph.Exists(id) {
  292. out.Write(sf.FormatProgress(utils.TruncateID(id), "Pulling metadata", nil))
  293. var (
  294. imgJSON []byte
  295. imgSize int
  296. err error
  297. img *image.Image
  298. )
  299. retries := 5
  300. for j := 1; j <= retries; j++ {
  301. imgJSON, imgSize, err = r.GetRemoteImageJSON(id, endpoint, token)
  302. if err != nil && j == retries {
  303. out.Write(sf.FormatProgress(utils.TruncateID(id), "Error pulling dependent layers", nil))
  304. return layers_downloaded, err
  305. } else if err != nil {
  306. time.Sleep(time.Duration(j) * 500 * time.Millisecond)
  307. continue
  308. }
  309. img, err = image.NewImgJSON(imgJSON)
  310. layers_downloaded = true
  311. if err != nil && j == retries {
  312. out.Write(sf.FormatProgress(utils.TruncateID(id), "Error pulling dependent layers", nil))
  313. return layers_downloaded, fmt.Errorf("Failed to parse json: %s", err)
  314. } else if err != nil {
  315. time.Sleep(time.Duration(j) * 500 * time.Millisecond)
  316. continue
  317. } else {
  318. break
  319. }
  320. }
  321. for j := 1; j <= retries; j++ {
  322. // Get the layer
  323. status := "Pulling fs layer"
  324. if j > 1 {
  325. status = fmt.Sprintf("Pulling fs layer [retries: %d]", j)
  326. }
  327. out.Write(sf.FormatProgress(utils.TruncateID(id), status, nil))
  328. layer, err := r.GetRemoteImageLayer(img.ID, endpoint, token, int64(imgSize))
  329. if uerr, ok := err.(*url.Error); ok {
  330. err = uerr.Err
  331. }
  332. if terr, ok := err.(net.Error); ok && terr.Timeout() && j < retries {
  333. time.Sleep(time.Duration(j) * 500 * time.Millisecond)
  334. continue
  335. } else if err != nil {
  336. out.Write(sf.FormatProgress(utils.TruncateID(id), "Error pulling dependent layers", nil))
  337. return layers_downloaded, err
  338. }
  339. layers_downloaded = true
  340. defer layer.Close()
  341. err = s.graph.Register(img, imgJSON,
  342. utils.ProgressReader(layer, imgSize, out, sf, false, utils.TruncateID(id), "Downloading"))
  343. if terr, ok := err.(net.Error); ok && terr.Timeout() && j < retries {
  344. time.Sleep(time.Duration(j) * 500 * time.Millisecond)
  345. continue
  346. } else if err != nil {
  347. out.Write(sf.FormatProgress(utils.TruncateID(id), "Error downloading dependent layers", nil))
  348. return layers_downloaded, err
  349. } else {
  350. break
  351. }
  352. }
  353. }
  354. out.Write(sf.FormatProgress(utils.TruncateID(id), "Download complete", nil))
  355. }
  356. return layers_downloaded, nil
  357. }
  358. func WriteStatus(requestedTag string, out io.Writer, sf *utils.StreamFormatter, layers_downloaded bool) {
  359. if layers_downloaded {
  360. out.Write(sf.FormatStatus("", "Status: Downloaded newer image for %s", requestedTag))
  361. } else {
  362. out.Write(sf.FormatStatus("", "Status: Image is up to date for %s", requestedTag))
  363. }
  364. }
  365. // downloadInfo is used to pass information from download to extractor
  366. type downloadInfo struct {
  367. imgJSON []byte
  368. img *image.Image
  369. tmpFile *os.File
  370. length int64
  371. downloaded bool
  372. err chan error
  373. }
  374. func (s *TagStore) pullV2Repository(eng *engine.Engine, r *registry.Session, out io.Writer, localName, remoteName, tag string, sf *utils.StreamFormatter, parallel bool) error {
  375. var layersDownloaded bool
  376. if tag == "" {
  377. log.Debugf("Pulling tag list from V2 registry for %s", remoteName)
  378. tags, err := r.GetV2RemoteTags(remoteName, nil)
  379. if err != nil {
  380. return err
  381. }
  382. for _, t := range tags {
  383. if downloaded, err := s.pullV2Tag(eng, r, out, localName, remoteName, t, sf, parallel); err != nil {
  384. return err
  385. } else if downloaded {
  386. layersDownloaded = true
  387. }
  388. }
  389. } else {
  390. if downloaded, err := s.pullV2Tag(eng, r, out, localName, remoteName, tag, sf, parallel); err != nil {
  391. return err
  392. } else if downloaded {
  393. layersDownloaded = true
  394. }
  395. }
  396. requestedTag := localName
  397. if len(tag) > 0 {
  398. requestedTag = localName + ":" + tag
  399. }
  400. WriteStatus(requestedTag, out, sf, layersDownloaded)
  401. return nil
  402. }
  403. func (s *TagStore) pullV2Tag(eng *engine.Engine, r *registry.Session, out io.Writer, localName, remoteName, tag string, sf *utils.StreamFormatter, parallel bool) (bool, error) {
  404. log.Debugf("Pulling tag from V2 registry: %q", tag)
  405. manifestBytes, err := r.GetV2ImageManifest(remoteName, tag, nil)
  406. if err != nil {
  407. return false, err
  408. }
  409. manifest, verified, err := s.verifyManifest(eng, manifestBytes)
  410. if err != nil {
  411. return false, fmt.Errorf("error verifying manifest: %s", err)
  412. }
  413. if len(manifest.FSLayers) != len(manifest.History) {
  414. return false, fmt.Errorf("length of history not equal to number of layers")
  415. }
  416. if verified {
  417. out.Write(sf.FormatStatus(localName+":"+tag, "The image you are pulling has been digitally signed by Docker, Inc."))
  418. } else {
  419. out.Write(sf.FormatStatus(tag, "Pulling from %s", localName))
  420. }
  421. if len(manifest.FSLayers) == 0 {
  422. return false, fmt.Errorf("no blobSums in manifest")
  423. }
  424. downloads := make([]downloadInfo, len(manifest.FSLayers))
  425. for i := len(manifest.FSLayers) - 1; i >= 0; i-- {
  426. var (
  427. sumStr = manifest.FSLayers[i].BlobSum
  428. imgJSON = []byte(manifest.History[i].V1Compatibility)
  429. )
  430. img, err := image.NewImgJSON(imgJSON)
  431. if err != nil {
  432. return false, fmt.Errorf("failed to parse json: %s", err)
  433. }
  434. downloads[i].img = img
  435. // Check if exists
  436. if s.graph.Exists(img.ID) {
  437. log.Debugf("Image already exists: %s", img.ID)
  438. continue
  439. }
  440. chunks := strings.SplitN(sumStr, ":", 2)
  441. if len(chunks) < 2 {
  442. return false, fmt.Errorf("expected 2 parts in the sumStr, got %#v", chunks)
  443. }
  444. sumType, checksum := chunks[0], chunks[1]
  445. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), "Pulling fs layer", nil))
  446. downloadFunc := func(di *downloadInfo) error {
  447. log.Debugf("pulling blob %q to V1 img %s", sumStr, img.ID)
  448. if c, err := s.poolAdd("pull", "img:"+img.ID); err != nil {
  449. if c != nil {
  450. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), "Layer already being pulled by another client. Waiting.", nil))
  451. <-c
  452. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), "Download complete", nil))
  453. } else {
  454. log.Debugf("Image (id: %s) pull is already running, skipping: %v", img.ID, err)
  455. }
  456. } else {
  457. defer s.poolRemove("pull", "img:"+img.ID)
  458. tmpFile, err := ioutil.TempFile("", "GetV2ImageBlob")
  459. if err != nil {
  460. return err
  461. }
  462. r, l, err := r.GetV2ImageBlobReader(remoteName, sumType, checksum, nil)
  463. if err != nil {
  464. return err
  465. }
  466. defer r.Close()
  467. io.Copy(tmpFile, utils.ProgressReader(r, int(l), out, sf, false, utils.TruncateID(img.ID), "Downloading"))
  468. out.Write(sf.FormatProgress(utils.TruncateID(img.ID), "Download complete", nil))
  469. log.Debugf("Downloaded %s to tempfile %s", img.ID, tmpFile.Name())
  470. di.tmpFile = tmpFile
  471. di.length = l
  472. di.downloaded = true
  473. }
  474. di.imgJSON = imgJSON
  475. return nil
  476. }
  477. if parallel {
  478. downloads[i].err = make(chan error)
  479. go func(di *downloadInfo) {
  480. di.err <- downloadFunc(di)
  481. }(&downloads[i])
  482. } else {
  483. err := downloadFunc(&downloads[i])
  484. if err != nil {
  485. return false, err
  486. }
  487. }
  488. }
  489. var layersDownloaded bool
  490. for i := len(downloads) - 1; i >= 0; i-- {
  491. d := &downloads[i]
  492. if d.err != nil {
  493. err := <-d.err
  494. if err != nil {
  495. return false, err
  496. }
  497. }
  498. if d.downloaded {
  499. // if tmpFile is empty assume download and extracted elsewhere
  500. defer os.Remove(d.tmpFile.Name())
  501. defer d.tmpFile.Close()
  502. d.tmpFile.Seek(0, 0)
  503. if d.tmpFile != nil {
  504. err = s.graph.Register(d.img, d.imgJSON,
  505. utils.ProgressReader(d.tmpFile, int(d.length), out, sf, false, utils.TruncateID(d.img.ID), "Extracting"))
  506. if err != nil {
  507. return false, err
  508. }
  509. // FIXME: Pool release here for parallel tag pull (ensures any downloads block until fully extracted)
  510. }
  511. out.Write(sf.FormatProgress(utils.TruncateID(d.img.ID), "Pull complete", nil))
  512. layersDownloaded = true
  513. } else {
  514. out.Write(sf.FormatProgress(utils.TruncateID(d.img.ID), "Already exists", nil))
  515. }
  516. }
  517. if err = s.Set(localName, tag, downloads[0].img.ID, true); err != nil {
  518. return false, err
  519. }
  520. return layersDownloaded, nil
  521. }