pull_v1.go 11 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345
  1. package graph
  2. import (
  3. "errors"
  4. "fmt"
  5. "io"
  6. "net"
  7. "net/url"
  8. "strings"
  9. "time"
  10. "github.com/Sirupsen/logrus"
  11. "github.com/docker/distribution/registry/client/transport"
  12. "github.com/docker/docker/image"
  13. "github.com/docker/docker/pkg/progressreader"
  14. "github.com/docker/docker/pkg/streamformatter"
  15. "github.com/docker/docker/pkg/stringid"
  16. "github.com/docker/docker/registry"
  17. "github.com/docker/docker/utils"
  18. )
  19. type v1Puller struct {
  20. *TagStore
  21. endpoint registry.APIEndpoint
  22. config *ImagePullConfig
  23. sf *streamformatter.StreamFormatter
  24. repoInfo *registry.RepositoryInfo
  25. session *registry.Session
  26. }
  27. func (p *v1Puller) Pull(tag string) (fallback bool, err error) {
  28. if utils.DigestReference(tag) {
  29. // Allowing fallback, because HTTPS v1 is before HTTP v2
  30. return true, registry.ErrNoSupport{errors.New("Cannot pull by digest with v1 registry")}
  31. }
  32. tlsConfig, err := p.registryService.TLSConfig(p.repoInfo.Index.Name)
  33. if err != nil {
  34. return false, err
  35. }
  36. // Adds Docker-specific headers as well as user-specified headers (metaHeaders)
  37. tr := transport.NewTransport(
  38. // TODO(tiborvass): was ReceiveTimeout
  39. registry.NewTransport(tlsConfig),
  40. registry.DockerHeaders(p.config.MetaHeaders)...,
  41. )
  42. client := registry.HTTPClient(tr)
  43. v1Endpoint, err := p.endpoint.ToV1Endpoint(p.config.MetaHeaders)
  44. if err != nil {
  45. logrus.Debugf("Could not get v1 endpoint: %v", err)
  46. return true, err
  47. }
  48. p.session, err = registry.NewSession(client, p.config.AuthConfig, v1Endpoint)
  49. if err != nil {
  50. // TODO(dmcgowan): Check if should fallback
  51. logrus.Debugf("Fallback from error: %s", err)
  52. return true, err
  53. }
  54. if err := p.pullRepository(tag); err != nil {
  55. // TODO(dmcgowan): Check if should fallback
  56. return false, err
  57. }
  58. out := p.config.OutStream
  59. out.Write(p.sf.FormatStatus("", "%s: this image was pulled from a legacy registry. Important: This registry version will not be supported in future versions of docker.", p.repoInfo.CanonicalName))
  60. return false, nil
  61. }
  62. func (p *v1Puller) pullRepository(askedTag string) error {
  63. out := p.config.OutStream
  64. out.Write(p.sf.FormatStatus("", "Pulling repository %s", p.repoInfo.CanonicalName))
  65. repoData, err := p.session.GetRepositoryData(p.repoInfo.RemoteName)
  66. if err != nil {
  67. if strings.Contains(err.Error(), "HTTP code: 404") {
  68. return fmt.Errorf("Error: image %s not found", utils.ImageReference(p.repoInfo.RemoteName, askedTag))
  69. }
  70. // Unexpected HTTP error
  71. return err
  72. }
  73. logrus.Debugf("Retrieving the tag list")
  74. tagsList := make(map[string]string)
  75. if askedTag == "" {
  76. tagsList, err = p.session.GetRemoteTags(repoData.Endpoints, p.repoInfo.RemoteName)
  77. } else {
  78. var tagID string
  79. tagID, err = p.session.GetRemoteTag(repoData.Endpoints, p.repoInfo.RemoteName, askedTag)
  80. tagsList[askedTag] = tagID
  81. }
  82. if err != nil {
  83. if err == registry.ErrRepoNotFound && askedTag != "" {
  84. return fmt.Errorf("Tag %s not found in repository %s", askedTag, p.repoInfo.CanonicalName)
  85. }
  86. logrus.Errorf("unable to get remote tags: %s", err)
  87. return err
  88. }
  89. for tag, id := range tagsList {
  90. repoData.ImgList[id] = &registry.ImgData{
  91. ID: id,
  92. Tag: tag,
  93. Checksum: "",
  94. }
  95. }
  96. logrus.Debugf("Registering tags")
  97. // If no tag has been specified, pull them all
  98. if askedTag == "" {
  99. for tag, id := range tagsList {
  100. repoData.ImgList[id].Tag = tag
  101. }
  102. } else {
  103. // Otherwise, check that the tag exists and use only that one
  104. id, exists := tagsList[askedTag]
  105. if !exists {
  106. return fmt.Errorf("Tag %s not found in repository %s", askedTag, p.repoInfo.CanonicalName)
  107. }
  108. repoData.ImgList[id].Tag = askedTag
  109. }
  110. errors := make(chan error)
  111. layersDownloaded := false
  112. imgIDs := []string{}
  113. sessionID := p.session.ID()
  114. defer func() {
  115. p.graph.Release(sessionID, imgIDs...)
  116. }()
  117. for _, image := range repoData.ImgList {
  118. downloadImage := func(img *registry.ImgData) {
  119. if askedTag != "" && img.Tag != askedTag {
  120. errors <- nil
  121. return
  122. }
  123. if img.Tag == "" {
  124. logrus.Debugf("Image (id: %s) present in this repository but untagged, skipping", img.ID)
  125. errors <- nil
  126. return
  127. }
  128. // ensure no two downloads of the same image happen at the same time
  129. poolKey := "img:" + img.ID
  130. broadcaster, found := p.poolAdd("pull", poolKey)
  131. broadcaster.Add(out)
  132. if found {
  133. errors <- broadcaster.Wait()
  134. return
  135. }
  136. defer p.poolRemove("pull", poolKey)
  137. // we need to retain it until tagging
  138. p.graph.Retain(sessionID, img.ID)
  139. imgIDs = append(imgIDs, img.ID)
  140. broadcaster.Write(p.sf.FormatProgress(stringid.TruncateID(img.ID), fmt.Sprintf("Pulling image (%s) from %s", img.Tag, p.repoInfo.CanonicalName), nil))
  141. success := false
  142. var lastErr, err error
  143. var isDownloaded bool
  144. for _, ep := range p.repoInfo.Index.Mirrors {
  145. ep += "v1/"
  146. broadcaster.Write(p.sf.FormatProgress(stringid.TruncateID(img.ID), fmt.Sprintf("Pulling image (%s) from %s, mirror: %s", img.Tag, p.repoInfo.CanonicalName, ep), nil))
  147. if isDownloaded, err = p.pullImage(broadcaster, img.ID, ep); err != nil {
  148. // Don't report errors when pulling from mirrors.
  149. logrus.Debugf("Error pulling image (%s) from %s, mirror: %s, %s", img.Tag, p.repoInfo.CanonicalName, ep, err)
  150. continue
  151. }
  152. layersDownloaded = layersDownloaded || isDownloaded
  153. success = true
  154. break
  155. }
  156. if !success {
  157. for _, ep := range repoData.Endpoints {
  158. broadcaster.Write(p.sf.FormatProgress(stringid.TruncateID(img.ID), fmt.Sprintf("Pulling image (%s) from %s, endpoint: %s", img.Tag, p.repoInfo.CanonicalName, ep), nil))
  159. if isDownloaded, err = p.pullImage(broadcaster, img.ID, ep); err != nil {
  160. // It's not ideal that only the last error is returned, it would be better to concatenate the errors.
  161. // As the error is also given to the output stream the user will see the error.
  162. lastErr = err
  163. broadcaster.Write(p.sf.FormatProgress(stringid.TruncateID(img.ID), fmt.Sprintf("Error pulling image (%s) from %s, endpoint: %s, %s", img.Tag, p.repoInfo.CanonicalName, ep, err), nil))
  164. continue
  165. }
  166. layersDownloaded = layersDownloaded || isDownloaded
  167. success = true
  168. break
  169. }
  170. }
  171. if !success {
  172. err := fmt.Errorf("Error pulling image (%s) from %s, %v", img.Tag, p.repoInfo.CanonicalName, lastErr)
  173. broadcaster.Write(p.sf.FormatProgress(stringid.TruncateID(img.ID), err.Error(), nil))
  174. errors <- err
  175. broadcaster.CloseWithError(err)
  176. return
  177. }
  178. broadcaster.Write(p.sf.FormatProgress(stringid.TruncateID(img.ID), "Download complete", nil))
  179. errors <- nil
  180. }
  181. go downloadImage(image)
  182. }
  183. var lastError error
  184. for i := 0; i < len(repoData.ImgList); i++ {
  185. if err := <-errors; err != nil {
  186. lastError = err
  187. }
  188. }
  189. if lastError != nil {
  190. return lastError
  191. }
  192. for tag, id := range tagsList {
  193. if askedTag != "" && tag != askedTag {
  194. continue
  195. }
  196. if err := p.Tag(p.repoInfo.LocalName, tag, id, true); err != nil {
  197. return err
  198. }
  199. }
  200. requestedTag := p.repoInfo.LocalName
  201. if len(askedTag) > 0 {
  202. requestedTag = utils.ImageReference(p.repoInfo.LocalName, askedTag)
  203. }
  204. writeStatus(requestedTag, out, p.sf, layersDownloaded)
  205. return nil
  206. }
  207. func (p *v1Puller) pullImage(out io.Writer, imgID, endpoint string) (layersDownloaded bool, err error) {
  208. var history []string
  209. history, err = p.session.GetRemoteHistory(imgID, endpoint)
  210. if err != nil {
  211. return false, err
  212. }
  213. out.Write(p.sf.FormatProgress(stringid.TruncateID(imgID), "Pulling dependent layers", nil))
  214. // FIXME: Try to stream the images?
  215. // FIXME: Launch the getRemoteImage() in goroutines
  216. sessionID := p.session.ID()
  217. // As imgID has been retained in pullRepository, no need to retain again
  218. p.graph.Retain(sessionID, history[1:]...)
  219. defer p.graph.Release(sessionID, history[1:]...)
  220. layersDownloaded = false
  221. for i := len(history) - 1; i >= 0; i-- {
  222. id := history[i]
  223. // ensure no two downloads of the same layer happen at the same time
  224. poolKey := "layer:" + id
  225. broadcaster, found := p.poolAdd("pull", poolKey)
  226. broadcaster.Add(out)
  227. if found {
  228. logrus.Debugf("Image (id: %s) pull is already running, skipping", id)
  229. err = broadcaster.Wait()
  230. if err != nil {
  231. return layersDownloaded, err
  232. }
  233. continue
  234. }
  235. // This must use a closure so it captures the value of err when
  236. // the function returns, not when the 'defer' is evaluated.
  237. defer func() {
  238. p.poolRemoveWithError("pull", poolKey, err)
  239. }()
  240. if !p.graph.Exists(id) {
  241. broadcaster.Write(p.sf.FormatProgress(stringid.TruncateID(id), "Pulling metadata", nil))
  242. var (
  243. imgJSON []byte
  244. imgSize int64
  245. err error
  246. img *image.Image
  247. )
  248. retries := 5
  249. for j := 1; j <= retries; j++ {
  250. imgJSON, imgSize, err = p.session.GetRemoteImageJSON(id, endpoint)
  251. if err != nil && j == retries {
  252. broadcaster.Write(p.sf.FormatProgress(stringid.TruncateID(id), "Error pulling dependent layers", nil))
  253. return layersDownloaded, err
  254. } else if err != nil {
  255. time.Sleep(time.Duration(j) * 500 * time.Millisecond)
  256. continue
  257. }
  258. img, err = image.NewImgJSON(imgJSON)
  259. layersDownloaded = true
  260. if err != nil && j == retries {
  261. broadcaster.Write(p.sf.FormatProgress(stringid.TruncateID(id), "Error pulling dependent layers", nil))
  262. return layersDownloaded, fmt.Errorf("Failed to parse json: %s", err)
  263. } else if err != nil {
  264. time.Sleep(time.Duration(j) * 500 * time.Millisecond)
  265. continue
  266. } else {
  267. break
  268. }
  269. }
  270. for j := 1; j <= retries; j++ {
  271. // Get the layer
  272. status := "Pulling fs layer"
  273. if j > 1 {
  274. status = fmt.Sprintf("Pulling fs layer [retries: %d]", j)
  275. }
  276. broadcaster.Write(p.sf.FormatProgress(stringid.TruncateID(id), status, nil))
  277. layer, err := p.session.GetRemoteImageLayer(img.ID, endpoint, imgSize)
  278. if uerr, ok := err.(*url.Error); ok {
  279. err = uerr.Err
  280. }
  281. if terr, ok := err.(net.Error); ok && terr.Timeout() && j < retries {
  282. time.Sleep(time.Duration(j) * 500 * time.Millisecond)
  283. continue
  284. } else if err != nil {
  285. broadcaster.Write(p.sf.FormatProgress(stringid.TruncateID(id), "Error pulling dependent layers", nil))
  286. return layersDownloaded, err
  287. }
  288. layersDownloaded = true
  289. defer layer.Close()
  290. err = p.graph.Register(img,
  291. progressreader.New(progressreader.Config{
  292. In: layer,
  293. Out: broadcaster,
  294. Formatter: p.sf,
  295. Size: imgSize,
  296. NewLines: false,
  297. ID: stringid.TruncateID(id),
  298. Action: "Downloading",
  299. }))
  300. if terr, ok := err.(net.Error); ok && terr.Timeout() && j < retries {
  301. time.Sleep(time.Duration(j) * 500 * time.Millisecond)
  302. continue
  303. } else if err != nil {
  304. broadcaster.Write(p.sf.FormatProgress(stringid.TruncateID(id), "Error downloading dependent layers", nil))
  305. return layersDownloaded, err
  306. } else {
  307. break
  308. }
  309. }
  310. }
  311. broadcaster.Write(p.sf.FormatProgress(stringid.TruncateID(id), "Download complete", nil))
  312. broadcaster.Close()
  313. }
  314. return layersDownloaded, nil
  315. }