Browse Source

change perms on release and cleanup docs

Docker-DCO-1.1-Signed-off-by: Jessie Frazelle <hugs@docker.com> (github: jfrazelle)

Docker-DCO-1.1-Signed-off-by: Jessie Frazelle <princess@docker.com> (github: jfrazelle)

Docker-DCO-1.1-Signed-off-by: Jessie Frazelle <jess@docker.com> (github: jfrazelle)

Docker-DCO-1.1-Signed-off-by: Jessica Frazelle <jess@docker.com> (github: jfrazelle)
Jessica Frazelle 10 năm trước cách đây
mục cha
commit
844c0f154a
1 tập tin đã thay đổi với 76 bổ sung73 xóa
  1. 76 73
      docs/release.sh

+ 76 - 73
docs/release.sh

@@ -1,4 +1,4 @@
-#!/usr/bin/env bash
+#!/bin/bash
 set -e
 
 set -o pipefail
@@ -16,57 +16,31 @@ make AWS_S3_BUCKET=docs-stage.docker.com docs-release
 will then push the documentation site to your s3 bucket.
 
  Note: you can add `OPTIONS=--dryrun` to see what will be done without sending to the server
+ You can also add NOCACHE=1 to publish without a cache, which is what we do for the master docs.
 EOF
 	exit 1
 }
 
-[ "$AWS_S3_BUCKET" ] || usage
-
-VERSION=$(cat VERSION)
-
-if [ "$AWS_S3_BUCKET" == "docs.docker.com" ]; then
-	if [ "${VERSION%-dev}" != "$VERSION" ]; then
-		echo "Please do not push '-dev' documentation to docs.docker.com ($VERSION)"
-		exit 1
-	fi
-	cat > ./sources/robots.txt <<'EOF'
-User-agent: *
-Allow: /
-EOF
-
-else
+create_robots_txt() {
 	cat > ./sources/robots.txt <<'EOF'
 User-agent: *
 Disallow: /
 EOF
-fi
-
-# Remove the last version - 1.0.2-dev -> 1.0
-MAJOR_MINOR="v${VERSION%.*}"
-export MAJOR_MINOR
-
-export BUCKET=$AWS_S3_BUCKET
-
-export AWS_CONFIG_FILE=$(pwd)/awsconfig
-[ -e "$AWS_CONFIG_FILE" ] || usage
-export AWS_DEFAULT_PROFILE=$BUCKET
-
-echo "cfg file: $AWS_CONFIG_FILE ; profile: $AWS_DEFAULT_PROFILE"
+}
 
 setup_s3() {
-	echo "Create $BUCKET"
 	# Try creating the bucket. Ignore errors (it might already exist).
+	echo "create $BUCKET if it does not exist"
 	aws s3 mb --profile $BUCKET s3://$BUCKET 2>/dev/null || true
+
 	# Check access to the bucket.
 	echo "test $BUCKET exists"
 	aws s3 --profile $BUCKET ls s3://$BUCKET
+
 	# Make the bucket accessible through website endpoints.
 	echo "make $BUCKET accessible as a website"
 	#aws s3 website s3://$BUCKET --index-document index.html --error-document jsearch/index.html
-	s3conf=$(cat s3_website.json | envsubst)
-	echo
-	echo $s3conf
-	echo
+	local s3conf=$(cat s3_website.json | envsubst)
 	aws s3api --profile $BUCKET put-bucket-website --bucket $BUCKET --website-configuration "$s3conf"
 }
 
@@ -86,28 +60,22 @@ upload_current_documentation() {
 		cache=no-cache
 	fi
 
-	echo
-	echo "Uploading $src"
-	echo "  to $dst"
-	echo
+	printf "\nUploading $src to $dst\n"
 
 	# a really complicated way to send only the files we want
 	# if there are too many in any one set, aws s3 sync seems to fall over with 2 files to go
 	#  versions.html_fragment
-		include="--recursive --include \"*.$i\" "
-		echo "uploading *.$i"
-		run="aws s3 cp $src $dst $OPTIONS --profile $BUCKET --cache-control $cache --acl public-read $include"
-		echo "======================="
-		echo "$run"
-		echo "======================="
-		$run
+	include="--recursive --include \"*.$i\" "
+	run="aws s3 cp $src $dst $OPTIONS --profile $BUCKET --cache-control $cache --acl public-read $include"
+	printf "\n=====\n$run\n=====\n"
+	$run
 
 	# Make sure the search_content.json.gz file has the right content-encoding
 	aws s3 cp --profile $BUCKET --cache-control $cache --content-encoding="gzip" --acl public-read "site/search_content.json.gz" "$dst"
 }
 
 invalidate_cache() {
-	if [ "" == "$DISTRIBUTION_ID" ]; then
+	if [[ -z "$DISTRIBUTION_ID" ]]; then
 		echo "Skipping Cloudfront cache invalidation"
 		return
 	fi
@@ -116,17 +84,20 @@ invalidate_cache() {
 
 	aws configure set preview.cloudfront true
 
-	files=$(find site/ -not -name "*.md*" -type f | sed 's/site\///g')
+	# Get all the files
+	# not .md~ files
+	# replace spaces w %20 so urlencoded
+	files=( $(find site/ -not -name "*.md*" -type f | sed 's/site//g' | sed 's/ /%20/g') )
 
 	len=${#files[@]}
+	last_file=${files[$((len-1))]}
 
 	echo "aws cloudfront  create-invalidation --profile $AWS_S3_BUCKET --distribution-id $DISTRIBUTION_ID --invalidation-batch '" > batchfile
 	echo "{\"Paths\":{\"Quantity\":$len," >> batchfile
 	echo "\"Items\": [" >> batchfile
 
-	for file in "${files[@]}"
-	do
-		if [ "$file" == "${files[${#files[@]}-1]}" ]; then
+	for file in "${files[@]}" ; do
+		if [[ $file == $last_file ]]; then
 			comma=""
 		else
 			comma=","
@@ -134,33 +105,65 @@ invalidate_cache() {
 		echo "\"$dst$file\"$comma" >> batchfile
 	done
 
-	echo "]}, \"CallerReference\":" >> batchfile
-	echo "\"$(date)\"}'" >> batchfile
+	echo "]}, \"CallerReference\":\"$(date)\"}'" >> batchfile
 
-
-	echo "-----"
-	cat batchfile
-	echo "-----"
 	sh batchfile
-	echo "-----"
 }
 
+main() {
+	[ "$AWS_S3_BUCKET" ] || usage
+
+	# Make sure there is an awsconfig file
+	export AWS_CONFIG_FILE=$(pwd)/awsconfig
+	[ -f "$AWS_CONFIG_FILE" ] || usage
 
-if [ "$OPTIONS" != "--dryrun" ]; then
-	setup_s3
-fi
+	# Get the version
+	VERSION=$(cat VERSION)
 
-# Default to only building the version specific docs so we don't clober the latest by accident with old versions
-if [ "$BUILD_ROOT" == "yes" ]; then
-	echo "Building root documentation"
+	# Disallow pushing dev docs to master
+	if [ "$AWS_S3_BUCKET" == "docs.docker.com" ] && [ "${VERSION%-dev}" != "$VERSION" ]; then
+		echo "Please do not push '-dev' documentation to docs.docker.com ($VERSION)"
+		exit 1
+	fi
+
+	# Clean version - 1.0.2-dev -> 1.0
+	export MAJOR_MINOR="v${VERSION%.*}"
+
+	export BUCKET=$AWS_S3_BUCKET
+	export AWS_DEFAULT_PROFILE=$BUCKET
+
+	# debug variables
+	echo "bucket: $BUCKET, full version: $VERSION, major-minor: $MAJOR_MINOR"
+	echo "cfg file: $AWS_CONFIG_FILE ; profile: $AWS_DEFAULT_PROFILE"
+
+	# create the robots.txt
+	create_robots_txt
+
+	if [ "$OPTIONS" != "--dryrun" ]; then
+		setup_s3
+	fi
+
+	# Default to only building the version specific docs
+	# so we don't clober the latest by accident with old versions
+	if [ "$BUILD_ROOT" == "yes" ]; then
+		echo "Building root documentation"
+		build_current_documentation
+
+		echo "Uploading root documentation"
+		upload_current_documentation
+		[ "$NOCACHE" ] || invalidate_cache
+	fi
+
+	#build again with /v1.0/ prefix
+	sed -i "s/^site_url:.*/site_url: \/$MAJOR_MINOR\//" mkdocs.yml
+	echo "Building the /$MAJOR_MINOR/ documentation"
 	build_current_documentation
-	upload_current_documentation
-	[ "$NOCACHE" ] || invalidate_cache
-fi
-
-#build again with /v1.0/ prefix
-sed -i "s/^site_url:.*/site_url: \/$MAJOR_MINOR\//" mkdocs.yml
-echo "Building the /$MAJOR_MINOR/ documentation"
-build_current_documentation
-upload_current_documentation "/$MAJOR_MINOR/"
-[ "$NOCACHE" ] || invalidate_cache "/$MAJOR_MINOR"
+
+	echo "Uploading the documentation"
+	upload_current_documentation "/$MAJOR_MINOR/"
+
+	# Invalidating cache
+	[ "$NOCACHE" ] || invalidate_cache "/$MAJOR_MINOR"
+}
+
+main