|
@@ -72,6 +72,9 @@ setup_s3() {
|
|
|
|
|
|
build_current_documentation() {
|
|
|
mkdocs build
|
|
|
+ cd site/
|
|
|
+ gzip -9k -f search_content.json
|
|
|
+ cd ..
|
|
|
}
|
|
|
|
|
|
upload_current_documentation() {
|
|
@@ -87,8 +90,6 @@ upload_current_documentation() {
|
|
|
echo "Uploading $src"
|
|
|
echo " to $dst"
|
|
|
echo
|
|
|
- #s3cmd --recursive --follow-symlinks --preserve --acl-public sync "$src" "$dst"
|
|
|
- #aws s3 cp --profile $BUCKET --cache-control "max-age=3600" --acl public-read "site/search_content.json" "$dst"
|
|
|
|
|
|
# a really complicated way to send only the files we want
|
|
|
# if there are too many in any one set, aws s3 sync seems to fall over with 2 files to go
|
|
@@ -100,6 +101,9 @@ upload_current_documentation() {
|
|
|
echo "$run"
|
|
|
echo "======================="
|
|
|
$run
|
|
|
+
|
|
|
+ # Make sure the search_content.json.gz file has the right content-encoding
|
|
|
+ aws s3 cp --profile $BUCKET --cache-control $cache --content-encoding="gzip" --acl public-read "site/search_content.json.gz" "$dst"
|
|
|
}
|
|
|
|
|
|
invalidate_cache() {
|