Compare commits
No commits in common. "master" and "v2.3" have entirely different histories.
30 changed files with 330 additions and 568 deletions
60
.github/workflows/buildx.yaml
vendored
60
.github/workflows/buildx.yaml
vendored
|
@ -1,60 +0,0 @@
|
|||
name: buildx
|
||||
|
||||
on:
|
||||
push:
|
||||
tags:
|
||||
- 'v*'
|
||||
|
||||
jobs:
|
||||
buildx:
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
-
|
||||
name: Checkout
|
||||
uses: actions/checkout@v2
|
||||
-
|
||||
name: Prepare
|
||||
id: prepare
|
||||
run: |
|
||||
DOCKER_IMAGE=andreimarcu/linx-server
|
||||
DOCKER_PLATFORMS=linux/amd64,linux/arm/v6,linux/arm/v7,linux/arm64/v8,linux/386
|
||||
VERSION=version-${GITHUB_REF#refs/tags/v}
|
||||
TAGS="--tag ${DOCKER_IMAGE}:${VERSION} --tag ${DOCKER_IMAGE}:latest"
|
||||
|
||||
echo ::set-output name=docker_image::${DOCKER_IMAGE}
|
||||
echo ::set-output name=version::${VERSION}
|
||||
echo ::set-output name=buildx_args::--platform ${DOCKER_PLATFORMS} \
|
||||
--build-arg VERSION=${VERSION} \
|
||||
--build-arg BUILD_DATE=$(date -u +'%Y-%m-%dT%H:%M:%SZ') \
|
||||
--build-arg VCS_REF=${GITHUB_SHA::8} \
|
||||
${TAGS} --file Dockerfile .
|
||||
-
|
||||
name: Set up Docker Buildx
|
||||
uses: crazy-max/ghaction-docker-buildx@v3
|
||||
-
|
||||
name: Docker Buildx (build)
|
||||
run: |
|
||||
docker buildx build --output "type=image,push=false" ${{ steps.prepare.outputs.buildx_args }}
|
||||
-
|
||||
name: Docker Login
|
||||
if: success()
|
||||
env:
|
||||
DOCKER_USERNAME: ${{ secrets.DOCKER_USERNAME }}
|
||||
DOCKER_PASSWORD: ${{ secrets.DOCKER_PASSWORD }}
|
||||
run: |
|
||||
echo "${DOCKER_PASSWORD}" | docker login --username "${DOCKER_USERNAME}" --password-stdin
|
||||
-
|
||||
name: Docker Buildx (push)
|
||||
if: success()
|
||||
run: |
|
||||
docker buildx build --output "type=image,push=true" ${{ steps.prepare.outputs.buildx_args }}
|
||||
-
|
||||
name: Docker Check Manifest
|
||||
if: always()
|
||||
run: |
|
||||
docker run --rm mplatform/mquery ${{ steps.prepare.outputs.docker_image }}:${{ steps.prepare.outputs.version }}
|
||||
-
|
||||
name: Clear
|
||||
if: always()
|
||||
run: |
|
||||
rm -f ${HOME}/.docker/config.json
|
5
.gitignore
vendored
5
.gitignore
vendored
|
@ -29,9 +29,8 @@ _testmain.go
|
|||
*.prof
|
||||
|
||||
linx-server
|
||||
linx-cleanup/linx-cleanup
|
||||
linx-genkey/linx-genkey
|
||||
linx-server.conf
|
||||
linx-cleanup
|
||||
linx-genkey
|
||||
files/
|
||||
meta/
|
||||
binaries/
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
FROM golang:1.14-alpine3.11 AS build
|
||||
FROM golang:alpine3.8 AS build
|
||||
|
||||
COPY . /go/src/github.com/andreimarcu/linx-server
|
||||
WORKDIR /go/src/github.com/andreimarcu/linx-server
|
||||
|
@ -8,7 +8,7 @@ RUN set -ex \
|
|||
&& go get -v . \
|
||||
&& apk del .build-deps
|
||||
|
||||
FROM alpine:3.11
|
||||
FROM alpine:3.8
|
||||
|
||||
COPY --from=build /go/bin/linx-server /usr/local/bin/linx-server
|
||||
|
||||
|
|
192
README.md
192
README.md
|
@ -1,136 +1,67 @@
|
|||
# Development on this repository has been frozen.
|
||||
|
||||
Feel free to send a pull request if you are maintaining an active fork of this project to add a link to your repository in this readme.
|
||||
|
||||
|
||||
### Active Forks
|
||||
- ZizzyDizzyMC: [https://github.com/ZizzyDizzyMC/linx-server/](https://github.com/ZizzyDizzyMC/linx-server/)
|
||||
- Seb3thehacker: [https://github.com/Seb3thehacker/linx-server](https://github.com/Seb3thehacker/linx-server)
|
||||
|
||||
|
||||
---
|
||||
|
||||
|
||||
linx-server
|
||||
linx-server
|
||||
======
|
||||
[](https://travis-ci.org/andreimarcu/linx-server)
|
||||
|
||||
Self-hosted file/media sharing website.
|
||||
|
||||
### Clients
|
||||
**Official**
|
||||
- CLI: **linx-client** - [Source](https://github.com/andreimarcu/linx-client)
|
||||
|
||||
**Unofficial**
|
||||
- Android: **LinxShare** - [Source](https://github.com/iksteen/LinxShare/) | [Google Play](https://play.google.com/store/apps/details?id=org.thegraveyard.linxshare)
|
||||
- CLI: **golinx** - [Source](https://github.com/mutantmonkey/golinx)
|
||||
### Demo
|
||||
You can see what it looks like using the demo: [https://demo.linx-server.net/](https://demo.linx-server.net/)
|
||||
|
||||
|
||||
### Features
|
||||
|
||||
- Display common filetypes (image, video, audio, markdown, pdf)
|
||||
- Display syntax-highlighted code with in-place editing
|
||||
- Documented API with keys for restricting uploads
|
||||
- Documented API with keys if need to restrict uploads (can use [linx-client](https://github.com/andreimarcu/linx-client) for uploading through command-line)
|
||||
- Torrent download of files using web seeding
|
||||
- File expiry, deletion key, file access key, and random filename options
|
||||
|
||||
|
||||
### Screenshots
|
||||
<img width="730" src="https://user-images.githubusercontent.com/4650950/76579039-03c82680-6488-11ea-8e23-4c927386fbd9.png" />
|
||||
|
||||
<img width="180" src="https://user-images.githubusercontent.com/4650950/76578903-771d6880-6487-11ea-8baf-a4a23fef4d26.png" /> <img width="180" src="https://user-images.githubusercontent.com/4650950/76578910-7be21c80-6487-11ea-9a0a-587d59bc5f80.png" /> <img width="180" src="https://user-images.githubusercontent.com/4650950/76578908-7b498600-6487-11ea-8994-ee7b6eb9cdb1.png" /> <img width="180" src="https://user-images.githubusercontent.com/4650950/76578907-7b498600-6487-11ea-8941-8f582bf87fb0.png" />
|
||||
<img width="200" src="https://user-images.githubusercontent.com/4650950/51735725-0033cf00-203d-11e9-8a97-f543330a92ec.png" /> <img width="200" src="https://user-images.githubusercontent.com/4650950/51735724-0033cf00-203d-11e9-8fe0-77442eaa8705.png" /> <img width="200" src="https://user-images.githubusercontent.com/4650950/51735726-0033cf00-203d-11e9-9fca-095a97e46ce8.png" /> <img width="200" src="https://user-images.githubusercontent.com/4650950/51735728-0033cf00-203d-11e9-90e9-4f2d36332fc4.png" />
|
||||
|
||||
|
||||
Getting started
|
||||
Get release and run
|
||||
-------------------
|
||||
|
||||
#### Using Docker
|
||||
1. Create directories ```files``` and ```meta``` and run ```chown -R 65534:65534 meta && chown -R 65534:65534 files```
|
||||
2. Create a config file (example provided in repo), we'll refer to it as __linx-server.conf__ in the following examples
|
||||
|
||||
|
||||
|
||||
Example running
|
||||
```
|
||||
docker run -p 8080:8080 -v /path/to/linx-server.conf:/data/linx-server.conf -v /path/to/meta:/data/meta -v /path/to/files:/data/files andreimarcu/linx-server -config /data/linx-server.conf
|
||||
```
|
||||
|
||||
Example with docker-compose
|
||||
```
|
||||
version: '2.2'
|
||||
services:
|
||||
linx-server:
|
||||
container_name: linx-server
|
||||
image: andreimarcu/linx-server
|
||||
command: -config /data/linx-server.conf
|
||||
volumes:
|
||||
- /path/to/files:/data/files
|
||||
- /path/to/meta:/data/meta
|
||||
- /path/to/linx-server.conf:/data/linx-server.conf
|
||||
network_mode: bridge
|
||||
ports:
|
||||
- "8080:8080"
|
||||
restart: unless-stopped
|
||||
```
|
||||
Ideally, you would use a reverse proxy such as nginx or caddy to handle TLS certificates.
|
||||
|
||||
#### Using a binary release
|
||||
|
||||
1. Grab the latest binary from the [releases](https://github.com/andreimarcu/linx-server/releases), then run ```go install```
|
||||
2. Run ```linx-server -config path/to/linx-server.conf```
|
||||
1. Grab the latest binary from the [releases](https://github.com/andreimarcu/linx-server/releases)
|
||||
2. Run ```./linx-server```
|
||||
|
||||
|
||||
Usage
|
||||
-----
|
||||
|
||||
#### Configuration
|
||||
All configuration options are accepted either as arguments or can be placed in a file as such (see example file linx-server.conf.example in repo):
|
||||
All configuration options are accepted either as arguments or can be placed in an ini-style file as such:
|
||||
```ini
|
||||
bind = 127.0.0.1:8080
|
||||
sitename = myLinx
|
||||
maxsize = 4294967296
|
||||
maxexpiry = 86400
|
||||
# ... etc
|
||||
```
|
||||
...and then run ```linx-server -config path/to/linx-server.conf```
|
||||
allowhotlink = true
|
||||
# etc
|
||||
```
|
||||
...and then invoke ```linx-server -config path/to/config.ini```
|
||||
|
||||
#### Options
|
||||
|
||||
|Option|Description
|
||||
|------|-----------
|
||||
| ```bind = 127.0.0.1:8080``` | what to bind to (default is 127.0.0.1:8080)
|
||||
| ```sitename = myLinx``` | the site name displayed on top (default is inferred from Host header)
|
||||
| ```siteurl = https://mylinx.example.org/``` | the site url (default is inferred from execution context)
|
||||
| ```selifpath = selif``` | path relative to site base url (the "selif" in mylinx.example.org/selif/image.jpg) where files are accessed directly (default: selif)
|
||||
| ```maxsize = 4294967296``` | maximum upload file size in bytes (default 4GB)
|
||||
| ```maxexpiry = 86400``` | maximum expiration time in seconds (default is 0, which is no expiry)
|
||||
| ```allowhotlink = true``` | Allow file hotlinking
|
||||
| ```contentsecuritypolicy = "..."``` | Content-Security-Policy header for pages (default is "default-src 'self'; img-src 'self' data:; style-src 'self' 'unsafe-inline'; frame-ancestors 'self';")
|
||||
| ```filecontentsecuritypolicy = "..."``` | Content-Security-Policy header for files (default is "default-src 'none'; img-src 'self'; object-src 'self'; media-src 'self'; style-src 'self' 'unsafe-inline'; frame-ancestors 'self';")
|
||||
| ```refererpolicy = "..."``` | Referrer-Policy header for pages (default is "same-origin")
|
||||
| ```filereferrerpolicy = "..."``` | Referrer-Policy header for files (default is "same-origin")
|
||||
| ```xframeoptions = "..." ``` | X-Frame-Options header (default is "SAMEORIGIN")
|
||||
| ```remoteuploads = true``` | (optionally) enable remote uploads (/upload?url=https://...)
|
||||
| ```nologs = true``` | (optionally) disable request logs in stdout
|
||||
| ```force-random-filename = true``` | (optionally) force the use of random filenames
|
||||
| ```custompagespath = custom_pages/``` | (optionally) specify path to directory containing markdown pages (must end in .md) that will be added to the site navigation (this can be useful for providing contact/support information and so on). For example, custom_pages/My_Page.md will become My Page in the site navigation
|
||||
|
||||
|
||||
#### Cleaning up expired files
|
||||
When files expire, access is disabled immediately, but the files and metadata
|
||||
will persist on disk until someone attempts to access them. You can set the following option to run cleanup every few minutes. This can also be done using a separate utility found the linx-cleanup directory.
|
||||
|
||||
|
||||
|Option|Description
|
||||
|------|-----------
|
||||
| ```cleanup-every-minutes = 5``` | How often to clean up expired files in minutes (default is 0, which means files will be cleaned up as they are accessed)
|
||||
|
||||
- ```-bind 127.0.0.1:8080``` -- what to bind to (default is 127.0.0.1:8080)
|
||||
- ```-sitename myLinx``` -- the site name displayed on top (default is inferred from Host header)
|
||||
- ```-siteurl "https://mylinx.example.org/"``` -- the site url (default is inferred from execution context)
|
||||
- ```-selifpath "selif"``` -- path relative to site base url (the "selif" in mylinx.example.org/selif/image.jpg) where files are accessed directly (default: selif)
|
||||
- ```-maxsize 4294967296``` -- maximum upload file size in bytes (default 4GB)
|
||||
- ```-maxexpiry 86400``` -- maximum expiration time in seconds (default is 0, which is no expiry)
|
||||
- ```-allowhotlink``` -- Allow file hotlinking
|
||||
- ```-contentsecuritypolicy "..."``` -- Content-Security-Policy header for pages (default is "default-src 'self'; img-src 'self' data:; style-src 'self' 'unsafe-inline'; frame-ancestors 'self';")
|
||||
- ```-filecontentsecuritypolicy "..."``` -- Content-Security-Policy header for files (default is "default-src 'none'; img-src 'self'; object-src 'self'; media-src 'self'; style-src 'self' 'unsafe-inline'; frame-ancestors 'self';")
|
||||
- ```-refererpolicy "..."``` -- Referrer-Policy header for pages (default is "same-origin")
|
||||
- ```-filereferrerpolicy "..."``` -- Referrer-Policy header for files (default is "same-origin")
|
||||
- ```-xframeoptions "..." ``` -- X-Frame-Options header (default is "SAMEORIGIN")
|
||||
- ```-remoteuploads``` -- (optionally) enable remote uploads (/upload?url=https://...)
|
||||
- ```-nologs``` -- (optionally) disable request logs in stdout
|
||||
- ```-force-random-filename``` -- (optionally) force the use of random filenames
|
||||
- ```-custompagespath "custom_pages"``` -- (optionally) specify path to directory containing markdown pages (must end in .md) that will be added to the site navigation (this can be useful for providing contact/support information and so on). For example, custom_pages/My_Page.md will become My Page in the site navigation
|
||||
|
||||
#### Require API Keys for uploads
|
||||
|
||||
|Option|Description
|
||||
|------|-----------
|
||||
| ```authfile = path/to/authfile``` | (optionally) require authorization for upload/delete by providing a newline-separated file of scrypted auth keys
|
||||
| ```remoteauthfile = path/to/remoteauthfile``` | (optionally) require authorization for remote uploads by providing a newline-separated file of scrypted auth keys
|
||||
| ```basicauth = true``` | (optionally) allow basic authorization to upload or paste files from browser when `-authfile` is enabled. When uploading, you will be prompted to enter a user and password - leave the user blank and use your auth key as the password
|
||||
- ```-authfile path/to/authfile``` -- (optionally) require authorization for upload/delete by providing a newline-separated file of scrypted auth keys
|
||||
- ```-remoteauthfile path/to/remoteauthfile``` -- (optionally) require authorization for remote uploads by providing a newline-separated file of scrypted auth keys
|
||||
- ```-basicauth``` -- (optionally) allow basic authorization to upload or paste files from browser when `-authfile` is enabled. When uploading, you will be prompted to enter a user and password - leave the user blank and use your auth key as the password
|
||||
|
||||
A helper utility ```linx-genkey``` is provided which hashes keys to the format required in the auth files.
|
||||
|
||||
|
@ -139,25 +70,37 @@ The following storage backends are available:
|
|||
|
||||
|Name|Notes|Options
|
||||
|----|-----|-------
|
||||
|LocalFS|Enabled by default, this backend uses the filesystem|```filespath = files/``` -- Path to store uploads (default is files/)<br />```metapath = meta/``` -- Path to store information about uploads (default is meta/)|
|
||||
|S3|Use with any S3-compatible provider.<br> This implementation will stream files through the linx instance (every download will request and stream the file from the S3 bucket). File metadata will be stored as tags on the object in the bucket.<br><br>For high-traffic environments, one might consider using an external caching layer such as described [in this article](https://blog.sentry.io/2017/03/01/dodging-s3-downtime-with-nginx-and-haproxy.html).|```s3-endpoint = https://...``` -- S3 endpoint<br>```s3-region = us-east-1``` -- S3 region<br>```s3-bucket = mybucket``` -- S3 bucket to use for files and metadata<br>```s3-force-path-style = true``` (optional) -- force path-style addresing (e.g. https://<span></span>s3.amazonaws.com/linx/example.txt)<br><br>Environment variables to provide:<br>```AWS_ACCESS_KEY_ID``` -- the S3 access key<br>```AWS_SECRET_ACCESS_KEY ``` -- the S3 secret key<br>```AWS_SESSION_TOKEN``` (optional) -- the S3 session token|
|
||||
|LocalFS|Enabled by default, this backend uses the filesystem|```-filespath files/``` -- Path to store uploads (default is files/)<br />```-metapath meta/``` -- Path to store information about uploads (default is meta/)|
|
||||
|S3|Use with any S3-compatible provider.<br> This implementation will stream files through the linx instance (every download will request and stream the file from the S3 bucket).<br><br>For high-traffic environments, one might consider using an external caching layer such as described [in this article](https://blog.sentry.io/2017/03/01/dodging-s3-downtime-with-nginx-and-haproxy.html).|```-s3-endpoint https://...``` -- S3 endpoint<br>```-s3-region us-east-1``` -- S3 region<br>```-s3-bucket mybucket``` -- S3 bucket to use for files and metadata<br>```-s3-force-path-style``` (optional) -- force path-style addresing (e.g. https://<span></span>s3.amazonaws.com/linx/example.txt)<br><br>Environment variables to provide:<br>```AWS_ACCESS_KEY_ID``` -- the S3 access key<br>```AWS_SECRET_ACCESS_KEY ``` -- the S3 secret key<br>```AWS_SESSION_TOKEN``` (optional) -- the S3 session token|
|
||||
|
||||
|
||||
#### SSL with built-in server
|
||||
|Option|Description
|
||||
|------|-----------
|
||||
| ```certfile = path/to/your.crt``` | Path to the ssl certificate (required if you want to use the https server)
|
||||
| ```keyfile = path/to/your.key``` | Path to the ssl key (required if you want to use the https server)
|
||||
- ```-certfile path/to/your.crt``` -- Path to the ssl certificate (required if you want to use the https server)
|
||||
- ```-keyfile path/to/your.key``` -- Path to the ssl key (required if you want to use the https server)
|
||||
|
||||
#### Use with http proxy
|
||||
|Option|Description
|
||||
|------|-----------
|
||||
| ```realip = true``` | let linx-server know you (nginx, etc) are providing the X-Real-IP and/or X-Forwarded-For headers.
|
||||
- ```-realip``` -- let linx-server know you (nginx, etc) are providing the X-Real-IP and/or X-Forwarded-For headers.
|
||||
|
||||
#### Use with fastcgi
|
||||
|Option|Description
|
||||
|------|-----------
|
||||
| ```fastcgi = true``` | serve through fastcgi
|
||||
- ```-fastcgi``` -- serve through fastcgi
|
||||
|
||||
|
||||
Cleaning up expired files
|
||||
-------------------------
|
||||
When files expire, access is disabled immediately, but the files and metadata
|
||||
will persist on disk until someone attempts to access them. If you'd like to
|
||||
automatically clean up files that have expired, you can use the included
|
||||
`linx-cleanup` utility. To run it automatically, use a cronjob or similar type
|
||||
of scheduled task.
|
||||
|
||||
You should be careful to ensure that only one instance of `linx-client` runs at
|
||||
a time to avoid unexpected behavior. It does not implement any type of locking.
|
||||
|
||||
#### Options
|
||||
- ```-filespath files/``` -- Path to stored uploads (default is files/)
|
||||
- ```-metapath meta/``` -- Path to stored information about uploads (default is meta/)
|
||||
- ```-nologs``` -- (optionally) disable deletion logs in stdout
|
||||
|
||||
|
||||
Deployment
|
||||
----------
|
||||
|
@ -182,16 +125,29 @@ server {
|
|||
}
|
||||
}
|
||||
```
|
||||
And run linx-server with the ```fastcgi = true``` option.
|
||||
And run linx-server with the ```-fastcgi``` option.
|
||||
|
||||
#### 2. Using the built-in https server
|
||||
Run linx-server with the ```certfile = path/to/cert.file``` and ```keyfile = path/to/key.file``` options.
|
||||
Run linx-server with the ```-certfile path/to/cert.file``` and ```-keyfile path/to/key.file``` options.
|
||||
|
||||
#### 3. Using the built-in http server
|
||||
Run linx-server normally.
|
||||
|
||||
#### 4. Using Docker with the built-in http server
|
||||
First, build the image:
|
||||
```docker build -t linx-server .```
|
||||
|
||||
You'll need some directories for the persistent storage. For the purposes of this example, we will use `/media/meta` and `/media/files`.
|
||||
|
||||
Then, run it:
|
||||
```docker run -p 8080:8080 -v /media/meta:/data/meta -v /media/files:/data/files linx-server```
|
||||
|
||||
|
||||
Development
|
||||
-----------
|
||||
Any help is welcome, PRs will be reviewed and merged accordingly.
|
||||
The official IRC channel is #linx on irc.oftc.net
|
||||
|
||||
1. ```go get -u github.com/andreimarcu/linx-server ```
|
||||
2. ```cd $GOPATH/src/github.com/andreimarcu/linx-server ```
|
||||
3. ```go build && ./linx-server```
|
||||
|
@ -216,4 +172,4 @@ along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|||
|
||||
Author
|
||||
-------
|
||||
Andrei Marcu, https://andreim.net/
|
||||
Andrei Marcu, http://andreim.net/
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
package apikeys
|
||||
package main
|
||||
|
||||
import (
|
||||
"bufio"
|
||||
|
@ -24,18 +24,16 @@ const (
|
|||
type AuthOptions struct {
|
||||
AuthFile string
|
||||
UnauthMethods []string
|
||||
BasicAuth bool
|
||||
SiteName string
|
||||
SitePath string
|
||||
}
|
||||
|
||||
type ApiKeysMiddleware struct {
|
||||
type auth struct {
|
||||
successHandler http.Handler
|
||||
failureHandler http.Handler
|
||||
authKeys []string
|
||||
o AuthOptions
|
||||
}
|
||||
|
||||
func ReadAuthKeys(authFile string) []string {
|
||||
func readAuthKeys(authFile string) []string {
|
||||
var authKeys []string
|
||||
|
||||
f, err := os.Open(authFile)
|
||||
|
@ -57,7 +55,7 @@ func ReadAuthKeys(authFile string) []string {
|
|||
return authKeys
|
||||
}
|
||||
|
||||
func CheckAuth(authKeys []string, key string) (result bool, err error) {
|
||||
func checkAuth(authKeys []string, key string) (result bool, err error) {
|
||||
checkKey, err := scrypt.Key([]byte(key), []byte(scryptSalt), scryptN, scryptr, scryptp, scryptKeyLen)
|
||||
if err != nil {
|
||||
return
|
||||
|
@ -75,74 +73,53 @@ func CheckAuth(authKeys []string, key string) (result bool, err error) {
|
|||
return
|
||||
}
|
||||
|
||||
func (a ApiKeysMiddleware) getSitePrefix() string {
|
||||
prefix := a.o.SitePath
|
||||
if len(prefix) <= 0 || prefix[0] != '/' {
|
||||
prefix = "/" + prefix
|
||||
}
|
||||
return prefix
|
||||
}
|
||||
|
||||
func (a ApiKeysMiddleware) goodAuthorizationHandler(w http.ResponseWriter, r *http.Request) {
|
||||
w.Header().Set("Location", a.getSitePrefix())
|
||||
w.WriteHeader(http.StatusFound)
|
||||
}
|
||||
|
||||
func (a ApiKeysMiddleware) badAuthorizationHandler(w http.ResponseWriter, r *http.Request) {
|
||||
if a.o.BasicAuth {
|
||||
rs := ""
|
||||
if a.o.SiteName != "" {
|
||||
rs = fmt.Sprintf(` realm="%s"`, a.o.SiteName)
|
||||
}
|
||||
w.Header().Set("WWW-Authenticate", `Basic`+rs)
|
||||
}
|
||||
http.Error(w, http.StatusText(http.StatusUnauthorized), http.StatusUnauthorized)
|
||||
}
|
||||
|
||||
func (a ApiKeysMiddleware) ServeHTTP(w http.ResponseWriter, r *http.Request) {
|
||||
var successHandler http.Handler
|
||||
prefix := a.getSitePrefix()
|
||||
|
||||
if r.URL.Path == prefix+"auth" {
|
||||
successHandler = http.HandlerFunc(a.goodAuthorizationHandler)
|
||||
} else {
|
||||
successHandler = a.successHandler
|
||||
}
|
||||
|
||||
if sliceContains(a.o.UnauthMethods, r.Method) && r.URL.Path != prefix+"auth" {
|
||||
func (a auth) ServeHTTP(w http.ResponseWriter, r *http.Request) {
|
||||
if sliceContains(a.o.UnauthMethods, r.Method) {
|
||||
// allow unauthenticated methods
|
||||
successHandler.ServeHTTP(w, r)
|
||||
a.successHandler.ServeHTTP(w, r)
|
||||
return
|
||||
}
|
||||
|
||||
key := r.Header.Get("Linx-Api-Key")
|
||||
if key == "" && a.o.BasicAuth {
|
||||
if key == "" && Config.basicAuth {
|
||||
_, password, ok := r.BasicAuth()
|
||||
if ok {
|
||||
key = password
|
||||
}
|
||||
}
|
||||
|
||||
result, err := CheckAuth(a.authKeys, key)
|
||||
result, err := checkAuth(a.authKeys, key)
|
||||
if err != nil || !result {
|
||||
http.HandlerFunc(a.badAuthorizationHandler).ServeHTTP(w, r)
|
||||
a.failureHandler.ServeHTTP(w, r)
|
||||
return
|
||||
}
|
||||
|
||||
successHandler.ServeHTTP(w, r)
|
||||
a.successHandler.ServeHTTP(w, r)
|
||||
}
|
||||
|
||||
func NewApiKeysMiddleware(o AuthOptions) func(*web.C, http.Handler) http.Handler {
|
||||
func UploadAuth(o AuthOptions) func(*web.C, http.Handler) http.Handler {
|
||||
fn := func(c *web.C, h http.Handler) http.Handler {
|
||||
return ApiKeysMiddleware{
|
||||
return auth{
|
||||
successHandler: h,
|
||||
authKeys: ReadAuthKeys(o.AuthFile),
|
||||
failureHandler: http.HandlerFunc(badAuthorizationHandler),
|
||||
authKeys: readAuthKeys(o.AuthFile),
|
||||
o: o,
|
||||
}
|
||||
}
|
||||
return fn
|
||||
}
|
||||
|
||||
func badAuthorizationHandler(w http.ResponseWriter, r *http.Request) {
|
||||
if Config.basicAuth {
|
||||
rs := ""
|
||||
if Config.siteName != "" {
|
||||
rs = fmt.Sprintf(` realm="%s"`, Config.siteName)
|
||||
}
|
||||
w.Header().Set("WWW-Authenticate", `Basic`+rs)
|
||||
}
|
||||
http.Error(w, http.StatusText(http.StatusUnauthorized), http.StatusUnauthorized)
|
||||
}
|
||||
|
||||
func sliceContains(slice []string, s string) bool {
|
||||
for _, v := range slice {
|
||||
if s == v {
|
|
@ -1,4 +1,4 @@
|
|||
package apikeys
|
||||
package main
|
||||
|
||||
import (
|
||||
"testing"
|
||||
|
@ -10,15 +10,15 @@ func TestCheckAuth(t *testing.T) {
|
|||
"vFpNprT9wbHgwAubpvRxYCCpA2FQMAK6hFqPvAGrdZo=",
|
||||
}
|
||||
|
||||
if r, err := CheckAuth(authKeys, ""); err != nil && r {
|
||||
if r, err := checkAuth(authKeys, ""); err != nil && r {
|
||||
t.Fatal("Authorization passed for empty key")
|
||||
}
|
||||
|
||||
if r, err := CheckAuth(authKeys, "thisisnotvalid"); err != nil && r {
|
||||
if r, err := checkAuth(authKeys, "thisisnotvalid"); err != nil && r {
|
||||
t.Fatal("Authorization passed for invalid key")
|
||||
}
|
||||
|
||||
if r, err := CheckAuth(authKeys, "haPVipRnGJ0QovA9nyqK"); err != nil && !r {
|
||||
if r, err := checkAuth(authKeys, "haPVipRnGJ0QovA9nyqK"); err != nil && !r {
|
||||
t.Fatal("Authorization failed for valid key")
|
||||
}
|
||||
}
|
|
@ -4,7 +4,6 @@ import (
|
|||
"encoding/json"
|
||||
"io"
|
||||
"io/ioutil"
|
||||
"net/http"
|
||||
"os"
|
||||
"path"
|
||||
"time"
|
||||
|
@ -83,18 +82,6 @@ func (b LocalfsBackend) Get(key string) (metadata backends.Metadata, f io.ReadCl
|
|||
return
|
||||
}
|
||||
|
||||
func (b LocalfsBackend) ServeFile(key string, w http.ResponseWriter, r *http.Request) (err error) {
|
||||
_, err = b.Head(key)
|
||||
if err != nil {
|
||||
return
|
||||
}
|
||||
|
||||
filePath := path.Join(b.filesPath, key)
|
||||
http.ServeFile(w, r, filePath)
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
func (b LocalfsBackend) writeMetadata(key string, metadata backends.Metadata) error {
|
||||
metaPath := path.Join(b.metaPath, key)
|
||||
|
||||
|
|
|
@ -3,7 +3,6 @@ package s3
|
|||
import (
|
||||
"io"
|
||||
"io/ioutil"
|
||||
"net/http"
|
||||
"os"
|
||||
"strconv"
|
||||
"time"
|
||||
|
@ -80,43 +79,6 @@ func (b S3Backend) Get(key string) (metadata backends.Metadata, r io.ReadCloser,
|
|||
return
|
||||
}
|
||||
|
||||
func (b S3Backend) ServeFile(key string, w http.ResponseWriter, r *http.Request) (err error) {
|
||||
var result *s3.GetObjectOutput
|
||||
|
||||
if r.Header.Get("Range") != "" {
|
||||
result, err = b.svc.GetObject(&s3.GetObjectInput{
|
||||
Bucket: aws.String(b.bucket),
|
||||
Key: aws.String(key),
|
||||
Range: aws.String(r.Header.Get("Range")),
|
||||
})
|
||||
|
||||
w.WriteHeader(206)
|
||||
w.Header().Set("Content-Range", *result.ContentRange)
|
||||
w.Header().Set("Content-Length", strconv.FormatInt(*result.ContentLength, 10))
|
||||
w.Header().Set("Accept-Ranges", "bytes")
|
||||
|
||||
} else {
|
||||
result, err = b.svc.GetObject(&s3.GetObjectInput{
|
||||
Bucket: aws.String(b.bucket),
|
||||
Key: aws.String(key),
|
||||
})
|
||||
|
||||
}
|
||||
|
||||
if err != nil {
|
||||
if aerr, ok := err.(awserr.Error); ok {
|
||||
if aerr.Code() == s3.ErrCodeNoSuchKey || aerr.Code() == "NotFound" {
|
||||
err = backends.NotFoundErr
|
||||
}
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
_, err = io.Copy(w, result.Body)
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
func mapMetadata(m backends.Metadata) map[string]*string {
|
||||
return map[string]*string{
|
||||
"Expiry": aws.String(strconv.FormatInt(m.Expiry.Unix(), 10)),
|
||||
|
|
|
@ -3,7 +3,6 @@ package backends
|
|||
import (
|
||||
"errors"
|
||||
"io"
|
||||
"net/http"
|
||||
"time"
|
||||
)
|
||||
|
||||
|
@ -14,7 +13,6 @@ type StorageBackend interface {
|
|||
Get(key string) (Metadata, io.ReadCloser, error)
|
||||
Put(key string, r io.Reader, expiry time.Time, deleteKey, accessKey string) (Metadata, error)
|
||||
PutMetadata(key string, m Metadata) error
|
||||
ServeFile(key string, w http.ResponseWriter, r *http.Request) error
|
||||
Size(key string) (int64, error)
|
||||
}
|
||||
|
||||
|
|
|
@ -100,7 +100,7 @@ func fileDisplayHandler(c web.C, w http.ResponseWriter, r *http.Request, fileNam
|
|||
bytes, err := ioutil.ReadAll(reader)
|
||||
if err == nil {
|
||||
extra["extension"] = extension
|
||||
extra["lang_hl"] = extensionToHlLang(extension)
|
||||
extra["lang_hl"], extra["lang_ace"] = extensionToHlAndAceLangs(extension)
|
||||
extra["contents"] = string(bytes)
|
||||
tpl = Templates["display/bin.html"]
|
||||
}
|
||||
|
@ -122,7 +122,6 @@ func fileDisplayHandler(c web.C, w http.ResponseWriter, r *http.Request, fileNam
|
|||
"forcerandom": Config.forceRandomFilename,
|
||||
"lines": lines,
|
||||
"files": metadata.ArchiveFiles,
|
||||
"siteurl": strings.TrimSuffix(getSiteURL(r), "/"),
|
||||
}, r, w)
|
||||
|
||||
if err != nil {
|
||||
|
|
11
fileserve.go
11
fileserve.go
|
@ -2,6 +2,7 @@ package main
|
|||
|
||||
import (
|
||||
"fmt"
|
||||
"io"
|
||||
"net/http"
|
||||
"net/url"
|
||||
"strconv"
|
||||
|
@ -60,12 +61,16 @@ func fileServeHandler(c web.C, w http.ResponseWriter, r *http.Request) {
|
|||
}
|
||||
|
||||
if r.Method != "HEAD" {
|
||||
|
||||
storageBackend.ServeFile(fileName, w, r)
|
||||
_, reader, err := storageBackend.Get(fileName)
|
||||
if err != nil {
|
||||
oopsHandler(c, w, r, RespAUTO, err.Error())
|
||||
oopsHandler(c, w, r, RespAUTO, "Unable to open file.")
|
||||
return
|
||||
}
|
||||
defer reader.Close()
|
||||
|
||||
if _, err = io.CopyN(w, reader, metadata.Size); err != nil {
|
||||
oopsHandler(c, w, r, RespAUTO, err.Error())
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
|
2
go.mod
2
go.mod
|
@ -8,7 +8,6 @@ require (
|
|||
github.com/dchest/uniuri v0.0.0-20200228104902-7aecb25e1fe5
|
||||
github.com/dustin/go-humanize v1.0.0
|
||||
github.com/flosch/pongo2 v0.0.0-20190707114632-bbf5a6c351f4
|
||||
github.com/gabriel-vasile/mimetype v1.1.1
|
||||
github.com/microcosm-cc/bluemonday v1.0.2
|
||||
github.com/minio/sha256-simd v0.1.1
|
||||
github.com/russross/blackfriday v1.5.1
|
||||
|
@ -16,4 +15,5 @@ require (
|
|||
github.com/zeebo/bencode v1.0.0
|
||||
github.com/zenazn/goji v0.9.0
|
||||
golang.org/x/crypto v0.0.0-20200302210943-78000ba7a073
|
||||
gopkg.in/h2non/filetype.v1 v1.0.5
|
||||
)
|
||||
|
|
4
go.sum
4
go.sum
|
@ -15,8 +15,6 @@ github.com/dustin/go-humanize v1.0.0 h1:VSnTsYCnlFHaM2/igO1h6X3HA71jcobQuxemgkq4
|
|||
github.com/dustin/go-humanize v1.0.0/go.mod h1:HtrtbFcZ19U5GC7JDqmcUSB87Iq5E25KnS6fMYU6eOk=
|
||||
github.com/flosch/pongo2 v0.0.0-20190707114632-bbf5a6c351f4 h1:GY1+t5Dr9OKADM64SYnQjw/w99HMYvQ0A8/JoUkxVmc=
|
||||
github.com/flosch/pongo2 v0.0.0-20190707114632-bbf5a6c351f4/go.mod h1:T9YF2M40nIgbVgp3rreNmTged+9HrbNTIQf1PsaIiTA=
|
||||
github.com/gabriel-vasile/mimetype v1.1.1 h1:qbN9MPuRf3bstHu9zkI9jDWNfH//9+9kHxr9oRBBBOA=
|
||||
github.com/gabriel-vasile/mimetype v1.1.1/go.mod h1:6CDPel/o/3/s4+bp6kIbsWATq8pmgOisOPG40CJa6To=
|
||||
github.com/go-check/check v0.0.0-20180628173108-788fd7840127 h1:0gkP6mzaMqkmpcJYCFOLkIBwI7xFExG03bbkOkCvUPI=
|
||||
github.com/go-check/check v0.0.0-20180628173108-788fd7840127/go.mod h1:9ES+weclKsC9YodN5RgxqK/VD9HM9JsCSh7rNhMZE98=
|
||||
github.com/go-sql-driver/mysql v1.5.0/go.mod h1:DCzpHaOWr8IXmIStZouvnhqoel9Qv2LBy8hT2VhHyBg=
|
||||
|
@ -70,6 +68,8 @@ golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ=
|
|||
golang.org/x/tools v0.0.0-20181221001348-537d06c36207/go.mod h1:n7NCudcB/nEzxVGmLbDWY5pfWTLqBcC2KZ6jyYvM4mQ=
|
||||
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405 h1:yhCVgyC4o1eVCa2tZl7eS0r+SDo693bJlVdllGtEeKM=
|
||||
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0=
|
||||
gopkg.in/h2non/filetype.v1 v1.0.5 h1:CC1jjJjoEhNVbMhXYalmGBhOBK2V70Q1N850wt/98/Y=
|
||||
gopkg.in/h2non/filetype.v1 v1.0.5/go.mod h1:M0yem4rwSX5lLVrkEuRRp2/NinFMD5vgJ4DlAhZcfNo=
|
||||
gopkg.in/mgo.v2 v2.0.0-20180705113604-9856a29383ce h1:xcEWjVhvbDy+nHP67nPDDpbYrY+ILlfndk4bRioVHaU=
|
||||
gopkg.in/mgo.v2 v2.0.0-20180705113604-9856a29383ce/go.mod h1:yeKp02qBN3iKW1OzL3MGk2IdtZzaj7SFntXj72NppTA=
|
||||
gopkg.in/yaml.v2 v2.2.2 h1:ZCJp+EgiOT7lHqUV2J862kp8Qj64Jo6az82+3Td9dZw=
|
||||
|
|
|
@ -7,8 +7,8 @@ import (
|
|||
"unicode"
|
||||
|
||||
"github.com/andreimarcu/linx-server/backends"
|
||||
"github.com/gabriel-vasile/mimetype"
|
||||
"github.com/minio/sha256-simd"
|
||||
"gopkg.in/h2non/filetype.v1"
|
||||
)
|
||||
|
||||
func GenerateMetadata(r io.Reader) (m backends.Metadata, err error) {
|
||||
|
@ -21,7 +21,7 @@ func GenerateMetadata(r io.Reader) (m backends.Metadata, err error) {
|
|||
|
||||
// Get first 512 bytes for mimetype detection
|
||||
header := make([]byte, 512)
|
||||
headerlen, err := teeReader.Read(header)
|
||||
_, err = teeReader.Read(header)
|
||||
if err != nil {
|
||||
return
|
||||
}
|
||||
|
@ -47,8 +47,17 @@ func GenerateMetadata(r io.Reader) (m backends.Metadata, err error) {
|
|||
|
||||
// Use the bytes we extracted earlier and attempt to determine the file
|
||||
// type
|
||||
kind := mimetype.Detect(header[:headerlen])
|
||||
m.Mimetype = kind.String()
|
||||
kind, err := filetype.Match(header)
|
||||
if err != nil {
|
||||
m.Mimetype = "application/octet-stream"
|
||||
return m, err
|
||||
} else if kind.MIME.Value != "" {
|
||||
m.Mimetype = kind.MIME.Value
|
||||
} else if printable(header) {
|
||||
m.Mimetype = "text/plain"
|
||||
} else {
|
||||
m.Mimetype = "application/octet-stream"
|
||||
}
|
||||
|
||||
return
|
||||
}
|
||||
|
|
|
@ -1,10 +1,8 @@
|
|||
package helpers
|
||||
|
||||
import (
|
||||
"bytes"
|
||||
"strings"
|
||||
"testing"
|
||||
"unicode/utf16"
|
||||
)
|
||||
|
||||
func TestGenerateMetadata(t *testing.T) {
|
||||
|
@ -19,7 +17,7 @@ func TestGenerateMetadata(t *testing.T) {
|
|||
t.Fatalf("Sha256sum was %q instead of expected value of %q", m.Sha256sum, expectedSha256sum)
|
||||
}
|
||||
|
||||
expectedMimetype := "text/plain; charset=utf-8"
|
||||
expectedMimetype := "text/plain"
|
||||
if m.Mimetype != expectedMimetype {
|
||||
t.Fatalf("Mimetype was %q instead of expected value of %q", m.Mimetype, expectedMimetype)
|
||||
}
|
||||
|
@ -29,45 +27,3 @@ func TestGenerateMetadata(t *testing.T) {
|
|||
t.Fatalf("Size was %d instead of expected value of %d", m.Size, expectedSize)
|
||||
}
|
||||
}
|
||||
|
||||
func TestTextCharsets(t *testing.T) {
|
||||
// verify that different text encodings are detected and passed through
|
||||
orig := "This is a text string"
|
||||
utf16 := utf16.Encode([]rune(orig))
|
||||
utf16LE := make([]byte, len(utf16)*2+2)
|
||||
utf16BE := make([]byte, len(utf16)*2+2)
|
||||
utf8 := []byte(orig)
|
||||
utf16LE[0] = 0xff
|
||||
utf16LE[1] = 0xfe
|
||||
utf16BE[0] = 0xfe
|
||||
utf16BE[1] = 0xff
|
||||
for i := 0; i < len(utf16); i++ {
|
||||
lsb := utf16[i] & 0xff
|
||||
msb := utf16[i] >> 8
|
||||
utf16LE[i*2+2] = byte(lsb)
|
||||
utf16LE[i*2+3] = byte(msb)
|
||||
utf16BE[i*2+2] = byte(msb)
|
||||
utf16BE[i*2+3] = byte(lsb)
|
||||
}
|
||||
|
||||
testcases := []struct {
|
||||
data []byte
|
||||
extension string
|
||||
mimetype string
|
||||
}{
|
||||
{mimetype: "text/plain; charset=utf-8", data: utf8},
|
||||
{mimetype: "text/plain; charset=utf-16le", data: utf16LE},
|
||||
{mimetype: "text/plain; charset=utf-16be", data: utf16BE},
|
||||
}
|
||||
|
||||
for i, testcase := range testcases {
|
||||
r := bytes.NewReader(testcase.data)
|
||||
m, err := GenerateMetadata(r)
|
||||
if err != nil {
|
||||
t.Fatalf("[%d] unexpected error return %v\n", i, err)
|
||||
}
|
||||
if m.Mimetype != testcase.mimetype {
|
||||
t.Errorf("[%d] Expected mimetype '%s', got mimetype '%s'\n", i, testcase.mimetype, m.Mimetype)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -1,19 +0,0 @@
|
|||
|
||||
linx-cleanup
|
||||
-------------------------
|
||||
When files expire, access is disabled immediately, but the files and metadata
|
||||
will persist on disk until someone attempts to access them.
|
||||
|
||||
If you'd like to automatically clean up files that have expired, you can use the included `linx-cleanup` utility. To run it automatically, use a cronjob or similar type
|
||||
of scheduled task.
|
||||
|
||||
You should be careful to ensure that only one instance of `linx-cleanup` runs at
|
||||
a time to avoid unexpected behavior. It does not implement any type of locking.
|
||||
|
||||
|
||||
|Option|Description
|
||||
|------|-----------
|
||||
| ```-filespath files/``` | Path to stored uploads (default is files/)
|
||||
| ```-nologs``` | (optionally) disable deletion logs in stdout
|
||||
| ```-metapath meta/``` | Path to stored information about uploads (default is meta/)
|
||||
|
|
@ -1,14 +1,26 @@
|
|||
package cleanup
|
||||
package main
|
||||
|
||||
import (
|
||||
"flag"
|
||||
"log"
|
||||
"time"
|
||||
|
||||
"github.com/andreimarcu/linx-server/backends/localfs"
|
||||
"github.com/andreimarcu/linx-server/expiry"
|
||||
)
|
||||
|
||||
func Cleanup(filesDir string, metaDir string, noLogs bool) {
|
||||
func main() {
|
||||
var filesDir string
|
||||
var metaDir string
|
||||
var noLogs bool
|
||||
|
||||
flag.StringVar(&filesDir, "filespath", "files/",
|
||||
"path to files directory")
|
||||
flag.StringVar(&metaDir, "metapath", "meta/",
|
||||
"path to metadata directory")
|
||||
flag.BoolVar(&noLogs, "nologs", false,
|
||||
"don't log deleted files")
|
||||
flag.Parse()
|
||||
|
||||
fileBackend := localfs.NewLocalfsBackend(metaDir, filesDir)
|
||||
|
||||
files, err := fileBackend.List()
|
||||
|
@ -32,11 +44,3 @@ func Cleanup(filesDir string, metaDir string, noLogs bool) {
|
|||
}
|
||||
}
|
||||
}
|
||||
|
||||
func PeriodicCleanup(minutes time.Duration, filesDir string, metaDir string, noLogs bool) {
|
||||
c := time.Tick(minutes)
|
||||
for range c {
|
||||
Cleanup(filesDir, metaDir, noLogs)
|
||||
}
|
||||
|
||||
}
|
|
@ -1,23 +0,0 @@
|
|||
package main
|
||||
|
||||
import (
|
||||
"flag"
|
||||
|
||||
"github.com/andreimarcu/linx-server/cleanup"
|
||||
)
|
||||
|
||||
func main() {
|
||||
var filesDir string
|
||||
var metaDir string
|
||||
var noLogs bool
|
||||
|
||||
flag.StringVar(&filesDir, "filespath", "files/",
|
||||
"path to files directory")
|
||||
flag.StringVar(&metaDir, "metapath", "meta/",
|
||||
"path to metadata directory")
|
||||
flag.BoolVar(&noLogs, "nologs", false,
|
||||
"don't log deleted files")
|
||||
flag.Parse()
|
||||
|
||||
cleanup.Cleanup(filesDir, metaDir, noLogs)
|
||||
}
|
|
@ -1,12 +0,0 @@
|
|||
[Unit]
|
||||
Description=Self-hosted file/code/media sharing (expired files cleanup)
|
||||
After=network.target
|
||||
|
||||
[Service]
|
||||
User=linx
|
||||
Group=linx
|
||||
ExecStart=/usr/bin/linx-cleanup
|
||||
WorkingDirectory=/srv/linx/
|
||||
|
||||
[Install]
|
||||
WantedBy=multi-user.target
|
|
@ -1,8 +0,0 @@
|
|||
[Unit]
|
||||
Description=Run linx-cleanup every hour
|
||||
|
||||
[Timer]
|
||||
OnUnitActiveSec=1h
|
||||
|
||||
[Install]
|
||||
WantedBy=timers.target
|
|
@ -1,12 +0,0 @@
|
|||
|
||||
bind = 127.0.0.1:8080
|
||||
sitename = myLinx
|
||||
siteurl = https://mylinx.example.org/
|
||||
selifpath = s
|
||||
maxsize = 4294967296
|
||||
maxexpiry = 86400
|
||||
allowhotlink = true
|
||||
remoteuploads = true
|
||||
nologs = true
|
||||
force-random-filename = false
|
||||
cleanup-every-minutes = 5
|
35
server.go
35
server.go
|
@ -16,11 +16,9 @@ import (
|
|||
"time"
|
||||
|
||||
rice "github.com/GeertJohan/go.rice"
|
||||
"github.com/andreimarcu/linx-server/auth/apikeys"
|
||||
"github.com/andreimarcu/linx-server/backends"
|
||||
"github.com/andreimarcu/linx-server/backends/localfs"
|
||||
"github.com/andreimarcu/linx-server/backends/s3"
|
||||
"github.com/andreimarcu/linx-server/cleanup"
|
||||
"github.com/flosch/pongo2"
|
||||
"github.com/vharitonsky/iniflags"
|
||||
"github.com/zenazn/goji/graceful"
|
||||
|
@ -73,7 +71,6 @@ var Config struct {
|
|||
forceRandomFilename bool
|
||||
accessKeyCookieExpiry uint64
|
||||
customPagesDir string
|
||||
cleanupEveryMinutes uint64
|
||||
}
|
||||
|
||||
var Templates = make(map[string]*pongo2.Template)
|
||||
|
@ -111,12 +108,9 @@ func setup() *web.Mux {
|
|||
mux.Use(AddHeaders(Config.addHeaders))
|
||||
|
||||
if Config.authFile != "" {
|
||||
mux.Use(apikeys.NewApiKeysMiddleware(apikeys.AuthOptions{
|
||||
mux.Use(UploadAuth(AuthOptions{
|
||||
AuthFile: Config.authFile,
|
||||
UnauthMethods: []string{"GET", "HEAD", "OPTIONS", "TRACE"},
|
||||
BasicAuth: Config.basicAuth,
|
||||
SiteName: Config.siteName,
|
||||
SitePath: Config.sitePath,
|
||||
}))
|
||||
}
|
||||
|
||||
|
@ -156,10 +150,6 @@ func setup() *web.Mux {
|
|||
storageBackend = s3.NewS3Backend(Config.s3Bucket, Config.s3Region, Config.s3Endpoint, Config.s3ForcePathStyle)
|
||||
} else {
|
||||
storageBackend = localfs.NewLocalfsBackend(Config.metaDir, Config.filesDir)
|
||||
if Config.cleanupEveryMinutes > 0 {
|
||||
go cleanup.PeriodicCleanup(time.Duration(Config.cleanupEveryMinutes)*time.Minute, Config.filesDir, Config.metaDir, Config.noLogs)
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
// Template setup
|
||||
|
@ -200,10 +190,29 @@ func setup() *web.Mux {
|
|||
mux.Get(Config.sitePath+"upload/", uploadRemote)
|
||||
|
||||
if Config.remoteAuthFile != "" {
|
||||
remoteAuthKeys = apikeys.ReadAuthKeys(Config.remoteAuthFile)
|
||||
remoteAuthKeys = readAuthKeys(Config.remoteAuthFile)
|
||||
}
|
||||
}
|
||||
|
||||
if Config.basicAuth {
|
||||
options := AuthOptions{
|
||||
AuthFile: Config.authFile,
|
||||
UnauthMethods: []string{},
|
||||
}
|
||||
okFunc := func(w http.ResponseWriter, r *http.Request) {
|
||||
w.Header().Set("Location", Config.sitePath)
|
||||
w.WriteHeader(http.StatusFound)
|
||||
}
|
||||
authHandler := auth{
|
||||
successHandler: http.HandlerFunc(okFunc),
|
||||
failureHandler: http.HandlerFunc(badAuthorizationHandler),
|
||||
authKeys: readAuthKeys(Config.authFile),
|
||||
o: options,
|
||||
}
|
||||
mux.Head(Config.sitePath+"auth", authHandler)
|
||||
mux.Get(Config.sitePath+"auth", authHandler)
|
||||
}
|
||||
|
||||
mux.Post(Config.sitePath+"upload", uploadPostHandler)
|
||||
mux.Post(Config.sitePath+"upload/", uploadPostHandler)
|
||||
mux.Put(Config.sitePath+"upload", uploadPutHandler)
|
||||
|
@ -302,8 +311,6 @@ func main() {
|
|||
flag.Uint64Var(&Config.accessKeyCookieExpiry, "access-cookie-expiry", 0, "Expiration time for access key cookies in seconds (set 0 to use session cookies)")
|
||||
flag.StringVar(&Config.customPagesDir, "custompagespath", "",
|
||||
"path to directory containing .md files to render as custom pages")
|
||||
flag.Uint64Var(&Config.cleanupEveryMinutes, "cleanup-every-minutes", 0,
|
||||
"How often to clean up expired files in minutes (default is 0, which means files will be cleaned up as they are accessed)")
|
||||
|
||||
iniflags.Parse()
|
||||
|
||||
|
|
|
@ -2,71 +2,102 @@
|
|||
|
||||
/* Tomorrow Comment */
|
||||
.hljs-comment,
|
||||
.hljs-quote {
|
||||
.hljs-title {
|
||||
color: #8e908c;
|
||||
}
|
||||
|
||||
/* Tomorrow Red */
|
||||
.hljs-variable,
|
||||
.hljs-template-variable,
|
||||
.hljs-attribute,
|
||||
.hljs-tag,
|
||||
.hljs-name,
|
||||
.hljs-selector-id,
|
||||
.hljs-selector-class,
|
||||
.hljs-regexp,
|
||||
.hljs-deletion {
|
||||
.ruby .hljs-constant,
|
||||
.xml .hljs-tag .hljs-title,
|
||||
.xml .hljs-pi,
|
||||
.xml .hljs-doctype,
|
||||
.html .hljs-doctype,
|
||||
.css .hljs-id,
|
||||
.css .hljs-class,
|
||||
.css .hljs-pseudo {
|
||||
color: #c82829;
|
||||
}
|
||||
|
||||
/* Tomorrow Orange */
|
||||
.hljs-number,
|
||||
.hljs-preprocessor,
|
||||
.hljs-pragma,
|
||||
.hljs-built_in,
|
||||
.hljs-builtin-name,
|
||||
.hljs-literal,
|
||||
.hljs-type,
|
||||
.hljs-params,
|
||||
.hljs-meta,
|
||||
.hljs-link {
|
||||
.hljs-constant {
|
||||
color: #f5871f;
|
||||
}
|
||||
|
||||
/* Tomorrow Yellow */
|
||||
.hljs-attribute {
|
||||
.ruby .hljs-class .hljs-title,
|
||||
.css .hljs-rules .hljs-attribute {
|
||||
color: #eab700;
|
||||
}
|
||||
|
||||
/* Tomorrow Green */
|
||||
.hljs-string,
|
||||
.hljs-symbol,
|
||||
.hljs-bullet,
|
||||
.hljs-addition {
|
||||
.hljs-value,
|
||||
.hljs-inheritance,
|
||||
.hljs-header,
|
||||
.ruby .hljs-symbol,
|
||||
.xml .hljs-cdata {
|
||||
color: #718c00;
|
||||
}
|
||||
|
||||
/* Tomorrow Aqua */
|
||||
.css .hljs-hexcolor {
|
||||
color: #3e999f;
|
||||
}
|
||||
|
||||
/* Tomorrow Blue */
|
||||
.hljs-title,
|
||||
.hljs-section {
|
||||
.hljs-function,
|
||||
.python .hljs-decorator,
|
||||
.python .hljs-title,
|
||||
.ruby .hljs-function .hljs-title,
|
||||
.ruby .hljs-title .hljs-keyword,
|
||||
.perl .hljs-sub,
|
||||
.javascript .hljs-title,
|
||||
.coffeescript .hljs-title {
|
||||
color: #4271ae;
|
||||
}
|
||||
|
||||
/* Tomorrow Purple */
|
||||
.hljs-keyword,
|
||||
.hljs-selector-tag {
|
||||
.javascript .hljs-function {
|
||||
color: #8959a8;
|
||||
}
|
||||
|
||||
.hljs {
|
||||
display: block;
|
||||
overflow-x: auto;
|
||||
background: white;
|
||||
color: #4d4d4c;
|
||||
padding: 0.5em;
|
||||
}
|
||||
|
||||
.hljs-emphasis {
|
||||
font-style: italic;
|
||||
.hljs-line-numbers {
|
||||
text-align: right;
|
||||
border-right: 1px solid #ccc;
|
||||
margin-right: 5px;
|
||||
color: #999;
|
||||
-webkit-touch-callout: none;
|
||||
-webkit-user-select: none;
|
||||
-khtml-user-select: none;
|
||||
-moz-user-select: none;
|
||||
-ms-user-select: none;
|
||||
user-select: none;
|
||||
}
|
||||
|
||||
.hljs-strong {
|
||||
font-weight: bold;
|
||||
}
|
||||
.coffeescript .javascript,
|
||||
.javascript .xml,
|
||||
.tex .hljs-formula,
|
||||
.xml .javascript,
|
||||
.xml .vbscript,
|
||||
.xml .css,
|
||||
.xml .hljs-cdata {
|
||||
opacity: 0.5;
|
||||
}
|
File diff suppressed because one or more lines are too long
|
@ -4,7 +4,7 @@
|
|||
<head>
|
||||
<title>{% block title %}{{ sitename }}{% endblock %}</title>
|
||||
<meta charset='utf-8' content='text/html' http-equiv='content-type'>
|
||||
<meta name='viewport' content='width=device-width, initial-scale=0.8'>
|
||||
<meta name='viewport' content='width=device-width, initial-scale=1.0'>
|
||||
<link href='{{ sitepath }}static/css/linx.css?v=1' media='screen, projection' rel='stylesheet' type='text/css'>
|
||||
<link href='{{ sitepath }}static/css/hint.css' rel='stylesheet' type='text/css'>
|
||||
<link href='{{ sitepath }}static/images/favicon.gif' rel='icon' type='image/gif'>
|
||||
|
@ -31,7 +31,7 @@
|
|||
{% block content %}{% endblock %}
|
||||
|
||||
<div id="footer">
|
||||
<a href="https://github.com/andreimarcu/linx-server">linx-server</a>
|
||||
<a href="https://github.com/andreimarcu/linx-server">linx</a>
|
||||
</div>
|
||||
|
||||
</div>
|
||||
|
|
|
@ -1,12 +1,9 @@
|
|||
{% extends "base.html" %}
|
||||
|
||||
{% block head %}
|
||||
<meta property="og:audio" content="{{ siteurl }}{{ sitepath }}{{ selifpath }}{{ filename }}" />
|
||||
{% endblock %}
|
||||
|
||||
{% block main %}
|
||||
<audio class="display-audio" controls preload='auto'>
|
||||
<source src='{{ sitepath }}{{ selifpath }}{{ filename }}'>
|
||||
<a href='{{ sitepath }}{{ selifpath }}{{ filename }}'>Download it instead</a>
|
||||
</audio>
|
||||
{% endblock %}
|
||||
{% endblock %}
|
||||
|
||||
|
|
|
@ -1,11 +1,7 @@
|
|||
{% extends "base.html" %}
|
||||
|
||||
{% block head %}
|
||||
<meta property="og:image" content="{{ siteurl }}{{ sitepath }}{{ selifpath }}{{ filename }}" />
|
||||
{% endblock %}
|
||||
|
||||
{% block main %}
|
||||
<a href="{{ sitepath }}{{ selifpath }}{{ filename }}">
|
||||
<img class="display-image" src="{{ sitepath }}{{ selifpath }}{{ filename }}" />
|
||||
</a>
|
||||
{% endblock %}
|
||||
{% endblock %}
|
||||
|
|
|
@ -1,12 +1,8 @@
|
|||
{% extends "base.html" %}
|
||||
|
||||
{% block head %}
|
||||
<meta property="og:video" content="{{ siteurl }}{{ sitepath }}{{ selifpath }}{{ filename }}" />
|
||||
{% endblock %}
|
||||
|
||||
{% block main %}
|
||||
<video class="display-video" controls autoplay>
|
||||
<source src="{{ sitepath }}{{ selifpath }}{{ filename }}" />
|
||||
<source src="{{ sitepath }}{{ selifpath }}{{ filename }}"/>
|
||||
<a href='{{ sitepath }}{{ selifpath }}{{ filename }}'>Download it instead</a>
|
||||
</video>
|
||||
{% endblock %}
|
||||
{% endblock %}
|
||||
|
|
27
upload.go
27
upload.go
|
@ -15,12 +15,11 @@ import (
|
|||
"strings"
|
||||
"time"
|
||||
|
||||
"github.com/andreimarcu/linx-server/auth/apikeys"
|
||||
"github.com/andreimarcu/linx-server/backends"
|
||||
"github.com/andreimarcu/linx-server/expiry"
|
||||
"github.com/dchest/uniuri"
|
||||
"github.com/gabriel-vasile/mimetype"
|
||||
"github.com/zenazn/goji/web"
|
||||
"gopkg.in/h2non/filetype.v1"
|
||||
)
|
||||
|
||||
var FileTooLargeError = errors.New("File too large.")
|
||||
|
@ -167,16 +166,13 @@ func uploadRemote(c web.C, w http.ResponseWriter, r *http.Request) {
|
|||
key = password
|
||||
}
|
||||
}
|
||||
result, err := apikeys.CheckAuth(remoteAuthKeys, key)
|
||||
result, err := checkAuth(remoteAuthKeys, key)
|
||||
if err != nil || !result {
|
||||
if Config.basicAuth {
|
||||
rs := ""
|
||||
if Config.siteName != "" {
|
||||
rs = fmt.Sprintf(` realm="%s"`, Config.siteName)
|
||||
}
|
||||
w.Header().Set("WWW-Authenticate", `Basic`+rs)
|
||||
badAuthorizationHandler(w, r)
|
||||
} else {
|
||||
unauthorizedHandler(c, w, r)
|
||||
}
|
||||
unauthorizedHandler(c, w, r)
|
||||
return
|
||||
}
|
||||
}
|
||||
|
@ -188,7 +184,6 @@ func uploadRemote(c web.C, w http.ResponseWriter, r *http.Request) {
|
|||
|
||||
upReq := UploadRequest{}
|
||||
grabUrl, _ := url.Parse(r.FormValue("url"))
|
||||
directURL := r.FormValue("direct_url") == "yes"
|
||||
|
||||
resp, err := http.Get(grabUrl.String())
|
||||
if err != nil {
|
||||
|
@ -220,11 +215,7 @@ func uploadRemote(c web.C, w http.ResponseWriter, r *http.Request) {
|
|||
return
|
||||
}
|
||||
|
||||
if directURL {
|
||||
http.Redirect(w, r, Config.sitePath+Config.selifPath+upload.Filename, 303)
|
||||
} else {
|
||||
http.Redirect(w, r, Config.sitePath+upload.Filename, 303)
|
||||
}
|
||||
http.Redirect(w, r, Config.sitePath+upload.Filename, 303)
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -267,11 +258,11 @@ func processUpload(upReq UploadRequest) (upload Upload, err error) {
|
|||
header = header[:n]
|
||||
|
||||
// Determine the type of file from header
|
||||
kind := mimetype.Detect(header)
|
||||
if len(kind.Extension()) < 2 {
|
||||
kind, err := filetype.Match(header)
|
||||
if err != nil || kind.Extension == "unknown" {
|
||||
extension = "file"
|
||||
} else {
|
||||
extension = kind.Extension()[1:] // remove leading "."
|
||||
extension = kind.Extension
|
||||
}
|
||||
}
|
||||
|
||||
|
|
145
util.go
145
util.go
|
@ -1,10 +1,15 @@
|
|||
package main
|
||||
|
||||
func extensionToHlLang(extension string) (hlExt string) {
|
||||
func extensionToHlAndAceLangs(extension string) (hlExt, aceExt string) {
|
||||
hlExt, exists := extensionToHl[extension]
|
||||
if !exists {
|
||||
hlExt = "text"
|
||||
}
|
||||
|
||||
aceExt, exists = extensionToAce[extension]
|
||||
if !exists {
|
||||
aceExt = "text"
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
|
@ -13,63 +18,83 @@ func supportedBinExtension(extension string) bool {
|
|||
return exists
|
||||
}
|
||||
|
||||
var extensionToHl = map[string]string{
|
||||
"ahk": "autohotkey",
|
||||
"apache": "apache",
|
||||
"applescript": "applescript",
|
||||
"bas": "basic",
|
||||
"bash": "sh",
|
||||
"bat": "dos",
|
||||
"c": "cpp",
|
||||
"cfc": "coldfusion",
|
||||
"clj": "clojure",
|
||||
"cmake": "cmake",
|
||||
"coffee": "coffee",
|
||||
"cpp": "c_cpp",
|
||||
"cs": "csharp",
|
||||
"css": "css",
|
||||
"d": "d",
|
||||
"dart": "dart",
|
||||
"diff": "diff",
|
||||
"dockerfile": "dockerfile",
|
||||
"elm": "elm",
|
||||
"erl": "erlang",
|
||||
"for": "fortran",
|
||||
"go": "go",
|
||||
"h": "cpp",
|
||||
"htm": "html",
|
||||
"html": "html",
|
||||
"ini": "ini",
|
||||
"java": "java",
|
||||
"js": "javascript",
|
||||
"json": "json",
|
||||
"jsp": "jsp",
|
||||
"kt": "kotlin",
|
||||
"less": "less",
|
||||
"lisp": "lisp",
|
||||
"lua": "lua",
|
||||
"m": "objectivec",
|
||||
"nginx": "nginx",
|
||||
"ocaml": "ocaml",
|
||||
"php": "php",
|
||||
"pl": "perl",
|
||||
"proto": "protobuf",
|
||||
"ps": "powershell",
|
||||
"py": "python",
|
||||
"rb": "ruby",
|
||||
"rs": "rust",
|
||||
"scala": "scala",
|
||||
"scm": "scheme",
|
||||
"scpt": "applescript",
|
||||
"scss": "scss",
|
||||
"sh": "sh",
|
||||
"sql": "sql",
|
||||
"tcl": "tcl",
|
||||
"tex": "latex",
|
||||
"toml": "ini",
|
||||
"ts": "typescript",
|
||||
"txt": "text",
|
||||
"xml": "xml",
|
||||
"yaml": "yaml",
|
||||
"yml": "yaml",
|
||||
var extensionToAce = map[string]string{
|
||||
"c": "c_cpp",
|
||||
"h": "c_cpp",
|
||||
"cpp": "c_cpp",
|
||||
"clj": "clojure",
|
||||
"coffee": "coffee",
|
||||
"cfc": "coldfusion",
|
||||
"cs": "csharp",
|
||||
"sh": "sh",
|
||||
"bash": "sh",
|
||||
"css": "css",
|
||||
"go": "golang",
|
||||
"diff": "diff",
|
||||
"html": "html",
|
||||
"xml": "xml",
|
||||
"ini": "ini",
|
||||
"java": "java",
|
||||
"js": "javascript",
|
||||
"json": "json",
|
||||
"jsp": "jsp",
|
||||
"tex": "latex",
|
||||
"lisp": "lisp",
|
||||
"less": "less",
|
||||
"lua": "lua",
|
||||
"md": "markdown",
|
||||
"ocaml": "ocaml",
|
||||
"tcl": "tcl",
|
||||
"yaml": "yaml",
|
||||
"php": "php",
|
||||
"pl": "perl",
|
||||
"py": "python",
|
||||
"rb": "ruby",
|
||||
"sql": "sql",
|
||||
"apache": "apache",
|
||||
"cmake": "cmake",
|
||||
"bat": "dos",
|
||||
"scala": "scala",
|
||||
"txt": "text",
|
||||
}
|
||||
|
||||
var extensionToHl = map[string]string{
|
||||
"c": "cpp",
|
||||
"h": "cpp",
|
||||
"cpp": "c_cpp",
|
||||
"clj": "clojure",
|
||||
"coffee": "coffee",
|
||||
"cfc": "coldfusion",
|
||||
"cs": "csharp",
|
||||
"sh": "sh",
|
||||
"bash": "sh",
|
||||
"css": "css",
|
||||
"go": "go",
|
||||
"diff": "diff",
|
||||
"html": "html",
|
||||
"htm": "html",
|
||||
"ini": "ini",
|
||||
"java": "java",
|
||||
"js": "javascript",
|
||||
"json": "json",
|
||||
"jsp": "jsp",
|
||||
"tex": "latex",
|
||||
"lisp": "lisp",
|
||||
"less": "less",
|
||||
"lua": "lua",
|
||||
"ocaml": "ocaml",
|
||||
"tcl": "tcl",
|
||||
"nginx": "nginx",
|
||||
"xml": "xml",
|
||||
"yaml": "yaml",
|
||||
"php": "php",
|
||||
"pl": "perl",
|
||||
"py": "python",
|
||||
"rb": "ruby",
|
||||
"sql": "sql",
|
||||
"apache": "apache",
|
||||
"cmake": "cmake",
|
||||
"bat": "dos",
|
||||
"scala": "scala",
|
||||
"txt": "text",
|
||||
}
|
||||
|
|
Loading…
Add table
Reference in a new issue