Skip to content

Latest commit

 

History

History
206 lines (137 loc) · 7.66 KB

readme.org

File metadata and controls

206 lines (137 loc) · 7.66 KB

Caddy Http Cache

This is a http cache plugin for caddy 2. The difference between this and cache-handler https://github.com/caddyserver/cache-handler is that this one is much easier to understand and configure. But it does not support a distrobuted ache, and it needs to be compiled with golang 1.20 and caddy version v2.6.4

In fact, I am not so familiar with the part of cdn cache’s mechanism so I reference the code from https://github.com/nicolasazrak/caddy-cache. which is the caddy v1’s cache module. Then, I migrate the codebase to be consist with the caddy2 architecture and develop more features based on that.

Currently, this doesn’t support distributed cache yet. It’s still in plan.

Features

Multi storage backends support

Now the following backends are supported.

  • file
  • inmemory
  • redis

In the latter part, I will show the example Caddyfile to serve different type of proxy cache server.

Conditional rule to cache the upstream response

  • uri path matcher
  • http header matcher

Set default cache’s max age

A default age for matched responses that do not have an explicit expiration.

Purge cache

There are exposed endpoints to purge cache. I implement it with admin apis endpoints. The following will show you how to purge cache

first you can list the current caches

Given that you specify the port 7777 to serve caddy admin, you can get the list of cache by the api below.

GET http://example.com:7777/caches

purge the cache with http DELETE request

It supports the regular expression.

DELETE http://example.com:7777/caches/purge
Content-Type: application/json

{
  "method": "GET",
  "host": "localhost",
  "uri": ".*\\.txt"
}

Support cluster with consul

NOTE: still under development and only the memory backend supports.

I’ve provided a simple example to mimic a cluster environment.

PROJECT_PATH=/app docker-compose --project-directory=./ -f example/distributed_cache/docker-compose.yaml up

How to build

In development, go to the cmd folder and type the following commands.

go build -ldflags="-w -s" -o caddy

To build linux binary by this

GOOS=linux GOARCH=amd64 go build -ldflags="-w -s" -o caddy

Or you can use the xcaddy to build the executable file. Ensure you’ve install it by go get -u github.com/caddyserver/xcaddy/cmd/xcaddy

xcaddy build v2.6.4 --with github.com/sillygod/cdp-cache

Xcaddy also provide a way to develop plugin locally.

xcaddy run --config cmd/Caddyfile

To remove unused dependencies

go mod tidy

How to run

You can directly run with the binary.

caddy run --config [caddy file] --adapter caddyfile

Or if you are preferred to use the docker

docker run -it -p80:80 -p443:443 -v [the path of caddyfile]:/app/Caddyfile docker.pkg.github.com/sillygod/cdp-cache/caddy:latest

Configuration

The following will list current support configs in the caddyfile.

status_header

The header to set cache status. default value: X-Cache-Status

match_path

Only the request’s path match the condition will be cached. Ex. / means all request need to be cached because all request’s path must start with /

match_methods

By default, only GET and POST methods are cached. If you would like to cache other methods as well you can configure here which methods should be cached, e.g.: GET HEAD POST.

To be able to distinguish different POST requests, it is advisable to include the body hash in the cache key, e.g.: {http.request.method} {http.request.host}{http.request.uri.path}?{http.request.uri.query} {http.request.contentlength} {http.request.bodyhash}

default_max_age

The cache’s expiration time.

stale_max_age

The duration that a cache entry is kept in the cache, even though it has already expired. The default duration is 0.

If this duration is > 0 and the upstream server answers with an HTTP status code >= 500 (server error) this plugin checks whether there is still an expired (stale) entry from a previous, successful call in the cache. In that case, this stale entry is used to answer instead of the 5xx response.

match_header

only the req’s header match the condtions ex.

match_header Content-Type image/jpg image/png “text/plain; charset=utf-8”

path

The position where to save the file. Only applied when the cache_type is file.

cache_key

The key of cache entry. The default value is {http.request.method} {http.request.host}{http.request.uri.path}?{http.request.uri.query}

cache_bucket_num

The bucket number of the mod of cache_key’s checksum. The default value is 256.

cache_type

Indicate to use which kind of cache’s storage backend. Currently, there are two choices. One is file and the other is in_memory

cache_max_memory_size

The max memory usage for in_memory backend.

distributed

Working in process. Currently, only support consul to establish the cluster of cache server node.

To see a example config, please refer this

service_name

specify your service to be registered in the consul agent.

addr

the address of the consul agent.

health_check

indicate the health_check endpoint which consul agent will use this endpoint to check the cache server is healthy

Example configs

You can go to the directory example. It shows you each type of cache’s configuration.

Benchmarks

Now, I just simply compares the performance between in-memory and disk.

Env

Caddy run with the config file under directory benchmark and tests were run on the mac book pro (1.4 GHz Intel Core i5, 16 GB 2133 MHz LPDDR3)

Test Result

The following benchmark is analysized by wrk -c 50 -d 30s --latency -t 4 http://localhost:9991/pg31674.txt without log open. Before running this, ensure you provision the tests data by bash benchmark/provision.sh

req/slatency (50% 90% 99%)
proxy + file cache138533.29ms / 4.09ms / 5.26ms
proxy + in memory cache206222.20ms / 3.03ms / 4.68ms

Todo list