k8s-csi-s3/README.md

176 lines
5.2 KiB
Markdown
Raw Normal View History

2018-07-14 08:52:55 +00:00
# CSI for S3
2018-07-14 08:52:55 +00:00
This is a Container Storage Interface ([CSI](https://github.com/container-storage-interface/spec/blob/master/spec.md)) for S3 (or S3 compatible) storage. This can dynamically allocate buckets and mount them via a fuse mount into any container.
2018-07-14 08:48:22 +00:00
## Status
2018-07-22 10:59:30 +00:00
This is still very experimental and should not be used in any production environment. Unexpected data loss could occur depending on what mounter and S3 storage backend is being used.
## Kubernetes installation
### Requirements
* Kubernetes 1.13+ (CSI v1.0.0 compatibility)
2018-07-14 08:48:22 +00:00
* Kubernetes has to allow privileged containers
* Docker daemon must allow shared mounts (systemd flag `MountFlags=shared`)
### 1. Create a secret with your S3 credentials
2018-07-14 08:48:22 +00:00
```yaml
apiVersion: v1
kind: Secret
metadata:
name: csi-s3-secret
# Namespace depends on the configuration in the storageclass.yaml
namespace: kube-system
2018-07-14 08:48:22 +00:00
stringData:
accessKeyID: <YOUR_ACCESS_KEY_ID>
secretAccessKey: <YOUR_SECRET_ACCES_KEY>
# For AWS set it to "https://s3.<region>.amazonaws.com"
2018-08-06 17:27:24 +00:00
endpoint: <S3_ENDPOINT_URL>
# If not on S3, set it to ""
2018-07-14 08:48:22 +00:00
region: <S3_REGION>
```
The region can be empty if you are using some other S3 compatible storage.
### 2. Deploy the driver
2018-07-14 08:48:22 +00:00
```bash
cd deploy/kubernetes
kubectl create -f provisioner.yaml
kubectl create -f attacher.yaml
kubectl create -f csi-s3.yaml
2018-07-14 08:48:22 +00:00
```
### 3. Create the storage class
2018-07-14 08:48:22 +00:00
```bash
kubectl create -f storageclass.yaml
2018-07-14 08:48:22 +00:00
```
### 4. Test the S3 driver
2019-05-15 19:14:18 +00:00
1. Create a pvc using the new storage class:
2018-07-14 08:48:22 +00:00
```bash
kubectl create -f pvc.yaml
2018-07-14 08:48:22 +00:00
```
2019-05-15 19:14:18 +00:00
2. Check if the PVC has been bound:
2018-07-14 08:48:22 +00:00
```bash
$ kubectl get pvc csi-s3-pvc
NAME STATUS VOLUME CAPACITY ACCESS MODES STORAGECLASS AGE
csi-s3-pvc Bound pvc-c5d4634f-8507-11e8-9f33-0e243832354b 5Gi RWO csi-s3 9s
2018-07-14 08:48:22 +00:00
```
2019-05-15 19:14:18 +00:00
3. Create a test pod which mounts your volume:
2018-07-14 08:48:22 +00:00
```bash
kubectl create -f pod.yaml
2018-07-14 08:48:22 +00:00
```
2018-07-14 08:48:22 +00:00
If the pod can start, everything should be working.
2019-05-15 19:14:18 +00:00
4. Test the mount
2018-07-14 08:48:22 +00:00
```bash
$ kubectl exec -ti csi-s3-test-nginx bash
$ mount | grep fuse
s3fs on /var/lib/www/html type fuse.s3fs (rw,nosuid,nodev,relatime,user_id=0,group_id=0,allow_other)
$ touch /var/lib/www/html/hello_world
```
2018-07-14 08:48:22 +00:00
If something does not work as expected, check the troubleshooting section below.
## Additional configuration
### Mounter
2018-08-06 17:27:24 +00:00
As S3 is not a real file system there are some limitations to consider here. Depending on what mounter you are using, you will have different levels of POSIX compability. Also depending on what S3 storage backend you are using there are not always [consistency guarantees](https://github.com/gaul/are-we-consistent-yet#observed-consistency).
2018-07-29 12:54:28 +00:00
The driver can be configured to use one of these mounters to mount buckets:
2018-07-14 08:48:22 +00:00
2019-03-07 19:27:02 +00:00
* [rclone](https://rclone.org/commands/rclone_mount)
2018-07-22 10:59:30 +00:00
* [s3fs](https://github.com/s3fs-fuse/s3fs-fuse)
* [goofys](https://github.com/kahing/goofys)
2018-07-27 11:34:24 +00:00
* [s3backer](https://github.com/archiecobbs/s3backer)
2018-07-22 10:59:30 +00:00
2018-07-29 12:54:28 +00:00
The mounter can be set as a parameter in the storage class. You can also create multiple storage classes for each mounter if you like.
2018-07-22 10:59:30 +00:00
All mounters have different strengths and weaknesses depending on your use case. Here are some characteristics which should help you choose a mounter:
#### rclone
2019-03-07 19:27:02 +00:00
* Almost full POSIX compatibility (depends on caching mode)
* Files can be viewed normally with any S3 client
#### s3fs
2018-07-22 10:59:30 +00:00
* Large subset of POSIX
* Files can be viewed normally with any S3 client
* Does not support appends or random writes
#### goofys
2018-07-22 10:59:30 +00:00
* Weak POSIX compatibility
* Performance first
* Files can be viewed normally with any S3 client
* Does not support appends or random writes
#### s3backer (experimental*)
2018-07-27 11:34:24 +00:00
* Represents a block device stored on S3
* Allows to use a real filesystem
* Files are not readable with other S3 clients
* Support appends
* Supports compression before upload (Not yet implemented in this driver)
* Supports encryption before upload (Not yet implemented in this driver)
*s3backer is experimental at this point because volume corruption can occur pretty quickly in case of an unexpected shutdown of a Kubernetes node or CSI pod.
The s3backer binary is not bundled with the normal docker image to keep that as small as possible. Use the `<version>-full` image tag for testing s3backer.
2018-08-06 17:27:24 +00:00
Fore more detailed limitations consult the documentation of the different projects.
2018-07-14 08:48:22 +00:00
## Troubleshooting
### Issues while creating PVC
2019-05-15 19:14:18 +00:00
Check the logs of the provisioner:
```bash
kubectl logs -l app=csi-provisioner-s3 -c csi-s3
2018-07-14 08:48:22 +00:00
```
### Issues creating containers
2019-05-15 19:14:18 +00:00
1. Ensure feature gate `MountPropagation` is not set to `false`
2. Check the logs of the s3-driver:
```bash
kubectl logs -l app=csi-s3 -c csi-s3
2018-07-14 08:48:22 +00:00
```
## Development
2018-07-14 08:48:22 +00:00
This project can be built like any other go application.
2018-07-14 08:48:22 +00:00
```bash
go get -u github.com/ctrox/csi-s3
2018-07-14 08:48:22 +00:00
```
### Build executable
2018-07-14 08:48:22 +00:00
```bash
make build
2018-07-14 08:48:22 +00:00
```
### Tests
2018-07-14 08:48:22 +00:00
Currently the driver is tested by the [CSI Sanity Tester](https://github.com/kubernetes-csi/csi-test/tree/master/pkg/sanity). As end-to-end tests require S3 storage and a mounter like s3fs, this is best done in a docker container. A Dockerfile and the test script are in the `test` directory. The easiest way to run the tests is to just use the make command:
2018-07-14 08:48:22 +00:00
```bash
make test
2018-07-14 08:48:22 +00:00
```