dynamic-localpv-provisioner
cstor-operators
Our great sponsors
dynamic-localpv-provisioner | cstor-operators | |
---|---|---|
3 | 4 | |
126 | 93 | |
9.5% | - | |
5.9 | 5.1 | |
9 days ago | 4 months ago | |
Go | Go | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
dynamic-localpv-provisioner
-
Using Kaniko to Build and Publish container image with Github action on Github Self-hosted Runners
# gha-runner-scale-set-value.yml githubConfigUrl: "https://github.com/myorg/myrepo" githubConfigSecret: github_token: "my-PAT" ## maxRunners is the max number of runners the autoscaling runner set will scale up to. maxRunners: 5 ## minRunners is the min number of idle runners. The target number of runners created will be ## calculated as a sum of minRunners and the number of jobs assigned to the scale set. minRunners: 1 containerMode: type: "kubernetes" ## type can be set to dind or kubernetes ## the following is required when containerMode.type=kubernetes kubernetesModeWorkVolumeClaim: accessModes: ["ReadWriteOnce"] # For local testing, use https://github.com/openebs/dynamic-localpv-provisioner/blob/develop/docs/quickstart.md to provide dynamic provision volume with storageClassName: openebs-hostpath storageClassName: "managed-csi" resources: requests: storage: 2Gi template: spec: securityContext: fsGroup: 123 ## needed to resolve permission issues with mounted volume. https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners-with-actions-runner-controller/troubleshooting-actions-runner-controller-errors#error-access-to-the-path-homerunner_work_tool-is-denied containers: - name: runner image: ghcr.io/actions/actions-runner:latest command: ["/home/runner/run.sh"] env: - name: ACTIONS_RUNNER_REQUIRE_JOB_CONTAINER value: "false" ## To allow jobs without a job container to run, set ACTIONS_RUNNER_REQUIRE_JOB_CONTAINER to false on your runner container. This instructs the runner to disable this check. volumes: - name: work ephemeral: volumeClaimTemplate: spec: accessModes: [ "ReadWriteOnce" ] storageClassName: "managed-csi" resources: requests: storage: 2Gi
-
A local maximum on bare metal k8s storage? OpenEBS ZFS LocalPV + Rancher Longhorn
Ahhhh So your'e using openebs localpv (openebs/dynamic-localpv-provisioner?). Yeah I couldn't go with that since it didn't properly limit space, but if you're running the DBs then you have controls of that stuff at a higher level (and it's probably good to not be too strict to not hurt user workloads). XFS supposedly does the limiting if you set up the underlying storage properly now but I couldn't get it to work.
-
Why OpenEBS 3.0 for Kubernetes and Storage?
OpenEBS Hostpath LocalPV (declared stable), the first and the most widely used LocalPV now supports enforcing XFS quotas and the ability to use a custom node label for node affinity (instead of the default 'kubernetes.io/hostname').
cstor-operators
-
Why OpenEBS 3.0 for Kubernetes and Storage?
OpenEBS CStor (declared stable), has added support for a CSI Driver and also improved customer resources and operators for managing the lifecycle of CStor Pools. This 3.0 version of the CStor includes:
-
32 Node KBN-I/5200 Kubernetes cluster running Debian 10/K8s 1.17/Intel Core i7/8GB RAM each. Used the standoffs and mounted each 8 stack to empty trays. Disks are 64GB USB3 sticks.
It's a little complicated at first but there's also OpenEBS and Longhorn. Longhorn is probably the most easiest to get going with, but I chose rook-ceph because it's very stable.
-
Looking for a sanity check on a project I'm working on at home, hoping you fine people can help - Raspberry Pi Kubernetes Cluster
- SMB CSI: https://github.com/kubernetes-csi/csi-driver-smb - OpenEBS if you got the hardware for it: https://openebs.io/
-
Best Open-Source Distributed Parallel Storage Option for an AI/ML Cluster?
Tried OpenEBS? These two have replication HA features. https://github.com/openebs/Mayastor https://github.com/openebs/cstor-operators
What are some alternatives?
dynamic-nfs-provisioner - Operator for dynamically provisioning an NFS server on any Kubernetes Persistent Volume. Also creates an NFS volume on the dynamically provisioned server for enabling Kubernetes RWX volumes.
Mayastor - Dynamically provision Stateful Persistent Replicated Cluster-wide Fabric Volumes & Filesystems for Kubernetes that is provisioned from an optimized NVME SPDK backend data storage stack.
jiva-operator - Kubernetes Operator for managing Jiva Volumes via custom resource.
openebs - Most popular & widely deployed Open Source Container Native Storage platform for Stateful Persistent Applications on Kubernetes.
lvm-localpv - Dynamically provision Stateful Persistent Node-Local Volumes & Filesystems for Kubernetes that is integrated with a backend LVM2 data storage stack.
longhorn - Cloud-Native distributed storage built on and for Kubernetes
rawfile-localpv - Dynamically deploy Stateful Persistent Node-Local Volumes & Filesystems for Kubernetes that is provisioned from RAW-device file loop mounted Local-Hostpath storage.
zfs-localpv - Dynamically provision Stateful Persistent Node-Local Volumes & Filesystems for Kubernetes that is integrated with a backend ZFS data storage stack.
device-localpv - CSI Driver for using Local Block Devices