kronform
cluster-api-provider-hetzner
kronform | cluster-api-provider-hetzner | |
---|---|---|
2 | 28 | |
59 | 511 | |
- | 4.7% | |
9.1 | 9.5 | |
4 days ago | 4 days ago | |
Shell | Go | |
MIT License | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
kronform
-
Bare-Metal Kubernetes, Part I: Talos on Hetzner
https://datavirke.dk/posts/bare-metal-kubernetes-first-incid...
Source code repository (set up in Part III) for node configuration and deployed services is available at https://github.com/MathiasPius/kronform
While the documentation was initially intended more as a future reference for myself as well as a log of decisions made, and why I made them, I've received some really good feedback and ideas already, and figured it might be interesting to the hacker community :)
cluster-api-provider-hetzner
-
Bare-Metal Kubernetes, Part I: Talos on Hetzner
Hetzner Cloud is officially supported, but that means setting up VPSs in Hetzner's Cloud offering, whereas this project was intended as a more or less independent pure bare-metal cluster. I see they offer Bare Metal support as well, but I haven't dived too deep into it.
I haven't used KubeOne, but I have previously used Syself's https://github.com/syself/cluster-api-provider-hetzner which I believe works in a similar fashion. I think the approach is very interesting and plays right into the Kubernetes Operator playbook and its self-healing ambitions.
That being said, the complexity of the approach, probably in trying to span and resolve inconsistencies across such a wide landscape of providers, caused me quite a bit of grief. I eventually abandoned this approach after having some operator somewhere consistently attempt and fail to spin up a secondary control plane VPS against my wishes. After poring over loads of documentation and half a dozen CRDs in an attempt to resolve it, I threw in my hat.
Of course, Kubermatic is not Syself, and this was about a year ago, so it is entirely possible that both projects are absolutely superb solutions to the problem at this point.
-
Fly.io Postgres cluster went down for 3 days, no word from them about it
For anyone interested in Kubernetes on Hetzner, there's a really interesting CAPI provider being actively developed:
https://github.com/syself/cluster-api-provider-hetzner
- Syself: Cluster API Provider Hetzner released
- Cluster API Provider Hetzner released
-
How many of you are running kubernetes on prem?
Just a hint running ML Workloads on Hetzner is pretty cheap! You could use for managing k8s: https://github.com/syself/cluster-api-provider-hetzner
-
Syself cluster-api-provider Hetzner v1.0.0-beta.16
we (Syself) release Cluster-API Provider Hetzner v1.0.0-beta.16.
-
NEW ARM-BASED CLOUD SERVER
ah okay they come from the upstream cluster-api project. The caph project implements only the infrastructure provider part of Cluster API.
-
Has anyone set up autoscaling on hetzner?
you can easily use it with https://github.com/syself/cluster-api-provider-hetzner
- Image digest of Go 1.19.7 changed?
-
What's the most sane way to operate a K8s cluster?
I would use cluster-api-provider-hetzner from Syself.