hcloud-cloud-controller-manager
virtual-kubelet
hcloud-cloud-controller-manager | virtual-kubelet | |
---|---|---|
8 | 10 | |
646 | 4,085 | |
1.4% | 0.7% | |
8.7 | 6.7 | |
5 days ago | 3 days ago | |
Go | Go | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
hcloud-cloud-controller-manager
-
Bare-Metal Kubernetes, Part I: Talos on Hetzner
When it comes to load balancing, I think the hcloud-cloud-controller-manager[1] is probably your best bet, and although I haven't tested it, I'm sure it can be coerced into some kind of working configuration with the vSwitch/Cloud Network coupling, even none of cluster is actually Cloud-based.
I haven't used Sidero Omni yet, but if it's as well architected as Talos is, I'm sure it's an excellent solution. It still leaves open the question of ordering and provisioning the servers themselves. For simpler use-cases it wouldn't be too difficult to hack together a script to interact with the Hetzner Robot API to achieve this goal, but if I wanted any level of robustness, and if you'll excuse the shameless plug, I think I'd write a custom operator in Rust using my hrobot-rs[2] library :)
[1] https://github.com/hetznercloud/hcloud-cloud-controller-mana...
-
Ditching ingress-nginx for Cloudflare Tunnels
Did you actually use the Hetzner LBs or did you just send traffic to the NodePort?
- Creating an ARM server with Ansible
-
Can any Hetzner user, please explain there workflow on Hetzner?
It's not even close to major public cloud providers, but this is my setup:
* https://github.com/kube-hetzner/terraform-hcloud-kube-hetzne... (Terraform, Kubernetes bootstrap)
* Flux for CI
* nginx-ingress + Hetzner Loadbalancer (thanks to https://github.com/hetznercloud/hcloud-cloud-controller-mana...)
* Hetzner storage volumes (thanks to https://github.com/hetznercloud/csi-driver)
Kube-Hetzner supports Hetzner Cloud loadbalancers and volumes out of the box, though it also supports other components.
- Tell HN: I got 10x Hetzner storage at the same price
- Error Message: missing prefix hcloud:// even with --cloud-provider=external
-
best practice kubernetes node networking
I plan to do the same or have already implemented it. I use the Hetzner private networks with the hcloud-cloud-controller-manager https://github.com/hetznercloud/hcloud-cloud-controller-manager. This creates routes so that the nodes are also connected to each other in the private network. All kubernetes resources listen on the private network. Except for the Kubernetes API, which listens on the external IP of the master so I can use kubectl from home. The actual gateway into the cluster is controlled by a load balancer from Hetzner.
-
Unable to use hcloud-cloud-controller-manager LB as ingress controller. Health check fails
I want to use Hetzner's Load Balancer as a Nginx Ingress controller. I follow this suggestion to install Nginx Ingress via helm.
virtual-kubelet
-
Bare-Metal Kubernetes, Part I: Talos on Hetzner
Speaking of k8s, anyone here know of ready-made solutions for getting XCode (i.e. xcodebuild) running in pods? As far as I'm aware, there are no good solutions for getting XCode running on Linux, so at the moment I'm just futzing about with a virtual-kubelet[0] implementation that spawns MacOS VMs. This works just fine, but the problem seems like such an obvious one that I expect there to be some existing solution(s) I just missed.
[0]:https://github.com/virtual-kubelet/virtual-kubelet/
-
Keeping Airflow tasks “cloud-native”
Have you looked into virtual kubelet yet? It allows you to make a virtual node in your on-prem cluster that schedules workloads on services like AWS Fargate or Azure Container Instances.
-
Similar to AWS Fargate provider?
If you are serious about implementing this yourself, you may want to look into virtual kubelet: https://virtual-kubelet.io/
- Nomad vs. Kubernetes
-
Deploy on prem Kubernetes. What is the best approach paid and unpaid to deploy a cluster on premise with burst to azure/aws? The only need is the ability to have some static pods. I do have a preference for free/open source solutions.
I just stumbled upon this project a while back and don't have experience with it, so I don't know how well it works and what caveats you may face, but there's Virtual Kubelet, which aims to do just that, i.e. running a virtual Kubernetes node outside the cluster. Its Kip provider sounds like the thing you're looking for.
-
Create a fake node
Are you looking something like the following: https://github.com/virtual-kubelet/virtual-kubelet
-
How to use the GitOps model to create, update and manage applications at the edge with KubeEdge and Argo
Kubeedge docs are light on self-justification... How does https://github.com/kubeedge/kubeedge differ from https://github.com/virtual-kubelet/virtual-kubelet or just running a regular kubelet on that edge machine?
-
Autoscaling Redis applications on Kubernetes 🚀🚀
If this sounds interesting, do check out Virtual Nodes in Azure Kubernetes Service to see how you can use them to seamlessly scale your applications to Azure Container Instances and benefit from quick provisioning of pods, and only pay per second for their execution time. The virtual nodes add-on for AKS, is based on the open source project Virtual Kubelet which is an open source Kubernetes kubelet implementation.
-
Infrastructure Engineering - Diving Deep
Use cases like these are made possible by projects like KubeEdge , K3s and Virtual Kubelets. You can read more about how they power the edge with different architectures and compromises here.
-
Evolving Container Security with Linux User Namespaces
This is a complicated question to answer.
This isn't my expertise (the cluster orchestration system), but I can answer to the best of my abilities: Titus, today is a system that sits on top of Kubernetes, and uses Kubernetes components to do its thing, but we've substituted many of the systems with our own. For example, closer to my area of knowledge, we've used our own executor / provider along with the Virtual Kubelet project (https://github.com/virtual-kubelet/virtual-kubelet) instead of Kubelet.
We're exploring where we can leverage the Kubernetes ecosystem, adapt components, or help contribute changes back that others can leverage to enable our use of more COTS components of Kubernetes.
tl;dr: We're swapping out the engines while in flight
What are some alternatives?
etcd - Distributed reliable key-value store for the most critical data of a distributed system [Moved to: https://github.com/etcd-io/etcd]
kubeedge - Kubernetes Native Edge Computing Framework (project under CNCF)
magento2-on-kubernetes - Configuration to deploy Magento 2 in a Kubernetes cluster
kubevirt - Kubernetes Virtualization API and runtime in order to define and manage virtual machines.
hetzner-lb-acmedns - A service to automatically update Letsencrypt SSL certificates on the Hetzner load-balancer using joohoi/acme-dns.
kubefed - Kubernetes Cluster Federation
cluster-api-provider-hetzner - Kubernetes Cluster API Provider Hetzner provides a consistent deployment and day 2 operations of "self-managed" Kubernetes clusters on Hetzner.
keda - KEDA is a Kubernetes-based Event Driven Autoscaling component. It provides event driven scale for any container running in Kubernetes
terraform-provider-hcloud - Terraform Hetzner Cloud provider
cni - Container Network Interface - networking for Linux containers
hetzner.hcloud - A collection to manage resources on Hetzner Cloud
charts - ⚠️(OBSOLETE) Curated applications for Kubernetes