actions-runner-controller
cert-manager
Our great sponsors
actions-runner-controller | cert-manager | |
---|---|---|
31 | 101 | |
4,216 | 11,457 | |
3.4% | 1.7% | |
9.0 | 9.8 | |
5 days ago | 1 day ago | |
Go | Go | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
actions-runner-controller
-
Using Kaniko to Build and Publish container image with Github action on Github Self-hosted Runners
To set-up the self-hosted runner, an Action Runner Controller (ARC) and Runner scale sets application will be installed via helm. This post will be using Azure Kubernetes Service and ARC that is officialy maintained by Github. There is another ARC that is maintained by the community. You can follow the discussion where github adopted the ARC project into a full Github product here
-
Show HN: DimeRun v2 β Run GitHub Actions on AWS EC2
Before this we were using https://github.com/actions/actions-runner-controller but that's running on K8s instead of VMs. So along with common limitations of running CI jos in K8s/container, it cannot have exactly the same environment as the official GitHub runners. Maintaining a K8s cluster was also very difficult.
-
Terraform module for scalable GitHub action runners on AWS
ARC is great for running GitHub Actions on Kubernetes:
https://github.com/actions/actions-runner-controller
-
Best CI/CD for AWS services?
Almost all of our cicd, builds run on GitHub. I'm talking cypress tests, deployments via terraform and helm to over 25 environments, all backend tests, daily test runs etc. Overall we were racking up a cost of almost 20k on GitHub. With the ARC deployed and using spot instances I think our total infrastructure costs went up about 4-5k even though we added more actions. If we switched back to their runners we'd probably be around 25k at this point.
-
Running helm from within network
What else needs to be moved to my artifactory (charts - https://github.com/actions/actions-runner-controller/tree/master/charts ) - if so tar or entire folder or anything else ? ) What should the above steps correspond to?
-
Action-runner-controller & Enterprise Git
You need to use the steps in the repo instead of the steps on the docs if you're using enterprise server.
-
GitHub support for Actions Runner Controller (ARC) emerging in docs!
Honestly not a fan of Github docs.....I feel like the ones in the repo are much clearer and easier to understand/read.
-
How much work does it take to operate a self-hosted GitHub runners?
Its pretty easy to set up honestly. Deploy this on your k8s cluster https://github.com/actions/actions-runner-controller and a runnerDeployment and youre good to go.
-
Self-Hosted runner on Kubernetes
Trying to use the Actions Runner Controller (https://github.com/actions/actions-runner-controller) to utilize self-hosted runners. I keep getting this error on the controller.
-
AKS cluster w/ GitHub App and Actions Runner Controller
I'm convinced one of (or a combination) of things is happening here in regards to authentication. This GH enterprise account is configured with SAML. I feel like that is a valid data point. I'm using https://github.com/actions/actions-runner-controller as a reference guide for what I should be doing. I suspect whoever is Owner of this organization has modified what I can do as a user. The steps in the doc where I can actually Install the Application isn't available to me. When configuring the GitHub App I'm given two options. I select the option for "this account only" knowing the documentation says it is possible to use this Github App with a repo in the Organization as long as I have Admin privileges or I'm the owner.
cert-manager
-
deploying a minio service to kubernetes
cert-manager
-
Upgrading Hundreds of Kubernetes Clusters
The second one is a combination of tools: External DNS, cert-manager, and NGINX ingress. Using these as a stack, you can quickly deploy an application, making it available through a DNS with a TLS without much effort via simple annotations. When I first discovered External DNS, I was amazed at its quality.
-
Run WebAssembly on DigitalOcean Kubernetes with SpinKube - In 4 Easy Steps
On top of its core components, SpinKube depends on cert-manager. cert-Manager is responsible for provisioning and managing TLS certificates that are used by the admission webhook system of the Spin Operator. Letβs install cert-manager and KWasm using the commands shown here:
-
Importing kubernetes manifests with terraform for cert-manager
terraform { required_providers { kubectl = { source = "gavinbunney/kubectl" version = "1.14.0" } } } # The reference to the current project or a AWS project data "google_client_config" "provider" {} # The reference to the current cluster or EKS data "google_container_cluster" "my_cluster" { name = var.cluster_name location = var.cluster_location } # We configure the kubectl provider to use those values for authenticating provider "kubectl" { host = data.google_container_cluster.my_cluster.endpoint token = data.google_client_config.provider.access_token cluster_ca_certificate = base64decode(data.google_container_cluster.my_cluster.master_auth[0].cluster_ca_certificate) } #Download the multiple manifests file. data "http" "cert_manager_crds" { url = "https://github.com/cert-manager/cert-manager/releases/download/v${var.cert_manager_version}/cert-manager.crds.yaml" } data "kubectl_file_documents" "cert_manager_crds" { content = data.http.cert_manager_crds.response_body lifecycle { precondition { condition = 200 == data.http.cert_manager_crds.status_code error_message = "Status code invalid" } } } # We use the for_each or else this kubectl_manifest will only import the first manifest in the file. resource "kubectl_manifest" "cert_manager_crds" { for_each = data.kubectl_file_documents.cert_manager_crds.manifests yaml_body = each.value }
-
An opinionated template for deploying a single k3s cluster with Ansible backed by Flux, SOPS, GitHub Actions, Renovate, Cilium, Cloudflare and more!
SSL certificates thanks to Cloudflare and cert-manager
-
Deploy Rancher on AWS EKS using Terraform & Helm Charts
kubectl apply -f https://github.com/cert-manager/cert-manager/releases/download/${CERT_MANAGER_VERSION}/cert-manager.crds.yaml
-
Setup/Design internal PKI
put the Sub-CA inside hashicorp vault to be used for automatic signing of services like https://cert-manager.io/ inside our k8s clusters.
-
Task vs Make - Final Thoughts
install-cert-manager: desc: Install cert-manager deps: - init-cluster cmds: - kubectl apply -f https://github.com/cert-manager/cert-manager/releases/download/{{.CERT_MANAGER_VERSION}}/cert-manager.yaml - echo "Waiting for cert-manager to be ready" && sleep 25 status: - kubectl -n cert-manager get pods | grep Running | wc -l | grep -q 3
-
Easy HTTPS for your private networks
I've been pretty frustrated with how private CAs are supported. Your private root CA can be maliciously used to MITM every domain on the Internet, even though you intend to use it for only a couple domain names. Most people forget to set Name Constraints when they create these and many helper tools lack support [1][2]. Worse, browser support for Name Constraints has been slow [3] and support isn't well tracked [4]. Public CAs give you certificate transparency and you can subscribe to events to detect mis-issuance. Some hosted private CAs like AWS's offer logs [5], but DIY setups don't.
Even still, there are a lot of folks happily using private CAs, they aren't the target audience for this initial release.
[1] https://github.com/FiloSottile/mkcert/issues/302
[2] https://github.com/cert-manager/cert-manager/issues/3655
[3] https://alexsci.com/blog/name-non-constraint/
[4] https://github.com/Netflix/bettertls/issues/19
[5] https://docs.aws.amazon.com/privateca/latest/userguide/secur...
-
βΈοΈ Managed Kubernetes : Our dev is on AWS, our prod is on OVH
the Cert Manager
What are some alternatives?
helm-charts - Jenkins helm charts
metallb - A network load-balancer implementation for Kubernetes using standard routing protocols
turnstyle - ποΈA GitHub Action for serializing workflow runs
aws-load-balancer-controller - A Kubernetes controller for Elastic Load Balancers
cache - Cache dependencies and build outputs in GitHub Actions
Portainer - Making Docker and Kubernetes management easy.
azure-pipelines-agent - Azure Pipelines Agent π
awx-operator - An Ansible AWX operator for Kubernetes built with Operator SDK and Ansible. π€
ghat - π Reuse GitHub Actions workflows across repositories
k3s - Lightweight Kubernetes
actions-runner-
oauth2-proxy - A reverse proxy that provides authentication with Google, Azure, OpenID Connect and many more identity providers.