k3s-oci-cluster
ingress-nginx
Our great sponsors
k3s-oci-cluster | ingress-nginx | |
---|---|---|
6 | 202 | |
208 | 16,638 | |
- | 1.5% | |
0.0 | 9.6 | |
5 months ago | 4 days ago | |
HCL | Go | |
GNU General Public License v3.0 only | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
k3s-oci-cluster
-
K8s cluster with OCI free-tier and Raspberry Pi4 (part 1)
After you've cloned the repo, go to oci/terraform.tfvars and edit all values with the ones from your notes file. This build uses the great terraform configuration files from this repo of garutilorenzo (using version 2.2; if you have errors running all of this, you should check what changed in this repo since v2.2, or 01.02.23). You can read here if you want to customize your configuration and edit the main.tf file. This is the diagram that garutilorenzo made and how your deployment will look like (this tutorial is without Longhorn and ArgoCD, with 1 server nodes + 3 worker nodes and with ingress controller set as Traefik): *note - I've got some problems with clock of WSL2 not being synced to Windows clock. And provisioning didn't worked so if you receive clock errors too, verify your time with datecommand, if out of sync just run sudo hwclock -s or sudo ntpdate time.windows.com. Now just run terraform plan and then terraform apply. If everything was ok you should have your resources created.
- Is there a step by step guide to Oracle Free Tier setups?
-
OCI free Kubernetes cluster, new terraform module release
This new release v2.0 introduces:
-
Deploy Kubernetes (K8s) on Amazon AWS using mixed on-demand and spot instances
We use the same stack used in this repository. This stack need longhorn and nginx ingress.
-
Deploy a Kubernetes cluster for free, using k3s and Oracle always free resources.
This module will deploy a Kubernetes cluster for free, using K3s and Oracle always free resources.
-
Deploy a Kubernetes cluster for free, using K3s and Oracle always free resources
Var Required Desc region yes set the correct OCI region based on your needs availability_domain yes Set the correct availability domain. See how to find the availability domain compartment_ocid yes Set the correct compartment ocid. See how to find the compartment ocid cluster_name yes the name of your K3s cluster. Default: k3s-cluster k3s_token yes The token of your K3s cluster. How to generate a random token my_public_ip_cidr yes your public ip in cidr format (Example: 195.102.xxx.xxx/32) environment yes Current work environment (Example: staging/dev/prod). This value is used for tag all the deployed resources compute_shape no Compute shape to use. Default VM.Standard.A1.Flex. NOTE Is mandatory to use this compute shape for provision 4 always free VMs os_image_id no Image id to use. Default image: Canonical-Ubuntu-20.04-aarch64-2022.01.18-0. See how to list all available OS images oci_core_vcn_cidr no VCN CIDR. Default: oci_core_vcn_cidr oci_core_subnet_cidr10 no First subnet CIDR. Default: 10.0.0.0/24 oci_core_subnet_cidr11 no Second subnet CIDR. Default: 10.0.1.0/24 oci_identity_dynamic_group_name no Dynamic group name. This dynamic group will contains all the instances of this specific compartment. Default: Compute_Dynamic_Group oci_identity_policy_name no Policy name. This policy will allow dynamic group 'oci_identity_dynamic_group_name' to read OCI api without auth. Default: Compute_To_Oci_Api_Policy kube_api_port no Kube api default port Default: 6443 public_lb_shape no LB shape for the public LB. Default: flexible. NOTE is mandatory to use this kind of shape to provision two always free LB (public and private) http_lb_port no http port used by the public LB. Default: 80 https_lb_port no http port used by the public LB. Default: 443 k3s_server_pool_size no Number of k3s servers deployed. Default 2 k3s_worker_pool_size no Number of k3s workers deployed. Default 2 install_longhorn no Boolean value, install longhorn "Cloud native distributed block storage for Kubernetes". Default: true longhorn_release no Longhorn release. Default: v1.2.3 unique_tag_key no Unique tag name used for tagging all the deployed resources. Default: k3s-provisioner unique_tag_value no Unique value used with unique_tag_key. Default: https://github.com/garutilorenzo/k3s-oci-cluster PATH_TO_PUBLIC_KEY no Path to your public ssh key (Default: "~/.ssh/id_rsa.pub) PATH_TO_PRIVATE_KEY no Path to your private ssh key (Default: "~/.ssh/id_rsa)
ingress-nginx
-
deploying a minio service to kubernetes
ingress-nginx
-
Upgrading Hundreds of Kubernetes Clusters
The second one is a combination of tools: External DNS, cert-manager, and NGINX ingress. Using these as a stack, you can quickly deploy an application, making it available through a DNS with a TLS without much effort via simple annotations. When I first discovered External DNS, I was amazed at its quality.
-
[06/52] Accessible Kubernetes with Terraform and DigitalOcean
resource "helm_release" "icrelease" { name = "nginx-ingress" repository = "https://kubernetes.github.io/ingress-nginx" chart = "ingress-nginx" version = "4.9.1" namespace = kubernetes_namespace.icnamespace.metadata[0].name set { name = "controller.ingressClassResource.default" value = "true" } }
-
Deploy Ghost with MySQL DB replication using helm chart
helm repo add ingress-nginx https://kubernetes.github.io/ingress-nginx helm repo update helm upgrade --install ingress-nginx ingress-nginx/ingress-nginx --namespace ingress-nginx --create-namespace -f custom/ghost/nginx.yaml
-
Kubernetes Gateway API v1.0: Should You Switch?
For example, if you chose Nginx Ingress, you will use some of its dozens of annotations that are not portable if you decide to switch to another Ingress implementation like Apache APISIX.
- nginx ingress controller installation
-
IP-Whitlisting: Is adjusting nginx-ingress-controller service a solution?
The controller is installed with helm upgrade --install ingress-nginx ingress-nginx --repo https://kubernetes.github.io/ingress-nginx --namespace ingress-nginx --create-namespace
-
Deploy Rancher on AWS EKS using Terraform & Helm Charts
helm repo add jetstack https://charts.jetstack.io helm repo add ingress-nginx https://kubernetes.github.io/ingress-nginx helm repo add rancher-latest https://releases.rancher.com/server-charts/latest helm repo update helm repo list
-
☸️ Kubernetes NGINX Ingress Controller: 10+ Complementary Configurations for Web Applications
Everything in the YAML snippets below — except for ingress configuration — relates to configuring the NGINX ingress controller. This includes customizing the default configuration.
-
Debugging NGINX Ingress Controller: A Comprehensive Guide
Keep Documentation Handy: Always have the official NGINX Ingress documentation accessible.
What are some alternatives?
oracle-cloud-terraform-examples - Oracle cloud terraform examples, provision oracle cloud resources using terraform
traefik - The Cloud Native Application Proxy
k3s-aws-terraform-cluster - Deploy an high available K3s cluster on Amazon AWS
emissary - open source Kubernetes-native API gateway for microservices built on the Envoy Proxy
longhorn - Cloud-Native distributed storage built on and for Kubernetes
metallb - A network load-balancer implementation for Kubernetes using standard routing protocols
terraform-oci-tdf-network-security - (OCI) Oracle Cloud Infrastructure module to manage security policies
cilium-cli - CLI to install, manage & troubleshoot Kubernetes clusters running Cilium
flannel - flannel is a network fabric for containers, designed for Kubernetes
haproxy-ingress - HAProxy Ingress
terraform-hcloud-kube-hetzner - Optimized and Maintenance-free Kubernetes on Hetzner Cloud in one command!
application-gateway-kubernetes-ingress - This is an ingress controller that can be run on Azure Kubernetes Service (AKS) to allow an Azure Application Gateway to act as the ingress for an AKS cluster.