eksctl
keda
eksctl | keda | |
---|---|---|
59 | 91 | |
4,791 | 7,791 | |
0.9% | 1.4% | |
9.5 | 9.5 | |
3 days ago | about 9 hours ago | |
Go | Go | |
GNU General Public License v3.0 or later | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
eksctl
-
Auto-scaling DynamoDB Streams applications on Kubernetes
There are a variety of ways in which you can create an Amazon EKS cluster. I prefer using eksctl CLI because of the convenience it offers. Creating an an EKS cluster using eksctl, can be as easy as this:
-
How to migrate Apache Solr from the existing cluster to Amazon EKS
There are many ways to create a cluster such as using eksctl. In my case, I will use terraform module cause itβs easy to reuse and comprehend.
-
Ultimate EKS Baseline Cluster: Part 1 - Provision EKS
eksctl [eksctl] is the tool that can provision EKS cluster as well as supporting VPC network infrastructure.
-
[AWS] EKS vs Self managed HA k3s running on 1x2 ec2 machines, for medium production workload
For this and many other reasons I recommend doing everything in Terraform EXCEPT EKS and its node groups. For that, I use https://eksctl.io/ because it much better manages the lifecycle of EKS and your node groups. I have an blog article better explaining why I recommend it, and another blog article explaining how to do zero-downtime upgrades with EKSCTL.
-
Automating Kong API Gateway deployment with Flux
eksctl
- Export a docker container to a VPC in AWS and exposing it publicly through a loadbalancer
-
Anybody using spot instances for worker nodes?
Second, make sure you create a spot instance group that attempts to launch MULTIPLE different instance types. This way if one instance type gets flushed, your autoscaler will kick in and launch a different type. Without this, you WILL HAVE DOWNTIME if a sudden price hike and flush occurs. If you're using eksctl I have example configurations that use multi-instance types on Github here.
-
Use AWS Controllers for Kubernetes to deploy a Serverless data processing solution with SQS, Lambda and DynamoDB
There are a variety of ways in which you can create an Amazon EKS cluster. I prefer using eksctl CLI because of the convenience it offers. Creating an an EKS cluster using eksctl, can be as easy as this:
-
strategy to upgrade eks cluster
I've written an article on this, with my recommended tool for managing eks EKSCTL.
-
Bootstrapping Kubernetes Cluster with CloudFormation
--- AWSTemplateFormatVersion: '2010-09-09' Parameters: VpcId: Type: AWS::EC2::VPC::Id Description: ID of the VPC in which to create the Kubernetes cluster SubnetIds: Type: List Description: List of Subnet IDs in which to create the Kubernetes cluster KeyPairName: Type: AWS::EC2::KeyPair::KeyName Description: Name of the EC2 Key Pair to use for SSH access to worker nodes ClusterName: Type: String Description: Name of the Kubernetes cluster to create Resources: ControlPlaneSecurityGroup: Type: AWS::EC2::SecurityGroup Properties: VpcId: !Ref VpcId GroupDescription: Allow inbound traffic to the Kubernetes control plane SecurityGroupIngress: - IpProtocol: tcp FromPort: 22 ToPort: 22 CidrIp: 0.0.0.0/0 WorkerNodeSecurityGroup: Type: AWS::EC2::SecurityGroup Properties: VpcId: !Ref VpcId GroupDescription: Allow inbound traffic to Kubernetes worker nodes SecurityGroupIngress: - IpProtocol: tcp FromPort: 22 ToPort: 22 CidrIp: 0.0.0.0/0 ControlPlaneInstanceProfile: Type: AWS::IAM::InstanceProfile Properties: Roles: - !Ref ControlPlaneRole ControlPlaneRole: Type: AWS::IAM::Role Properties: AssumeRolePolicyDocument: Version: '2012-10-17' Statement: - Effect: Allow Principal: Service: - ec2.amazonaws.com Action: - sts:AssumeRole ManagedPolicyArns: - arn:aws:iam::aws:policy/AmazonEKSClusterPolicy - arn:aws:iam::aws:policy/AmazonEKSServicePolicy ControlPlaneInstance: Type: AWS::EC2::Instance Properties: ImageId: ami-0b69ea66ff7391e80 InstanceType: t2.micro KeyName: !Ref KeyPairName NetworkInterfaces: - DeviceIndex: 0 AssociatePublicIpAddress: true GroupSet: - !Ref ControlPlaneSecurityGroup SubnetId: !Select [0, !Ref SubnetIds] IamInstanceProfile: !Ref ControlPlaneInstanceProfile UserData: Fn::Base64: !Sub | #!/bin/bash echo 'net.bridge.bridge-nf-call-iptables=1' | tee -a /etc/sysctl.conf sysctl -p yum update -y amazon-linux-extras install docker -y service docker start usermod -a -G docker ec2-user curl -o /usr/local/bin/kubectl https://amazon-eks.s3.us-west-2.amazonaws.com/1.21.2/2021-07-05/bin/linux/amd64/kubectl chmod +x /usr/local/bin/kubectl echo 'export PATH=$PATH:/usr/local/bin' >> /etc/bashrc curl --silent --location "https://github.com/weaveworks/eksctl/releases
keda
-
Ask HN: What's the right way to scale K8s for GPU workloads?
It seems you want something like KEDA (https://keda.sh)
-
Tortoise: Shell-Shockingly-Good Kubernetes Autoscaling
Most just utilize out of the box macro resources available in HPA.
For more advanced use cases there is keda - https://keda.sh/
-
Root Cause Chronicles: Quivering Queue
Thankfully KEDA operator was already part of the cluster, and all Robin had to do was create a ScaledObject manifest targeting the Dispatch ScaleUp event, based on the rabbitmq_global_messages_received_total metric from Prometheus.
-
Five tools to add to your K8s cluster
Keda
- K8s latencies in chained services - Using RL?
-
Best Kubernetes DevOps Tools: A Comprehensive Guide
KEDA introduces event-driven scaling to Kubernetes workloads. It integrates with Kubernetes Horizontal Pod Autoscalers and can scale pods based on external metrics from services like databases and message queues (Kafka, RabbitMQ, MongoDB).
-
Auto-scaling DynamoDB Streams applications on Kubernetes
# update version 2.8.2 if required kubectl apply -f https://github.com/kedacore/keda/releases/download/v2.8.2/keda-2.8.2.yaml
- KEDA
-
What is the difference in production for scale to zero usecases - Keda vs Lambda ?
This is traditionally a AWS Lambda usecase - or an OpenFaas kind of usecase. But very recently i discovered https://keda.sh/ and it seems it is specifically meant for this in a kubernetes environment.
-
Ingesting Data into OpenSearch using Apache Kafka and Go
If you deploy the application to Amazon EKS, you can also consider using KEDA to auto-scale your consumer application based on the number of messages in the MSK topic.
What are some alternatives?
terraform-aws-eks - Terraform module to create AWS Elastic Kubernetes (EKS) resources πΊπ¦
k8s-prometheus-adapter - An implementation of the custom.metrics.k8s.io API using Prometheus
kops - Kubernetes Operations (kOps) - Production Grade k8s Installation, Upgrades and Management
argo - Workflow Engine for Kubernetes
argo-cd - Declarative Continuous Deployment for Kubernetes
istio - Connect, secure, control, and observe services.
terraform-aws-eks-blueprints - Configure and deploy complete EKS clusters.
karpenter-provider-aws - Karpenter is a Kubernetes Node Autoscaler built for flexibility, performance, and simplicity.
cluster-api - Home for Cluster API, a subproject of sig-cluster-lifecycle
helm - The Kubernetes Package Manager
eks-anywhere - Run Amazon EKS on your own infrastructure π
http-add-on - Add-on for KEDA to scale HTTP workloads