containers-roadmap
eksctl
Our great sponsors
containers-roadmap | eksctl | |
---|---|---|
80 | 59 | |
5,137 | 4,765 | |
0.6% | 0.9% | |
2.0 | 9.6 | |
9 months ago | 8 days ago | |
Shell | Go | |
GNU General Public License v3.0 or later | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
containers-roadmap
-
General Availability of the AWS SDK for Rust
Thanks for showing up and answering questions. Congratulations on the release.
What kind of plans for support of Rust's evolving async ecosystem?
Any particular reason why the public roadmap does not show the columns similar to "Researching", "We're Working On It" like the other similar public AWS Roadmaps? See example for Containers: https://github.com/aws/containers-roadmap/projects/1
Would be nice to have fully working examples on Github, for most common scenarios across most AWS services. This is something that historically
-
Architecting for Resilience: Crafting Opinionated EKS Clusters with Karpenter & Cilium Cluster Mesh β Part 1
Second, We will only rely on one managed node group, but we will leverage Karpenter; however, karpenter needs to be deployed on a node. (This may change soon once the Karpenter is available on the EKS Control Plane.) [EKS] Karpenter inside control plane Β· Issue #1792 Β· aws/containers-roadmap
-
Running a Web Application with 100% AWS Fargate Spot Containers π€
As written in the AWS documentation, during periods of extremely high demand, Fargate Spot capacity might be unavailable. In concrete terms, if your ECS service is set up to execute tasks in 100% Spot, there is a risk of running out of capacity. A workaround has been created in the hope that one day this issue will be implemented by the AWS team. This workaround allows you to set up two ECS services :
-
Kubernetes SidecarContainers feature is merged
Deploying Fargate with CDK has to have been the most pleasant developer experience I have ever had with any product so far.
If image caching becomes a reality with Fargate I can't imagine a need to ever use anything else
- AWS Config supports recording exclusions by resource type
- Announcing pull through cache for registry.k8s.io in Amazon Elastic Container Registry
-
EKS/Spot vs EKS Fargate/Spot?
Eks Fargate doesn't support spot yet https://github.com/aws/containers-roadmap/issues/622
- audit logging of the master plane in EKS
-
How to use Podman inside of a container
Until podman could be used with AWS ECR/ECS it's pretty much moot in my case: https://github.com/aws/containers-roadmap/issues/626
- How to keep 100% availability with a single ec2 spot instance?
eksctl
-
Auto-scaling DynamoDB Streams applications on Kubernetes
There are a variety of ways in which you can create an Amazon EKS cluster. I prefer using eksctl CLI because of the convenience it offers. Creating an an EKS cluster using eksctl, can be as easy as this:
-
How to migrate Apache Solr from the existing cluster to Amazon EKS
There are many ways to create a cluster such as using eksctl. In my case, I will use terraform module cause itβs easy to reuse and comprehend.
-
Ultimate EKS Baseline Cluster: Part 1 - Provision EKS
eksctl [eksctl] is the tool that can provision EKS cluster as well as supporting VPC network infrastructure.
-
[AWS] EKS vs Self managed HA k3s running on 1x2 ec2 machines, for medium production workload
For this and many other reasons I recommend doing everything in Terraform EXCEPT EKS and its node groups. For that, I use https://eksctl.io/ because it much better manages the lifecycle of EKS and your node groups. I have an blog article better explaining why I recommend it, and another blog article explaining how to do zero-downtime upgrades with EKSCTL.
-
Automating Kong API Gateway deployment with Flux
eksctl
- Export a docker container to a VPC in AWS and exposing it publicly through a loadbalancer
-
Anybody using spot instances for worker nodes?
Second, make sure you create a spot instance group that attempts to launch MULTIPLE different instance types. This way if one instance type gets flushed, your autoscaler will kick in and launch a different type. Without this, you WILL HAVE DOWNTIME if a sudden price hike and flush occurs. If you're using eksctl I have example configurations that use multi-instance types on Github here.
-
Use AWS Controllers for Kubernetes to deploy a Serverless data processing solution with SQS, Lambda and DynamoDB
There are a variety of ways in which you can create an Amazon EKS cluster. I prefer using eksctl CLI because of the convenience it offers. Creating an an EKS cluster using eksctl, can be as easy as this:
-
strategy to upgrade eks cluster
I've written an article on this, with my recommended tool for managing eks EKSCTL.
-
Bootstrapping Kubernetes Cluster with CloudFormation
--- AWSTemplateFormatVersion: '2010-09-09' Parameters: VpcId: Type: AWS::EC2::VPC::Id Description: ID of the VPC in which to create the Kubernetes cluster SubnetIds: Type: List Description: List of Subnet IDs in which to create the Kubernetes cluster KeyPairName: Type: AWS::EC2::KeyPair::KeyName Description: Name of the EC2 Key Pair to use for SSH access to worker nodes ClusterName: Type: String Description: Name of the Kubernetes cluster to create Resources: ControlPlaneSecurityGroup: Type: AWS::EC2::SecurityGroup Properties: VpcId: !Ref VpcId GroupDescription: Allow inbound traffic to the Kubernetes control plane SecurityGroupIngress: - IpProtocol: tcp FromPort: 22 ToPort: 22 CidrIp: 0.0.0.0/0 WorkerNodeSecurityGroup: Type: AWS::EC2::SecurityGroup Properties: VpcId: !Ref VpcId GroupDescription: Allow inbound traffic to Kubernetes worker nodes SecurityGroupIngress: - IpProtocol: tcp FromPort: 22 ToPort: 22 CidrIp: 0.0.0.0/0 ControlPlaneInstanceProfile: Type: AWS::IAM::InstanceProfile Properties: Roles: - !Ref ControlPlaneRole ControlPlaneRole: Type: AWS::IAM::Role Properties: AssumeRolePolicyDocument: Version: '2012-10-17' Statement: - Effect: Allow Principal: Service: - ec2.amazonaws.com Action: - sts:AssumeRole ManagedPolicyArns: - arn:aws:iam::aws:policy/AmazonEKSClusterPolicy - arn:aws:iam::aws:policy/AmazonEKSServicePolicy ControlPlaneInstance: Type: AWS::EC2::Instance Properties: ImageId: ami-0b69ea66ff7391e80 InstanceType: t2.micro KeyName: !Ref KeyPairName NetworkInterfaces: - DeviceIndex: 0 AssociatePublicIpAddress: true GroupSet: - !Ref ControlPlaneSecurityGroup SubnetId: !Select [0, !Ref SubnetIds] IamInstanceProfile: !Ref ControlPlaneInstanceProfile UserData: Fn::Base64: !Sub | #!/bin/bash echo 'net.bridge.bridge-nf-call-iptables=1' | tee -a /etc/sysctl.conf sysctl -p yum update -y amazon-linux-extras install docker -y service docker start usermod -a -G docker ec2-user curl -o /usr/local/bin/kubectl https://amazon-eks.s3.us-west-2.amazonaws.com/1.21.2/2021-07-05/bin/linux/amd64/kubectl chmod +x /usr/local/bin/kubectl echo 'export PATH=$PATH:/usr/local/bin' >> /etc/bashrc curl --silent --location "https://github.com/weaveworks/eksctl/releases
What are some alternatives?
eks-nvme-ssd-provisioner - EKS NVMe SSD provisioner for Amazon EC2 Instance Stores
terraform-aws-eks - Terraform module to create AWS Elastic Kubernetes (EKS) resources πΊπ¦
kube-fledged - A kubernetes operator for creating and managing a cache of container images directly on the cluster worker nodes, so application pods start almost instantly
kops - Kubernetes Operations (kOps) - Production Grade k8s Installation, Upgrades and Management
netshoot - a Docker + Kubernetes network trouble-shooting swiss-army container
argo-cd - Declarative Continuous Deployment for Kubernetes
kraken - P2P Docker registry capable of distributing TBs of data in seconds
terraform-aws-eks-blueprints - Configure and deploy complete EKS clusters.
juicefs - JuiceFS is a distributed POSIX file system built on top of Redis and S3.
cluster-api - Home for Cluster API, a subproject of sig-cluster-lifecycle
piku - The tiniest PaaS you've ever seen. Piku allows you to do git push deployments to your own servers.
eks-anywhere - Run Amazon EKS on your own infrastructure π