containers-roadmap
autoscaler
containers-roadmap | autoscaler | |
---|---|---|
80 | 89 | |
5,146 | 7,633 | |
0.4% | 0.7% | |
2.0 | 9.7 | |
9 months ago | about 5 hours ago | |
Shell | Go | |
GNU General Public License v3.0 or later | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
containers-roadmap
-
General Availability of the AWS SDK for Rust
Thanks for showing up and answering questions. Congratulations on the release.
What kind of plans for support of Rust's evolving async ecosystem?
Any particular reason why the public roadmap does not show the columns similar to "Researching", "We're Working On It" like the other similar public AWS Roadmaps? See example for Containers: https://github.com/aws/containers-roadmap/projects/1
Would be nice to have fully working examples on Github, for most common scenarios across most AWS services. This is something that historically
-
Architecting for Resilience: Crafting Opinionated EKS Clusters with Karpenter & Cilium Cluster Mesh — Part 1
Second, We will only rely on one managed node group, but we will leverage Karpenter; however, karpenter needs to be deployed on a node. (This may change soon once the Karpenter is available on the EKS Control Plane.) [EKS] Karpenter inside control plane · Issue #1792 · aws/containers-roadmap
-
Running a Web Application with 100% AWS Fargate Spot Containers 🤘
As written in the AWS documentation, during periods of extremely high demand, Fargate Spot capacity might be unavailable. In concrete terms, if your ECS service is set up to execute tasks in 100% Spot, there is a risk of running out of capacity. A workaround has been created in the hope that one day this issue will be implemented by the AWS team. This workaround allows you to set up two ECS services :
-
Kubernetes SidecarContainers feature is merged
Deploying Fargate with CDK has to have been the most pleasant developer experience I have ever had with any product so far.
If image caching becomes a reality with Fargate I can't imagine a need to ever use anything else
https://github.com/aws/containers-roadmap/issues/696
- AWS Config supports recording exclusions by resource type
- Announcing pull through cache for registry.k8s.io in Amazon Elastic Container Registry
-
EKS/Spot vs EKS Fargate/Spot?
Eks Fargate doesn't support spot yet https://github.com/aws/containers-roadmap/issues/622
- audit logging of the master plane in EKS
-
How to use Podman inside of a container
Until podman could be used with AWS ECR/ECS it's pretty much moot in my case: https://github.com/aws/containers-roadmap/issues/626
- How to keep 100% availability with a single ec2 spot instance?
autoscaler
-
Upgrading Hundreds of Kubernetes Clusters
We use Cluster Autoscaler to automatically adjust the number of nodes (cluster size) based on your actual usage to ensure efficiency. Additionally, we deploy Vertical and Horizontal Pod Autoscalers to scale your applications' resources as their needs change automatically.
-
Not Everything Is Google's Fault (Just Most Things)
> * Hetzner: cheap, good service, the finest pets in the world, no cattle
You can absolutely do cattle with Hetzner. They support imaging and immutable infrastructure. They don't have a native auto scaling equivalent, but if you're using Kubernetes, they have a cluster autoscaler: https://github.com/kubernetes/autoscaler/blob/master/cluster...
-
Kubernetes(K8s) Autoscaler — a detailed look at the design and implementation of VPA
Here we take the VPA as a starting point to analyze the design and implementation principles of the VPA in Autoscaler. The source code for this article is based on Autoscaler HEAD fbe25e1.
- Scaling with Karpenter and Empty Pod(A.k.a Overprovisioning)
-
Reducing Cloud Costs on Kubernetes Dev Envs
Autoscaling over EKS can be accomplished using either the cluster-autoscaler project or Karpenter. If you want to use Spot instances, consider using Karpenter, as it has better integrations with AWS for optimizing spot pricing and availability, minimizing interruptions, and falling back to on-demand nodes if no spot instances are available.
-
☸️ Managed Kubernetes : Our dev is on AWS, our prod is on OVH
Autoscaling is already provided on OVH, but we don't use it for now. Autoscaler has to be manually installed on the AWS/EKS cluster.
-
relevant way of scaling pods
do you mean this: https://github.com/kubernetes/autoscaler/blob/master/vertical-pod-autoscaler/pkg/recommender/README.md
-
Kubernetes Cluster Maintenance
Read more about this scaler in detail here!
-
Anyone running Windows nodes in your clusters?
We have a default node group of Linux hosts, but there's a secondary nodegroup of Windows hosts that is typically scaled down to 0. When a team's build runs, a pod is scheduled based on their definition. Cluster-autoscaler will check the nodeSelector and automatically spin up a node from that nodegroup if necessary.
-
How to make sure Kubernetes autoscaler not deleting the nodes which runs specific pod
I am running a Kubernetes cluster(AWS EKS one) with Autoscaler pod So that Cluster will autoscale according to the resource request within the cluster.
What are some alternatives?
eks-nvme-ssd-provisioner - EKS NVMe SSD provisioner for Amazon EC2 Instance Stores
karpenter-provider-aws - Karpenter is a Kubernetes Node Autoscaler built for flexibility, performance, and simplicity.
kube-fledged - A kubernetes operator for creating and managing a cache of container images directly on the cluster worker nodes, so application pods start almost instantly
cluster-proportional-autoscaler - Kubernetes Cluster Proportional Autoscaler Container
netshoot - a Docker + Kubernetes network trouble-shooting swiss-army container
aws-ebs-csi-driver - CSI driver for Amazon EBS https://aws.amazon.com/ebs/
kraken - P2P Docker registry capable of distributing TBs of data in seconds
keda - KEDA is a Kubernetes-based Event Driven Autoscaling component. It provides event driven scale for any container running in Kubernetes
juicefs - JuiceFS is a distributed POSIX file system built on top of Redis and S3.
descheduler - Descheduler for Kubernetes
copilot-cli - The AWS Copilot CLI is a tool for developers to build, release and operate production ready containerized applications on AWS App Runner or Amazon ECS on AWS Fargate.
k3s-aws-terraform-cluster - Deploy an high available K3s cluster on Amazon AWS