Complete container management platform (by rancher)

Rancher Alternatives

Similar projects and alternatives to rancher

  • lens

    rancher VS lens

    Lens - The way the world runs Kubernetes

  • Portainer

    rancher VS Portainer

    Making Docker and Kubernetes management easy.

  • Zigi

    Delete the most useless function ever: context switching.. Zigi monitors Jira and GitHub updates, pings you when PRs need approval and lets you take fast actions - all directly from Slack! Plus it reduces cycle time by up to 75%.

  • k3s

    rancher VS k3s

    Lightweight Kubernetes

  • podman

    rancher VS podman

    Podman: A tool for managing OCI containers and pods.

  • microk8s

    rancher VS microk8s

    MicroK8s is a small, fast, single-package Kubernetes for developers, IoT and edge.

  • cluster-api

    Home for Cluster API, a subproject of sig-cluster-lifecycle

  • ingress-nginx

    Ingress-NGINX Controller for Kubernetes

  • Scout APM

    Truly a developer’s best friend. Scout APM is great for developers who want to find and fix performance issues in their applications. With Scout, we'll take care of the bugs so you can focus on building great things 🚀.

  • lima

    rancher VS lima

    Linux virtual machines, typically on macOS, for running containerd

  • Dokku

    rancher VS Dokku

    A docker-powered PaaS that helps you build and manage the lifecycle of applications

  • kops

    rancher VS kops

    Kubernetes Operations (kOps) - Production Grade k8s Installation, Upgrades and Management

  • kubesphere

    rancher VS kubesphere

    The container platform tailored for Kubernetes multi-cloud, datacenter, and edge management ⎈ 🖥 ☁️

  • k9s

    rancher VS k9s

    🐶 Kubernetes CLI To Manage Your Clusters In Style!

  • kubernetes

    rancher VS kubernetes

    Production-Grade Container Scheduling and Management

  • prometheus

    rancher VS prometheus

    The Prometheus monitoring system and time series database.

  • Nomad

    rancher VS Nomad

    Nomad is an easy-to-use, flexible, and performant workload orchestrator that can deploy a mix of microservice, batch, containerized, and non-containerized applications. Nomad is easy to operate and scale and has native Consul and Vault integrations.

  • kind

    rancher VS kind

    Kubernetes IN Docker - local clusters for testing Kubernetes

  • Flynn

    rancher VS Flynn

    [UNMAINTAINED] A next generation open source platform as a service (PaaS)

  • metallb

    rancher VS metallb

    A network load-balancer implementation for Kubernetes using standard routing protocols

  • harvester

    rancher VS harvester

    Open source hyperconverged infrastructure (HCI) software

  • Grafana

    rancher VS Grafana

    The open and composable observability and data visualization platform. Visualize metrics, logs, and traces from multiple sources like Prometheus, Loki, Elasticsearch, InfluxDB, Postgres and many more.

  • InfluxDB

    Build time-series-based applications quickly and at scale.. InfluxDB is the Time Series Data Platform where developers build real-time applications for analytics, IoT and cloud-native services in less time with less code.

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a better rancher alternative or higher similarity.

rancher reviews and mentions

Posts with mentions or reviews of rancher. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2022-08-30.
  • Terraform code for kubernetes on vsphere?
    3 projects | reddit.com/r/devops | 30 Aug 2022
    I don't know in which extend you plan to use Kubernetes in the future, but if it is aimed to become several huge production clusters, you should looks into Apps like Rancher: https://rancher.com
  • The Container Orchestrator Landscape
    8 projects | news.ycombinator.com | 24 Aug 2022
    This seems like a pretty well written overview!

    As someone who rather liked Docker Swarm (and still likes it, running my homelab and private cloud stuff on it), it is a bit sad to see it winding down like it, even though there were attempts to capitalize on the nice set of simple functionality that it brought to the table like CapRover: https://caprover.com/

    Even though there is still some nice software to manage installs of it, like Portainer: https://www.portainer.io/ (which also works for Kubernetes, like a smaller version of Rancher)

    Even though the resource usage is far lower than that of almost any Kubernetes distro that I've used (microk8s, K3s and K0s included), the Compose format being pretty much amazing for most smaller deployments and Compose still being one of the better ways to run things locally in addition to Swarm for remote deployments (Skaffold or other K8s local cluster solutions just feel complex in comparison).

    And yet, that's probably not where the future lies. Kubernetes won. Well, Nomad is also pretty good, admittedly.

    Though if you absolutely do need Kubernetes, personally I'd suggest that you look in the direction of Rancher for a simple UI to manage it, or at least drill down into the cluster state, without needing too much digging through a CLI: https://rancher.com/

    Lots of folks actually like k9s as well, if you do like the TUI approach a bit more: https://k9scli.io/

    But for the actual clusters, assuming that you ever want to self-host one, ideally a turnkey solution, RKE is good, K0s is also promising, but personally I'd go with K3s: https://k3s.io/ which has been really stable on DEB distros and mostly works okay on RPM ones (if you cannot afford OpenShift or to wait for MicroShift), with my only pet peeve being that the Traefik ingress is a little bit under-documented (e.g. how to configure common use cases, like a SSL certificate, one with an intermediate certificate, maybe a wildcard, or perhaps just use Let's Encrypt, how to set defaults vs defining them per domain).

    For the folks with thicker wallets, though, I'd suggest to just give in and pay someone to run a cluster for you: that way you'll get something vaguely portable, will make lots of the aspects in regards to running it someone else's problem and will be able to leverage the actual benefits of working with the container orchestrator.

    > To extend its reach across multiple hosts, Docker introduced Swarm mode in 2016. This is actually the second product from Docker to bear the name "Swarm" — a product from 2014 implemented a completely different approach to running containers across multiple hosts, but it is no longer maintained. It was replaced by SwarmKit, which provides the underpinnings of the current version of Docker Swarm.

    On an unrelated note, this, at least to me, feels like pretty bad naming and management of the whole initiative, though. Of course, if the features are there, it shouldn't be enough to scare anyone away from the project, but at the same time it could have been a bit simpler.

  • I want to provide some free support for community, how should I start?
    2 projects | reddit.com/r/devops | 3 Aug 2022
    But I think once you have a good understanding of K8S internal (components, how thing work underlying, etc.), you can use some tool to help you provision / maintain k8s cluster easier (look for https://rancher.com/ and alternatives).
  • Rancher monitoring v1 to v2 upgrade fails with "V1 should be disabled but the operator is still being deployed"
    2 projects | reddit.com/r/rancher | 11 Jul 2022
    Monitoring V1 should be disabled but the operator is still being deployed. Please file a bug with Rancher at https://github.com/rancher/rancher/issues/new.
  • Ask HN: What is your Kubernetes nightmare?
    8 projects | news.ycombinator.com | 27 Jun 2022
    Late to the party, but figured I'd share my own story (some details obviously changed, but hopefully the spirit of the experience remains).

    Suppose that you work in an org that successfully ships software in a variety of ways - as regular packaged software that runs on an OS directly (e.g. a .jar that expects a certain JDK version in the VM), or maybe even uses containers sometimes, be it with Nomad, Swarm or something else.

    And then a project comes along that needs Kubernetes, because someone else made that choice for you (in some orgs, it might be a requirement from the side of clients, others might want to be able to claim that their software runs on Kubernets, in other cases some dev might be padding their CV and leave) and now you need to deal with its consequences.

    But here's the thing - if the organization doesn't have enough buy-in into Kubernetes, it's as if you're starting everything from 0, especially if paying some cloud vendor to give you a managed cluster isn't in the cards, be it because of data storage requirements (even for dev environments), other compliance reasons or even just corporate policy.

    So, I might be given a single VM on a server, with 8 GB of RAM for launching 4 or so Java/.NET services, as that is a decent amount of resources for doing things the old way. But now, I need to fit a whole Kubernetes cluster in there, which in most configurations eats resources like there's no tomorrow. Oh, and the colleagues also don't have too much experience working with Kubernetes, so some sort of a helpful UI might be nice to have, except that the org uses RPM distros and there are no resources for an install of OpenShift on that VM.

    But how much can I even do with that amount of resources, then? Well, I did manage to get K3s (a certified K8s distro by Rancher) up and running, though my hopes of connecting it with the actual Rancher tool (https://rancher.com/) to act as a good web UI didn't succeed. Mostly because of some weirdness with the cgroups support and Rancher running as a Docker container in many cases, which just kind of broke. I did get Portainer (https://www.portainer.io/) up and running instead, but back then I think there were certain problems with the UI, as it's still very much in active development and gradually receives lots of updates. I might have just gone with Kubernetes dashboard, but admittedly the whole login thing isn't quite as intuitive as the alternatives.

    That said, everything kind of broke down for a bit as I needed to setup the ingress. What if you have a wildcard certificate along the lines of .something.else.org.com and want it to be used for all of your apps? Back in the day, you'd just setup Nginx or Apache as your reverse proxy and let it worry about SSL/TLS termination. A duty which is now taken over by Kubernetes, except that by default K3s comes with Traefik as their ingress controller of choice and the documentation isn't exactly stellar.

    So for getting this sort of configuration up and running, I needed to think about a HelmChartConfig for Traefik, a ConfigMap which references the secrets, a TLSStore to contain them, as well as creating the actual tls-secrets themselves with the appropriate files off of the file system, which still feels a bit odd and would probably be an utter mess to get particular certificates up and running for some other paths, as well as Let's Encrypt for other ones yet. In short, what previously would have been those very same files living on the file system and a few (dozen?) lines inside of the reverse proxy configuration, is now a distributed mess of abstractions and actions which certainly need some getting used to.

    Oh, and Portainer sometimes just gets confused and fails to figure out how to properly setup the routes, though I do have to say that at least MetalLB does its job nicely.

    And then? Well, we can't just ship manifests directly, we also need Helm charts! But of course, in addition to writing those and setting up the CI for packaging them, you also need something running to store them, as well as any Docker images that you want. In lieu of going through all of the red tape to set that up on shared infrastructure (which would need cleanup policies, access controls and lots of planning so things don't break for other parties using it), instead I crammed in an instance of Nexus/Artifactory/Harbor/... on that very same server, with the very same resource limits, with deadlines still looming over my head.

    But that's not it, for software isn't developed in a vacuum. Throw in all of the regular issues with developing software, like not being 100% clear on each of the configuration values that the apps need (because developers are fallible, of course), changes to what they want to use, problems with DB initialization (of course, still needing an instance of PostgreSQL/MariaDB running on the very same server, which for whatever reason might get used as a shared DB) and so on.

    In short, you take a process that already has pain points in most orgs and make it needlessly more complex. There are tangible benefits for using Kubernetes. Once you find a setup that works (personally, Ubuntu LTS or a similar distro, full Rancher install, maybe K3s as the underlying cluster or RKE/K3s/k0s on separate nodes, with Nginx for ingress, or a 100% separately managed ingress) then it's great and the standardization is almost like a superpower (as long as you don't go crazy with CRDs). Yet, you need to pay a certain cost up front.

    What could be done to alleviate some of the pain points?

    In short, I think that:

      - expect to need a lot more resources than previously: always have a separate node for managing your cluster and put any sorts of tools on it as well (like Portainer/Rancher), but run your app workloads on other nodes (K3s or k0s can still be not too demanding with resources for the most part)
  • Don't Use Kubernetes, Yet
    7 projects | news.ycombinator.com | 18 Jun 2022
    A few years, I would have said no. Now, I'm cautiously optimistic about it.

    Personally, I think that you can use something like Rancher (https://rancher.com/) or Portainer (https://www.portainer.io/) for easier management and/or dashboard functionality, to make the learning curve a bit more approachable. For example, you can create a deployment through the UI by following a wizard that also offers you configuration that you might want to use (e.g. resource limits) and then later retrieve the YAML manifest, should you wish to do that. They also make interacting with Helm charts (pre-made packages) more easy.

    Furthermore, there are certified distributions which are not too resource hungry, especially if you need to self-host clusters, for example K3s (https://k3s.io/) and k0s (https://k0sproject.io/) are both production ready up to a certain scale, don't consume a lot of memory, are easy to setup and work with whilst being mostly OS agnostic (DEB distros will always work best, RPM ones have challenges as soon as you look elsewhere instead of at OpenShift, which is probably only good for enterprises).

    If you can automated cluster setup with Ansible and treat the clusters as something that you can easily re-deploy when you inevitably screw up (you might not do that, but better to plan for failure), you should be good! Even Helm charts have gotten pretty easy to write and deploy and K8s works nicely with most CI/CD tools out there, given that kubectl lends itself pretty well to scripting.

  • Building an Internal Kubernetes Platform
    6 projects | dev.to | 16 Jun 2022
    Alternatively, it is also possible to use a multi-cloud or hybrid-cloud approach, which combines several cloud providers or even public and private clouds. Special tools such as Rancher and OpenShift can be very useful to run this type of system.
  • Five Dex Alternatives for Kubernetes Authentication
    6 projects | dev.to | 16 Jun 2022
    Rancher provides a Rancher authentication proxy that allows user authentication from a central location. With this proxy, you can set the credential for authenticating users that want to access your Kubernetes clusters. You can create, view, update, or delete users through Rancher’s UI and API.
  • I WANT TO LEARN. Roast me/ Humble me if need be. It's the only way.
    2 projects | reddit.com/r/HomeServer | 27 May 2022
    Not sure if it works on arm but I heard that rancher and harvester are a good combo for clustering k3s and kvm if you want to try both technologies. https://rancher.com/ https://harvesterhci.io/
  • Opinion about k3s
    2 projects | reddit.com/r/kubernetes | 24 May 2022
    - https://github.com/rancher/rancher/issues/16454
  • A note from our sponsor - SonarLint
    www.sonarlint.org | 1 Dec 2022
    Up your coding game and discover issues early. SonarLint is a free plugin that helps you find & fix bugs and security issues from the moment you start writing code. Install from your favorite IDE marketplace today. Learn more →


Basic rancher repo stats
6 days ago
Clean code begins in your IDE with SonarLint
Up your coding game and discover issues early. SonarLint is a free plugin that helps you find & fix bugs and security issues from the moment you start writing code. Install from your favorite IDE marketplace today.