Enzyme.jl VS Nim

Compare Enzyme.jl vs Nim and see what are their differences.

Nim

Nim is a statically typed compiled systems programming language. It combines successful concepts from mature languages like Python, Ada and Modula. Its design focuses on efficiency, expressiveness, and elegance (in that order of priority). (by nim-lang)
InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
Enzyme.jl Nim
10 347
401 16,079
2.7% 0.5%
9.5 9.9
about 6 hours ago 4 days ago
Julia Nim
MIT License GNU General Public License v3.0 or later
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

Enzyme.jl

Posts with mentions or reviews of Enzyme.jl. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2022-06-25.
  • Custom gradients in Enzyme
    1 project | /r/Julia | 27 Nov 2022
    It's possible but at this time it's not recommended or documented as right now it requires writing some LLVM-level stuff and a better system is coming soon (see https://github.com/EnzymeAD/Enzyme.jl/pull/177)
  • “Why I still recommend Julia”
    11 projects | news.ycombinator.com | 25 Jun 2022
    Can you point to a concrete example of one that someone would run into when using the differential equation solvers with the default and recommended Enzyme AD for vector-Jacobian products? I'd be happy to look into it, but there do not currently seem to be any correctness issues in the Enzyme issue tracker that are current (3 issues are open but they all seem to be fixed, other than https://github.com/EnzymeAD/Enzyme.jl/issues/278 which is actually an activity analysis bug in LLVM). So please be more specific. The issue with Enzyme right now seems to moreso be about finding functional forms that compile, and it throws compile-time errors in the event that it cannot fully analyze the program and if it has too much dynamic behavior (example: https://github.com/EnzymeAD/Enzyme.jl/issues/368).

    Additional note, we recently did a overhaul of SciMLSensitivity (https://sensitivity.sciml.ai/dev/) and setup a system which amounts to 15 hours of direct unit tests doing a combinatoric check of arguments with 4 hours of downstream testing (https://github.com/SciML/SciMLSensitivity.jl/actions/runs/25...). What that identified is that any remaining issues that can arise are due to the implicit parameters mechanism in Zygote (Zygote.params). To counteract this upstream issue, we (a) try to default to never default to Zygote VJPs whenever we can avoid it (hence defaulting to Enzyme and ReverseDiff first as previously mentioned), and (b) put in a mechanism for early error throwing if Zygote hits any not implemented derivative case with an explicit error message (https://github.com/SciML/SciMLSensitivity.jl/blob/v7.0.1/src...). We have alerted the devs of the machine learning libraries, and from this there has been a lot of movement. In particular, a globals-free machine learning library, Lux.jl, was created with fully explicit parameters https://lux.csail.mit.edu/dev/, and thus by design it cannot have this issue. In addition, the Flux.jl library itself is looking to do a redesign that eliminates implicit parameters (https://github.com/FluxML/Flux.jl/issues/1986). Which design will be the one in the end, that's uncertain right now, but it's clear that no matter what the future designs of the deep learning libraries will fully cut out that part of Zygote.jl. And additionally, the other AD libraries (Enzyme and Diffractor for example) do not have this "feature", so it's an issue that can only arise from a specific (not recommended) way of using Zygote (which now throws explicit error messages early and often if used anywhere near SciML because I don't tolerate it).

    So from this, SciML should be rather safe and if not, please share some details and I'd be happy to dig in.

  • The Julia language has a number of correctness flaws
    19 projects | news.ycombinator.com | 16 May 2022
    Lots of things are being rewritten. Remember we just released a new neural network library the other day, SimpleChains.jl, and showed that it gave about a 10x speed improvement on modern CPUs with multithreading enabled vs Jax Equinox (and 22x when AVX-512 is enabled) for smaller neural network and matrix-vector types of cases (https://julialang.org/blog/2022/04/simple-chains/). Then there's Lux.jl fixing some major issues of Flux.jl (https://github.com/avik-pal/Lux.jl). Pretty much everything is switching to Enzyme which improves performance quite a bit over Zygote and allows for full mutation support (https://github.com/EnzymeAD/Enzyme.jl). So an entire machine learning stack is already seeing parts release.

    Right now we're in a bit of an uncomfortable spot where we have to use Zygote for a few things and then Enzyme for everything else, but the custom rules system is rather close and that's the piece that's needed to make the full transition.

  • Engineering Trade-Offs in Automatic Differentiation: from TensorFlow and PyTorch to Jax and Julia
    1 project | /r/Julia | 26 Dec 2021
    enzyme.jl is probably the quickest way to play with enzyme: https://github.com/wsmoses/Enzyme.jl
  • Useful Algorithms That Are Not Optimized by Jax, PyTorch, or TensorFlow
    2 projects | news.ycombinator.com | 22 Jul 2021
    "Maybe they let you declare some subgraph as 'dynamic' to avoid static optimizations?" What you just described is Tensorflow Eager and why it has some performance issues. XLA makes some pretty strong assumptions and I don't that should change. Tensorflow's ability to automatically generate good parallelized production code stems from the restrictions it has imposed. So I wouldn't even try for a "one true AD to rule them all" since making things more flexible will reduce the amount of compiler optimizations that can be automatically performed.

    To get the more flexible form, you really would want to do it in a way that uses a full programming language's IR as its target. I think trying to use a fully dynamic programming language IR directly (Python, R, etc.) directly would be pretty insane because it would be hard to enforce rules and get performance. So some language that has a front end over an optimizing compiler (LLVM) would probably make the most sense. Zygote and Diffractor uses Julia's IR, but there are other ways to do this as well. Enzyme (https://github.com/wsmoses/Enzyme.jl) uses the LLVM IR directly for doing source-to-source translations. Using some dialect of LLVM (provided by MLIR) might be an interesting place to write a more ML-focused flexible AD system. Swift for Tensorflow used the Swift IR. This mindset starts to show why those tools were chosen.

  • Julia Computing Raises $24M Series A
    5 projects | news.ycombinator.com | 19 Jul 2021
    Have you explored the SciML landscape at all (?):

    https://sciml.ai/

    There are a number of components here which enable (what I would call) the expression of more advanced models using Julia's nice compositional properties.

    Flux.jl is of course what most people would think of here (one of Julia's deep learning frameworks). But the reality behind Flux.jl is that it is just Julia code -- nothing too fancy.

    There's ongoing work for AD in several directions -- including a Julia interface to Enzyme: https://github.com/wsmoses/Enzyme.jl

    Also, a new AD system which Keno (who you'll see comment below or above) has been working on -- see Diffractor.jl on the JuliaCon schedule (for example).

    Long story short -- there's quite a lot of work going on.

    It may not seem like there is a "unified" package -- but that's because packages compose so well together in Julia, there's really no need for that.

  • Swift for TensorFlow Shuts Down
    13 projects | news.ycombinator.com | 12 Feb 2021
    The name of the LLVM AD tool is actually Enzyme [http://enzyme.mit.edu/] (Zygote is a Julia tool)
  • Enzyme – High-performance automatic differentiation of LLVM (r/MachineLearning)
    1 project | /r/datascienceproject | 8 Feb 2021
    1 project | /r/datascienceproject | 7 Feb 2021
  • Enzyme – High-performance automatic differentiation of LLVM
    3 projects | news.ycombinator.com | 4 Feb 2021
    Also see the Julia package that makes it acessible with a high level interface and probably one of the easier ways to play with it: https://github.com/wsmoses/Enzyme.jl.

Nim

Posts with mentions or reviews of Nim. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-04-26.
  • 3 years of fulltime Rust game development, and why we're leaving Rust behind
    21 projects | news.ycombinator.com | 26 Apr 2024
  • Top Paying Programming Technologies 2024
    19 projects | dev.to | 6 Mar 2024
    22. Nim - $80,000
  • "14 Years of Go" by Rob Pike
    2 projects | news.ycombinator.com | 27 Feb 2024
    I think the right answer to your question would be NimLang[0]. In reality, if you're seeking to use this in any enterprise context, you'd most likely want to select the subset of C++ that makes sense for you or just use C#.

    [0]https://nim-lang.org/

  • Odin Programming Language
    23 projects | news.ycombinator.com | 1 Jan 2024
  • Ask HN: Interest in a Rust-Inspired Language Compiling to JavaScript?
    5 projects | news.ycombinator.com | 24 Dec 2023
    I don't think it's a rust-inspired language, but since it has strong typing and compiles to javascript, did you give a look at nim [0] ?

    For what it takes, I find the language very expressive without the verbosity in rust that reminds me java. And it is also very flexible.

    [0] : https://nim-lang.org/

  • The nim website and the downloads are insecure
    1 project | /r/nim | 11 Dec 2023
    I see a valid cert for https://nim-lang.org/
  • Nim
    5 projects | news.ycombinator.com | 6 Dec 2023
    FYI, on the front page, https://nim-lang.org, in large type you have this:

    > Nim is a statically typed compiled systems programming language. It combines successful concepts from mature languages like Python, Ada and Modula.

  • Things I've learned about building CLI tools in Python
    16 projects | news.ycombinator.com | 24 Oct 2023
    You better off with using a compiled language.

    If you interested in a language that's compiled, fast, but as easy and pleasant as Python - I'd recommend you take a look at [Nim](https://nim-lang.org).

    And to prove what Nim's capable of - here's a cool repo with 100+ cli apps someone wrote in Nim: [c-blake/bu](https://github.com/c-blake/bu)

  • Mojo is now available on Mac
    13 projects | news.ycombinator.com | 19 Oct 2023
    Chapel has at least several full-time developers at Cray/HPE and (I think) the US national labs, and has had some for almost two decades. That's much more than $100k.

    Chapel is also just one of many other projects broadly interested in developing new programming languages for "high performance" programming. Out of that large field, Chapel is not especially related to the specific ideas or design goals of Mojo. Much more related are things like Codon (https://exaloop.io), and the metaprogramming models in Terra (https://terralang.org), Nim (https://nim-lang.org), and Zig (https://ziglang.org).

    But Chapel is great! It has a lot of good ideas, especially for distributed-memory programming, which is its historical focus. It is more related to Legion (https://legion.stanford.edu, https://regent-lang.org), parallel & distributed Fortran, ZPL, etc.

  • NIR: Nim Intermediate Representation
    1 project | /r/hackernews | 2 Oct 2023

What are some alternatives?

When comparing Enzyme.jl and Nim you can also consider the following projects:

ChainRules.jl - forward and reverse mode automatic differentiation primitives for Julia Base + StdLibs

zig - General-purpose programming language and toolchain for maintaining robust, optimal, and reusable software.

ForwardDiff.jl - Forward Mode Automatic Differentiation for Julia

go - The Go programming language

MLJ.jl - A Julia machine learning framework

Odin - Odin Programming Language

swift - Swift for TensorFlow

rust - Empowering everyone to build reliable and efficient software.

Lux.jl - Explicitly Parameterized Neural Networks in Julia

crystal - The Crystal Programming Language

NBodySimulator.jl - A differentiable simulator for scientific machine learning (SciML) with N-body problems, including astrophysical and molecular dynamics

v - Simple, fast, safe, compiled language for developing maintainable software. Compiles itself in <1s with zero library dependencies. Supports automatic C => V translation. https://vlang.io