SaaSHub helps you find the best software and product alternatives Learn more →
Top 13 Julia Ode Projects
-
DifferentialEquations.jl
Multi-language suite for high-performance solvers of differential equations and scientific machine learning (SciML) components. Ordinary differential equations (ODEs), stochastic differential equations (SDEs), delay differential equations (DDEs), differential-algebraic equations (DAEs), and more in Julia.
-
ModelingToolkit.jl
An acausal modeling framework for automatically parallelized scientific machine learning (SciML) in Julia. A computer algebra system for integrated symbolics for physics-informed machine learning and automated transformations of differential equations
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
-
NeuralPDE.jl
Physics-Informed Neural Networks (PINN) Solvers of (Partial) Differential Equations for Scientific Machine Learning (SciML) accelerated simulation
-
OrdinaryDiffEq.jl
High performance ordinary differential equation (ODE) and differential-algebraic equation (DAE) solvers, including neural ordinary differential equations (neural ODEs) and scientific machine learning (SciML)
-
Catalyst.jl
Chemical reaction network and systems biology interface for scientific machine learning (SciML). High performance, GPU-parallelized, and O(1) solvers in open source software.
-
DataDrivenDiffEq.jl
Data driven modeling and automated discovery of dynamical systems for the SciML Scientific Machine Learning organization
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
SciMLSensitivity.jl
A component of the DiffEq ecosystem for enabling sensitivity analysis for scientific machine learning (SciML). Optimize-then-discretize, discretize-then-optimize, adjoint methods, and more for ODEs, SDEs, DDEs, DAEs, etc.
-
DiffEqBase.jl
The lightweight Base library for shared types and functionality for defining differential equation and scientific machine learning (SciML) problems
-
DiffEqGPU.jl
GPU-acceleration routines for DifferentialEquations.jl and the broader SciML scientific machine learning ecosystem
-
DiffEqDevTools.jl
Benchmarking, testing, and development tools for differential equations and scientific machine learning (SciML)
The documentation has a manifest associated with it: https://docs.sciml.ai/NeuralPDE/dev/#Reproducibility. Instantiating the manifest will give you all of the exact versions used for the documentation build (https://github.com/SciML/NeuralPDE.jl/blob/gh-pages/v5.7.0/assets/Manifest.toml). You just ]instantiate folder_of_manifest. Or you can use the Project.toml.
There has been a lot of research in Runge Kutta methods in the last couple decades which resulted in all kind of specialized Runge Kutta methods. You have high order ones, RK methods for stiff problems, embedded RK methods which benefit from adaprive step size control, RK-Nystrom methods for second order Problems, symplectic RK methods which preserve energy (eg. hamiltonian) ando so on. If you are interested in the numerics and the use cases I highly recommend checking out the Julia Libary OrdinaryDiffEq (https://github.com/SciML/OrdinaryDiffEq.jl). If you look into the documentation you find A LOT of implemented RK methods for all kind of use cases.
Indeed, and this year we created a system for compiling ODE code not just optimized CUDA kernels but also OneAPI kernels, AMD GPU kernels, and Metal. Peer reviewed version is here (https://www.sciencedirect.com/science/article/abs/pii/S00457...), open access is here (https://arxiv.org/abs/2304.06835), and the open source code is at https://github.com/SciML/DiffEqGPU.jl. The key that the paper describes is that in this case kernel generation is about 20x-100x faster than PyTorch and Jax (see the Jax compilation in multiple ways in this notebook https://colab.research.google.com/drive/1d7G-O5JX31lHbg7jTzz..., extra overhead though from calling Julia from Python but still shows a 10x).
The point really is that while deep learning libraries are amazing, at the end of the day they are DSL and really pull towards one specific way of computing and parallelization. It turns out that way of parallelizing is good for deep learning, but not for all things you may want to accelerate. Sometimes (i.e. cases that aren't dominated by large linear algebra) building problem-specific kernels is a major win, and it's over-extrapolating to see ML frameworks do well with GPUs and think that's the only thing that's required. There are many ways to parallelize a code, ML libraries hardcode a very specific way, and it's good for what they are used for but not every problem that can arise.
Project mention: Julia as a unifying end-to-end workflow language on the Frontier exascale system | news.ycombinator.com | 2023-11-19
Julia Ode related posts
- 2023 was the year that GPUs stood still
- Julia as a unifying end-to-end workflow language on the Frontier exascale system
- Modern Numerical Solving methods
- Why Fortran is a scientific powerhouse
- Mathematically Modelling a PRV
- How much useful are Runge-Kutta methods of order 9 and higher within double-precision arithmetic/floating point accuracy?
- Interpolant Coefficients for the BS5 Runge-Kutta method
-
A note from our sponsor - SaaSHub
www.saashub.com | 23 Apr 2024
Index
What are some of the best open-source Ode projects in Julia? This list will help you:
Project | Stars | |
---|---|---|
1 | DifferentialEquations.jl | 2,754 |
2 | ModelingToolkit.jl | 1,333 |
3 | NeuralPDE.jl | 901 |
4 | Latexify.jl | 529 |
5 | OrdinaryDiffEq.jl | 498 |
6 | Catalyst.jl | 419 |
7 | DataDrivenDiffEq.jl | 398 |
8 | SciMLSensitivity.jl | 308 |
9 | DiffEqBase.jl | 295 |
10 | DiffEqGPU.jl | 267 |
11 | SciMLStyle | 192 |
12 | ModelingToolkitStandardLibrary.jl | 98 |
13 | DiffEqDevTools.jl | 46 |
Sponsored