Our great sponsors
-
eigenlearning
codebase for "A Theory of the Inductive Bias and Generalization of Kernel Regression and Wide Neural Networks"
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
Fwiw (and shameless plug disclaimer), I made a library for easily toying around with evolution based NAS. It uses mixed integer programming to align the neurons (e.g. when resizing a layer) which it gives much higher degrees of freedom over what architectures one can allow in the search space over anything else I have seen: https://github.com/DrChainsaw/NaiveGAflux.jl
In addition to using a validation set, there is research w.r.t. the neural tangent kernel which claims that metrics correlated to both training speed and generalization can be computed from the NTK. This would then remove/reduce the need for training as one can substitute it with computing the NTK. I don’t have a complete list of references, but here is one example and another where they have applied NTK (and number of linear regions) to NAS.
Related posts
- Julia 1.10 Released
- [D] ClosedAI license, open-source license which restricts only OpenAI, Microsoft, Google, and Meta from commercial use
- [D] Why are we stuck with Python for something that require so much speed and parallelism (neural networks)?
- Deep Learning With Flux: Loss Doesn't Converge
- Flux: The Elegant Machine Learning Stack