Our great sponsors
-
An Evolutionary Approach to Dynamic Introduction of Tasks in Large-scale Multitask Learning Systems - Google 2022 – Pathways - Jeff Dean! - Network grows with amount of tasks and data! Paper: https://arxiv.org/abs/2205.12755 Github: https://github.com/google-research/google-research/tree/master/muNet
-
EfficientZero
Open-source codebase for EfficientZero, from "Mastering Atari Games with Limited Data" at NeurIPS 2021.
Mastering Atari Games with Limited Data – EfficientZero ( Human sample -efficiency! ) Paper: https://arxiv.org/abs/2111.00210 Lesswrong article about the paper: https://www.lesswrong.com/posts/mRwJce3npmzbKfxws/efficientzero-how-it-works Github: https://github.com/YeWR/EfficientZero
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
-
msn
Discontinued Masked Siamese Networks for Label-Efficient Learning (https://arxiv.org/abs/2204.07141)
Masked Siamese Networks for Label-Efficient Learning similar to EfficientZero in structure Paper: https://arxiv.org/abs/2204.07141 Github: https://github.com/facebookresearch/msn
-
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness Paper: https://arxiv.org/abs/2205.14135 Github: https://github.com/HazyResearch/flash-attention and https://github.com/lucidrains/flash-attention-jax
-
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness Paper: https://arxiv.org/abs/2205.14135 Github: https://github.com/HazyResearch/flash-attention and https://github.com/lucidrains/flash-attention-jax
-
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
DeepSpeed Inference: Enabling Efficient Inference of Transformer Models at Unprecedented Scale - Microsoft 2022 Paper: https://arxiv.org/pdf/2207.00032.pdf Github: https://github.com/microsoft/DeepSpeed
-
General-purpose, long-context autoregressive modeling with Perceiver AR - Deepmind 2022 Paper: https://arxiv.org/abs/2202.07765 Deepmind: https://www.deepmind.com/publications/perceiver-ar-general-purpose-long-context-autoregressive-generation Code: https://github.com/google-research/perceiver-ar
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
RHO-LOSS - Prioritized Training on Points that are Learnable, Worth Learning, and Not Yet Learnt - Trains Models 18x faster with higher accuracy Paper: https://arxiv.org/abs/2206.07137 Github: https://github.com/OATML/RHO-Loss
-
CodeRL
This is the official code for the paper CodeRL: Mastering Code Generation through Pretrained Models and Deep Reinforcement Learning (NeurIPS22).
CodeRL: Mastering Code Generation through Pretrained Models and Deep Reinforcement Learning Paper: https://arxiv.org/pdf/2207.01780.pdf Github: https://github.com/salesforce/CodeRL
-
XMem: Long-Term Video Object Segmentation with an Atkinson-Shiffrin Memory Model ( Added because of the Atkinson-Shiffrin Memory Model ) Paper: https://arxiv.org/abs/2207.07115 Github: https://github.com/hkchengrex/XMem