-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
Have anyone tried to optimize the forward and backward using custom Cuda code or fused kernel to speed up the training time of current LLMs? I only have seen FasterTransformer ( NVIDIA/FasterTransformer) and other similar tools but they're only focusing on inference.
NOTE:
The number of mentions on this list indicates mentions on common posts plus user suggested alternatives.
Hence, a higher number means a more popular project.
Related posts
-
Why are self attention not as deployment friendly?
-
Whether the ML computation engineering expertise will be valuable, is the question.
-
Lack of activation in transformer feedforward layer?
-
How to Build an AI Text Generator: Text Generation with a GPT-2 Model
-
AI leaderboards are no longer useful. It's time to switch to Pareto curves