LMOps
unilm
LMOps | unilm | |
---|---|---|
6 | 40 | |
3,186 | 18,358 | |
2.7% | 1.5% | |
8.1 | 9.0 | |
14 days ago | 8 days ago | |
Python | Python | |
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
LMOps
-
Has anyone found a way to use Microsoft's Promptist in Stable Diffusion?
Microsoft released an open-sourced prompt optimizer. Has anyone used https://github.com/microsoft/LMOps for prompt optimization? They have it in this demo: https://huggingface.co/spaces/microsoft/Promptist, I'm not sure how to use it for 1.5.
-
🚀 Microsoft Open Source LMOps: An AI Prompt Optimization Toolkit For Generative AI Models
Quick Read: https://www.marktechpost.com/2023/02/14/microsoft-open-source-lmops-an-ai-prompt-optimization-toolkit-for-generative-ai-models/ Paper: https://arxiv.org/pdf/2212.09611.pdf Github: https://github.com/microsoft/LMOps
- General technology for enabling AI capabilities with LLMs and Generative models
- microsoft/LMOps: General technology for enabling AI capabilities w/ LLMs and Generative AI models
-
Microsoft Promptist : Optimising Stable Diffusion prompts via a language model fine-tuned with reinforcement learning
Project Page : https://github.com/microsoft/LMOps/tree/main/promptist
unilm
- The Era of 1-Bit LLMs: Training_Tips, Code And_FAQ [pdf]
- The Era of 1-Bit LLMs: Training Tips, Code and FAQ
-
The Era of 1-bit LLMs: ternary parameters for cost-effective computing
+1 On this, the real proof would have been testing both models side-by-side.
It seems that it may be published on GitHub [1] according to HuggingFace [2].
[1] https://github.com/microsoft/unilm/tree/master/bitnet
[2] https://huggingface.co/papers/2402.17764
- I'm an Old Fart and AI Makes Me Sad
-
On building a semantic search engine
e5-mistral is essentially a distillation from gpt-4 to a smaller model. You can see here https://github.com/microsoft/unilm/blob/16da2f193b9c1dab0a69...
they actually have custom prompts for each dataset being tested.
Question would be, if you haven't seen the task before, what is a good prompt to prepend for your task?
IMO e5-mistral is overfit to MTEB
-
Leveraging GPT-4 for PDF Data Extraction: A Comprehensive Guide
Layout LM v1, v2 and v3 models [ Github ] DocBERT [ Github ]
-
Microsoft Publishes LongNet: Scaling Transformers to 1,000,000,000 Tokens
The repository is available here.
-
Recommended open LLMs with image input modality?
It is missing kosmos-2. I remember its image captioning was(demo currently down) really good and it's almost as fast as llava and lavin.
-
LongNet: Scaling Transformers to 1,000,000,000 Tokens
Should be this: https://github.com/microsoft/unilm/
-
[R] LongNet: Scaling Transformers to 1,000,000,000 Tokens
This is from Microsoft Research (Asia). https://aka.ms/GeneralAI
What are some alternatives?
langchain - ⚡ Building applications with LLMs through composability ⚡ [Moved to: https://github.com/langchain-ai/langchain]
transformers - 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
langchain - 🦜🔗 Build context-aware reasoning applications
ERNIE - Official implementations for various pre-training models of ERNIE-family, covering topics of Language Understanding & Generation, Multimodal Understanding & Generation, and beyond.
Awesome-Efficient-LLM - A curated list for Efficient Large Language Models
involution - [CVPR 2021] Involution: Inverting the Inherence of Convolution for Visual Recognition, a brand new neural operator
lora-instruct - Finetune Falcon, LLaMA, MPT, and RedPajama on consumer hardware using PEFT LoRA
gensim - Topic Modelling for Humans
CUPCAKEAGI - 🧁👋 Welcome to CupcakeAGI, where we bake up some sweet and creamy AGI goodness! 🍰🤖
maelstrom - A workbench for writing toy implementations of distributed systems.
ImageNet21K - Official Pytorch Implementation of: "ImageNet-21K Pretraining for the Masses"(NeurIPS, 2021) paper
rasa - 💬 Open source machine learning framework to automate text- and voice-based conversations: NLU, dialogue management, connect to Slack, Facebook, and more - Create chatbots and voice assistants