vertex-ai-samples
transformers.js
vertex-ai-samples | transformers.js | |
---|---|---|
24 | 26 | |
1,358 | 7,507 | |
4.0% | - | |
9.8 | 9.4 | |
about 20 hours ago | 8 days ago | |
Jupyter Notebook | JavaScript | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
vertex-ai-samples
- Gemini 1.5 outshines GPT-4-Turbo-128K on long code prompts, HVM author
-
Let's build your first ML app in Google Cloud Run
Google Cloud Platform (GCP) provides a very befitting Machine Learning solution called Vertex Ai that handles Google Cloud's unified platform for building, deploying, and managing machine learning (ML) models. Our goal is to build a simple Machine Learning application that optimizes all that GCP provides plus an implementation of continuous integration and continuous development (CI/CD).
-
Google Gemini Pro API Available Through AI Studio
Cross posting some links from another post that HNers found helpful
- https://cloud.google.com/vertex-ai (marketing page)
- https://cloud.google.com/vertex-ai/docs (docs entry point)
- https://console.cloud.google.com/vertex-ai (cloud console)
- https://console.cloud.google.com/vertex-ai/model-garden (all the models)
- https://console.cloud.google.com/vertex-ai/generative (studio / playground)
VertexAI is the umbrella for all of the Google models available through their cloud platform.
-
Google Imagen 2
For the peer comments
- https://cloud.google.com/vertex-ai (main page)
- https://cloud.google.com/vertex-ai/docs/start/introduction-u... (docs entry point)
- https://console.cloud.google.com/vertex-ai (cloud console)
-
Introducing Gemini: our largest and most capable AI model
Starting on December 13, developers and enterprise customers can access Gemini Pro via the Gemini API in Google AI Studio or Google Cloud Vertex AI.
-
How to Use AI/ML Models for Your Projects
Google Cloud Platform (https://cloud.google.com/vertex-ai): Conversely, Google Cloud Platform (GCP) provides a comprehensive suite of AI and machine learning services, including APIs for vision, language, conversation, and structured data analysis. Whether you're analyzing images, interpreting human speech, or diving deep into data patterns, GCP has something for you.
-
Create a ChatBot with VertexAI and LibreChat
VertexAI is a machine learning platform available on Google Cloud. It offers a variety of services to train and deploy AI models, including those for Generative AI.
- Tune PaLM 2 with your own RLHF training data
-
Any better alternatives to fine-tuning GPT-3 yet to create a custom chatbot persona based on provided knowledge for others to use?
Depending on how much work you want to put into it, you can get started at HuggingFace with their models and datasets, but you'd need compute power, multiple MLOps, etc. I was introduced to the concept in this video, since Google has their Vertex AI tools on Google Cloud, and there's always LangChain but I'm not sure about anything recent.
- Google Cloud Learning Machine
transformers.js
-
Transformers.js: Machine Learning for the Web
We have some other WebGPU demos, including:
- WebGPU embedding benchmark: https://huggingface.co/spaces/Xenova/webgpu-embedding-benchm...
- Real-time object detection: https://huggingface.co/spaces/Xenova/webgpu-video-object-det...
- Real-time background removal: https://huggingface.co/spaces/Xenova/webgpu-video-background...
- WebGPU depth estimation: https://huggingface.co/spaces/Xenova/webgpu-depth-anything
- Image background removal: https://huggingface.co/spaces/Xenova/remove-background-webgp...
You can follow the progress for full WebGPU support in the v3 development branch (https://github.com/xenova/transformers.js/pull/545).
To answer your question, while there are certain ops missing, the main limitation at the moment is for models with decoders... which are not very fast (yet) due to inefficient buffer reuse and many redundant copies between CPU and GPU. We're working closely with the ORT team to fix these issues though!
-
Deep Learning in JavaScript
BTW: you might want to add support for typed arrays.
See: https://github.com/xenova/transformers.js/blob/8804c36591d11...
This is really old, but added as part of the shape of the vector as well: https://github.com/nicolaspanel/numjs/blob/master/src/dtypes...
-
Deja-Vu your AI✦ Bookmarking Tool
Made possible by Xenova and Supabase / gte-small
-
Extracting YouTube video data with OpenAI and LangChain
To build the application, you’ll use the youtube-transcript package to retrieve YouTube video transcripts. You will then use LangChain and the Transformers.js package to generate free Hugging Face embeddings for the given transcript and store them in a vector store instead of relying on potentially expensive OpenAI embeddings. Lastly, you will use LangChain and an OpenAI model to retrieve information stored in the vector store.
- Transformers.js releases Zero-shot audio classification support
-
How to Use AI/ML Models for Your Projects
Transformers.js: A groundbreaking library, Transformers.js brings transformer models like GPT-3, BERT, and Whisper straight to your browser. With the introduction of technologies like webGPU and LLM, Transformers.js has garnered significant attention. If you’d like to learn how to integrate a small model in the UI, check out their code and examples here.
-
Show HN: Tiny LLMs – Browser-based private AI models for a wide array of tasks
The announcement seems somewhat disingenuous. The PR[1] found from their release notes[2] seems to contain only boilerplate and no real support for Mistral models or their weights.
[1]: https://github.com/xenova/transformers.js/pull/379
- Transformers.js
- Transformers.js: Run Machine Learning models directly in the browser
- What is the most cost-efficient way to have an embedding generator endpoint that is using an open-source embedding model? [Q]
What are some alternatives?
mlops-with-vertex-ai - An end-to-end example of MLOps on Google Cloud using TensorFlow, TFX, and Vertex AI
onnxruntime - ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
awesome-mlops - A curated list of references for MLOps
web-stable-diffusion - Bringing stable diffusion models to web browsers. Everything runs inside the browser with no server support.
MLflow - Open source platform for the machine learning lifecycle
web-ai - Run modern deep learning models in the browser.
VevestaX - 2 Lines of code to track ML experiments + EDA + check into Github
spark-nlp - State of the Art Natural Language Processing
jina - ☁️ Build multimodal AI applications with cloud-native stack
memory64 - Memory with 64-bit indexes
rasa - 💬 Open source machine learning framework to automate text- and voice-based conversations: NLU, dialogue management, connect to Slack, Facebook, and more - Create chatbots and voice assistants
openai-java - OpenAI Api Client in Java