pykale
open_flamingo
Our great sponsors
pykale | open_flamingo | |
---|---|---|
2 | 4 | |
427 | 3,459 | |
1.6% | 3.6% | |
9.1 | 7.8 | |
about 1 month ago | about 1 month ago | |
Python | Python | |
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
pykale
-
PyKale Preprint: Knowledge-Aware Machine Learning from Multiple Sources in Python [P][R]
A 10-page preprint is on arXiv to describe the green machine learning design principles behind our pipeline-based API below as well as features and examples in our PyKale library for multimodal learning and transfer learning with deep learning and dimensionality reduction on graphs, images, texts, and videos to enable and accelerate interdisciplinary research: [2106.09756] PyKale: Knowledge-Aware Machine Learning from Multiple Sources in Python (arxiv.org)
-
[P] An introduction to PyKale https://github.com/pykale/pykale, a PyTorch library that provides a unified pipeline-based API for knowledge-aware multimodal learning and transfer learning on graphs, images, texts, and videos to accelerate interdisciplinary research. Welcome feedback/contribution!
Thank you LargeYellowBus for your valuable feedback! That's really helpful. It is weird to hear that the GitHub link was broken when you tried. It is here https://github.com/pykale/pykale and there have been 100+ unique visitors today. So it might be temporary on your side and you may try again.
open_flamingo
-
Are there any multimodal AI models I can use to provide a paired text *and* image input, to then generate an expanded descriptive text output? [D]
Maybe the recent OpenFlamingo gives you better results (they have a demo on HF).
- [D] Multi modal for visual qna based on a given image. Need suggestions.
- Open Flamingo: An open-source framework for training large multimodal models
-
Announcing OpenFlamingo: An open-source framework for training vision-language models with in-context learning | LAION
Code here: https://github.com/mlfoundations/open_flamingo
What are some alternatives?
EasyOCR - Ready-to-use OCR with 80+ supported languages and all popular writing scripts including Latin, Chinese, Arabic, Devanagari, Cyrillic and etc.
transformers - 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
AdaTime - [TKDD 2023] AdaTime: A Benchmarking Suite for Domain Adaptation on Time Series Data
Emu - Emu Series: Generative Multimodal Models from BAAI
Multimodal-Toolkit - Multimodal model for text and tabular data with HuggingFace transformers as building block for text data
speechbrain - A PyTorch-based Speech Toolkit
Meta-SelfLearning - Meta Self-learning for Multi-Source Domain Adaptation: A Benchmark
icl-ceil - [ICML 2023] Code for our paper “Compositional Exemplars for In-context Learning”.
social-balance - A library-agnostic project for calculating exactly and efficiently social balance, based on the Aref, Mason and Wilson paper (https://arxiv.org/abs/1611.09030)
jina - ☁️ Build multimodal AI applications with cloud-native stack
pytorch-adapt - Domain adaptation made easy. Fully featured, modular, and customizable.
valhalla-nmt - Code repository for CVPR 2022 paper "VALHALLA: Visual Hallucination for Machine Translation"