MiniCPM-Llama3-V 2.5: A GPT-4V Level Multimodal LLM on Your Phone
-
Updated
Jun 12, 2024 - Python
MiniCPM-Llama3-V 2.5: A GPT-4V Level Multimodal LLM on Your Phone
Start building LLM-empowered multi-agent applications in an easier way.
Transformer Based Model Paper Review and PyTorch Code
Generative Model Paper Review and PyTorch Code
Unified embedding generation and search engine. Also available on cloud - cloud.marqo.ai
GPT4V-level open-source multi-modal model based on Llama3-8B
[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4V. 接近GPT-4V表现的可商用开源多模态对话模型
Here we will track the latest AI Multimodal Models, including Multimodal Foundation Models, LLM, Agent, Audio, Image, Video, Music and 3D content. 🔥
A one-stop data processing system to make data higher-quality, juicier, and more digestible for LLMs! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷为大语言模型提供更高质量、更丰富、更易”消化“的数据!
VisualRWKV is the visual-enhanced version of the RWKV language model, enabling RWKV to handle various visual tasks.
ModelScope: bring the notion of Model-as-a-Service to life.
Represent, send, store and search multimodal data
Open-source evaluation toolkit of large vision-language models (LVLMs), support GPT-4v, Gemini, QwenVLPlus, 50+ HF models, 20+ benchmarks
[EMNLP 2022] An Open Toolkit for Knowledge Graph Extraction and Construction
Build high-performance AI models with modular building blocks
Open Source Routing Engine for OpenStreetMap
Basic implementation code for multimodal models and some applications or fine-tuning tasks based on them.
[ICLR 2024 Spotlight] This is the official code for the paper "SNIP: Bridging Mathematical Symbolic and Numeric Realms with Unified Pre-training"
Language Modeling Research Hub, a comprehensive compendium for enthusiasts and scholars delving into the fascinating realm of language models (LMs), with a particular focus on large language models (LLMs)
Add a description, image, and links to the multi-modal topic page so that developers can more easily learn about it.
To associate your repository with the multi-modal topic, visit your repo's landing page and select "manage topics."