MiniCPM-Llama3-V 2.5: A GPT-4V Level Multimodal LLM on Your Phone
-
Updated
Jul 16, 2024 - Python
MiniCPM-Llama3-V 2.5: A GPT-4V Level Multimodal LLM on Your Phone
ModelScope: bring the notion of Model-as-a-Service to life.
a state-of-the-art-level open visual language model | 多模态预训练模型
Implementation / replication of DALL-E, OpenAI's Text to Image Transformer, in Pytorch
Unified embedding generation and search engine. Also available on cloud - cloud.marqo.ai
Open Source Routing Engine for OpenStreetMap
[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4V. 接近GPT-4V表现的可商用开源多模态对话模型
Chinese and English multimodal conversational language model | 多模态中英双语对话语言模型
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
[EMNLP 2022] An Open Toolkit for Knowledge Graph Extraction and Construction
Start building LLM-empowered multi-agent applications in an easier way.
Represent, send, store and search multimodal data
Video-LLaVA: Learning United Visual Representation by Alignment Before Projection
Mixture-of-Experts for Large Vision-Language Models
A robust, all-in-one GPT interface for Discord. ChatGPT-style conversations, image generation, AI-moderation, custom indexes/knowledgebase, youtube summarizer, and more!
A one-stop data processing system to make data higher-quality, juicier, and more digestible for (multimodal) LLMs! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷为大模型提供更高质量、更丰富、更易”消化“的数据!
Project Page for "LISA: Reasoning Segmentation via Large Language Model"
GPT4V-level open-source multi-modal model based on Llama3-8B
🥂 Gracefully face hCaptcha challenge with MoE(ONNX) embedded solution.
Add a description, image, and links to the multi-modal topic page so that developers can more easily learn about it.
To associate your repository with the multi-modal topic, visit your repo's landing page and select "manage topics."