minicpm-v

There are 8 repositories under minicpm-v topic.

  • OpenBMB/MiniCPM-V

    MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone

    Language:Python13.1k107617914
  • modelscope/ms-swift

    Use PEFT or Full-parameter to finetune 400+ LLMs (Qwen2.5, Llama3.2, GLM4, Internlm2.5, Yi1.5, Mistral, Baichuan2, DeepSeek, ...) or 150+ MLLMs (Qwen2-VL, Qwen2-Audio, Llama3.2-Vision, Llava, InternVL2.5, MiniCPM-V-2.6, GLM4v, Xcomposer2.5, Yi-VL, DeepSeek-VL2, Phi3.5-Vision, GOT-OCR2, ...).

    Language:Python5k231.5k433
  • PaddlePaddle/PaddleMIX

    Paddle Multimodal Integration and eXploration, supporting mainstream multi-modal tasks, including end-to-end large-scale multi-modal pretrain models and diffusion model toolbox. Equipped with high performance and flexibility.

    Language:Python46422163169
  • RLHF-V/RLAIF-V

    RLAIF-V: Aligning MLLMs through Open-Source AI Feedback for Super GPT-4V Trustworthiness

    Language:Python26663610
  • AXERA-TECH/ax-llm

    Explore LLM model deployment based on AXera's AI chips

    Language:C++674410
  • sitamgithub-MSIT/PicQ

    PicQ: Demo for MiniCPM-V 2.6 to answer questions about images using natural language.

    Language:Python4200
  • Kazuhito00/MiniCPM-V2.6-Colaboratory-Sample

    軽量VLMのMiniCPM-V2.6のColaboratoryサンプル

    Language:Jupyter Notebook210
  • sitamgithub-MSIT/VidiQA

    VidiQA: Demo for MiniCPM-V 2.6 to answer questions about videos using natural language.

    Language:Python220