open-source-llm

There are 26 repositories under open-source-llm topic.

  • bentoml/OpenLLM

    Run any open-source LLMs, such as DeepSeek and Llama, as OpenAI compatible API endpoint in the cloud.

    Language:Python11.9k56279788
  • kennethleungty/Llama-2-Open-Source-LLM-CPU-Inference

    Running Llama 2 and other Open-Source LLMs on CPU Inference Locally for Document Q&A

    Language:Python9671324211
  • ghimiresunil/LLM-PowerHouse-A-Curated-Guide-for-Large-Language-Models-with-Custom-Training-and-Inferencing

    LLM-PowerHouse: Unleash LLMs' potential through curated tutorials, best practices, and ready-to-use code for custom training and inferencing.

    Language:Jupyter Notebook713170119
  • rohan-paul/LLM-FineTuning-Large-Language-Models

    LLM (Large Language Model) FineTuning

    Language:Jupyter Notebook56482137
  • aikit

    kaito-project/aikit

    🏗️ Fine-tune, build, and deploy open-source LLMs easily!

    Language:Go49076551
  • Kenza-AI/sagify

    LLMs and Machine Learning done easily

    Language:Python441165368
  • PetroIvaniuk/llms-tools

    A list of LLMs Tools & Projects

  • ArdaGnsrn/ollama-php

    This is a PHP library for Ollama. Ollama is an open-source project that serves as a powerful and user-friendly platform for running LLMs on your local machine. It acts as a bridge between the complexities of LLM technology and the desire for an accessible and customizable AI experience.

    Language:PHP19131020
  • BodhiSearch/BodhiApp

    Run Open Source/Open Weight LLMs locally with OpenAI compatible APIs

    Language:TypeScript1236149
  • SAP-samples/btp-generative-ai-hub-use-cases

    Samples on how to build industry solution leveraging generative AI capabilities on top of SAP BTP and integrated with SAP S/4HANA Cloud.

    Language:HTML6317967
  • EmbeddedLLM/embeddedllm

    EmbeddedLLM: API server for Embedded Device Deployment. Currently support CUDA/OpenVINO/IpexLLM/DirectML/CPU

    Language:Python442191
  • Talnz007/VulkanIlm

    GPU-accelerated LLaMA inference wrapper for legacy Vulkan-capable systems a Pythonic way to run AI with knowledge (Ilm) on fire (Vulkan).

    Language:Python250
  • Praveen76/Finetune-Open-Source-LLMs-on-custom-data

    Fine-tune open-source Language Models (LLMs) on E-commerce data, leveraging Amazon's sales data. Showcase a tailored solution for enhanced language understanding and generation with a focus on custom E-commerce datasets.

    Language:Jupyter Notebook10101
  • Utshav-paudel/Docbot

    Read your local files and answer your queries

    Language:Python5102
  • Mehdi-Letafati/multi-agent-workflows-from-scratch

    Multi-agent workflows with Llama3: A private on-device multi-agent framework

    Language:Python3100
  • AdvaithVellanki/Leads-Qualifier-Agent

    This project contains the code and documentation for an autonomous AI agent that classifies, enriches, and scores inbound business leads. It is built with a FastAPI backend, a LangGraph agent workflow powered by a local Ollama LLM, and a Streamlit frontend for demonstration.

    Language:Python2
  • SSR-web-cloud/LocalPrompt

    LocalPrompt is an AI-powered tool designed to refine and optimize AI prompts, helping users run locally hosted AI models like Mistral-7B for privacy and efficiency. Ideal for developers seeking to run LLMs locally without external APIs.

  • AyaFergany/RAG-App-Using-a-LLM

    In this project, we leverage Weaviate, a vector database, to power our retrieval-augmented generation (RAG) application. Weaviate enables efficient vector similarity search, which is crucial for building effective RAG systems. Additionally, we use local language model (LLM) and embedding models.

    Language:Jupyter Notebook1101
  • cr4kn4x/the-stix-intern

    *the-stix-intern* a minimalistic framework for the automized extraction of CTI from unstructured texts

    Language:HTML1
  • sandy-sp/gittxt

    Gittxt is an AI-focused CLI and plugin tool for extracting, filtering, and packaging text from GitHub repos. Build LLM-compatible datasets, prep code for prompt engineering, and power AI workflows with structured .txt, .json, .md, or .zip outputs.

    Language:Python01
  • srinivasansaravanamuthu/PrivateLLM

    Create a Small LLM using EleutherAI/gpt-neo-2.7B - Fine Tune It for a Specalized Purpouse and Leverage as a Co-Pilot

    Language:HTML0100
  • AlexPraxedes12/LearnSynth

    Turn text/audio/video into study packs (summaries, concept maps, quizzes). Flutter + FastAPI. Uses GPT-OSS-20B (Replicate) by default, with optional fallbacks (OpenAI/Anthropic) and OSS modes (Ollama/vLLM).

    Language:Dart
  • FarazF19/Conversational-QnA-Chatbot

    This project implements a production-ready Conversational (RAG) chatbot using LangChain, Groq’s LLaMA 3 model, ChromaDB, and HuggingFace embeddings. It supports persistent chat history across sessions using RunnableWithMessageHistory, and integrates modular prompting, document loading, and retriever-aware chains to handle context-aware multi-chats.

    Language:Jupyter Notebook10
  • mostafaaladham/UXI-LLM

    UXI-LLM is a modular hybrid language model that combines neural and symbolic reasoning for enhanced AI capabilities. Built for flexibility and local use, it empowers developers and researchers to customize their AI solutions with ease. 🐙💻

    Language:Python
  • mshobari/local_ollama_project

    🤖 Build a local AI assistant that categorizes user input with structured JSON output, ensuring privacy and rapid performance for iOS/mobile integration.

  • sawadkk/LocalPrompt

    LocalPrompt is an AI-powered tool designed to refine and optimize AI prompts, helping users run locally hosted AI models like Mistral-7B for privacy and efficiency. Ideal for developers seeking to run LLMs locally without external APIs.

    Language:Python