Pinned Repositories
awesome-local-ai
An awesome repository of local AI tools
cortex
Drop-in, local AI alternative to the OpenAI stack. Multi-engine (llama.cpp, TensorRT-LLM). Powers 👋 Jan
cortex.llamacpp
cortex.python
C++ code that run Python embedding
cortex.tensorrt-llm
Nitro is an C++ inference server on top of TensorRT-LLM. OpenAI-compatible API. Run blazing fast inference on Nvidia GPUs. Used in Jan
docs
Jan.ai Website & Documentation
extension-template
jan
Jan is an open source alternative to ChatGPT that runs 100% offline on your computer. Multiple engine support (llama.cpp, TensorRT-LLM)
model-converter
models
Jan's Repositories
janhq/jan
Jan is an open source alternative to ChatGPT that runs 100% offline on your computer. Multiple engine support (llama.cpp, TensorRT-LLM)
janhq/cortex
Drop-in, local AI alternative to the OpenAI stack. Multi-engine (llama.cpp, TensorRT-LLM). Powers 👋 Jan
janhq/awesome-local-ai
An awesome repository of local AI tools
janhq/cortex.tensorrt-llm
Nitro is an C++ inference server on top of TensorRT-LLM. OpenAI-compatible API. Run blazing fast inference on Nvidia GPUs. Used in Jan
janhq/model-converter
janhq/docs
Jan.ai Website & Documentation
janhq/extension-template
janhq/models
janhq/cortex.llamacpp
janhq/cortex.python
C++ code that run Python embedding
janhq/thinking-machines
Thinking Machines
janhq/architecture
janhq/cortex-node
The official Node.js / Typescript library for the OpenAI API
janhq/cortex-python
The official Python library for the OpenAI API
janhq/node-nitro
janhq/open-foundry
R&D experiments
janhq/triton_tensorrt_llm
janhq/charts
This repository contains helm chart for our team
janhq/plugin-catalog
janhq/py-nitro
janhq/cortex-web
janhq/cortex.hub
janhq/cortex.onnx
janhq/infinity
The AI-native database built for LLM applications, providing incredibly fast vector and full-text search
janhq/langchainjs
janhq/llama.cpp-avx-vnni
Port of Facebook's LLaMA model in C/C++
janhq/openai_trtllm
OpenAI compatible API for TensorRT LLM triton backend
janhq/pymaker
Make the py
janhq/tensorrtllm_backend
The Triton TensorRT-LLM Backend
janhq/trt-llm-as-openai-windows
This reference can be used with any existing OpenAI integrated apps to run with TRT-LLM inference locally on GeForce GPU on Windows instead of cloud.