Pinned Repositories
dockerpackage
dockerpackage
urjtag
Local copy of urjtag
vllm-pascal
A fork of vLLM enabling Pascal architecture GPUs
write-a-C-interpreter
Write a simple interpreter of C. Inspired by c4 and largely based on it.
xmem
Atmel External Memory Board XMEM interface
TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
infinity
Infinity is a high-throughput, low-latency REST API for serving text-embeddings, reranking models and clip
open-webui
User-friendly WebUI for LLMs (Formerly Ollama WebUI)
llm
Access large language models from the command-line
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
cduk's Repositories
cduk/vllm-pascal
A fork of vLLM enabling Pascal architecture GPUs
cduk/dockerpackage
dockerpackage
cduk/urjtag
Local copy of urjtag
cduk/write-a-C-interpreter
Write a simple interpreter of C. Inspired by c4 and largely based on it.
cduk/xmem
Atmel External Memory Board XMEM interface