EmbeddedLLM
EmbeddedLLM is the creator behind JamAI Base, a platform designed to orchestrate AI with spreadsheet-like simplicity.
Singapore
Pinned Repositories
embeddedllm
EmbeddedLLM: API server for Embedded Device Deployment. Currently support CUDA/OpenVINO/IpexLLM/DirectML/CPU
JamAIBase
The collaborative spreadsheet for AI. Chain cells into powerful pipelines, experiment with prompts and models, and evaluate LLM responses in real-time. Work together seamlessly to build and iterate on AI applications.
jamaibase-cookbook
JamAI Base cookbook repo
jamaibase-nextjs-vercel
jamaibase-ts-docs
Typescript Documentation of JamAISDK
LLaVA-Plus-Serve
LLaVA-Plus: Large Language and Vision Assistants that Plug and Learn to Use Skills
mamba-rocm
unstructured-executable
Open source libraries and APIs to build custom preprocessing pipelines for labeling, training, or production machine learning pipelines.
vllm
vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs
xformers-rocm
Strip down to support flash attention v2 ROCM.
EmbeddedLLM's Repositories
EmbeddedLLM/JamAIBase
The collaborative spreadsheet for AI. Chain cells into powerful pipelines, experiment with prompts and models, and evaluate LLM responses in real-time. Work together seamlessly to build and iterate on AI applications.
EmbeddedLLM/vllm
vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs
EmbeddedLLM/embeddedllm
EmbeddedLLM: API server for Embedded Device Deployment. Currently support CUDA/OpenVINO/IpexLLM/DirectML/CPU
EmbeddedLLM/mamba-rocm
EmbeddedLLM/jamaibase-cookbook
JamAI Base cookbook repo
EmbeddedLLM/xformers-rocm
Strip down to support flash attention v2 ROCM.
EmbeddedLLM/jamaibase-nextjs-vercel
EmbeddedLLM/LLaVA-Plus-Serve
LLaVA-Plus: Large Language and Vision Assistants that Plug and Learn to Use Skills
EmbeddedLLM/causal-conv1d-rocm
Causal depthwise conv1d in CUDA, with a PyTorch interface
EmbeddedLLM/dspy
DSPy: The framework for programming—not prompting—foundation models
EmbeddedLLM/EAGLE
EAGLE: Lossless Acceleration of LLM Decoding by Feature Extrapolation
EmbeddedLLM/grouped_gemm-rocm
PyTorch bindings for CUTLASS grouped GEMM.
EmbeddedLLM/jamaibase-ts-docs
Typescript Documentation of JamAISDK
EmbeddedLLM/megablocks-rocm
EmbeddedLLM/stk-rocm
EmbeddedLLM/unstructured-executable
Open source libraries and APIs to build custom preprocessing pipelines for labeling, training, or production machine learning pipelines.
EmbeddedLLM/workshop
EmbeddedLLM/ai-town
A MIT-licensed, deployable starter kit for building and customizing your own version of AI town - a virtual town where AI characters live, chat and socialize.
EmbeddedLLM/arena-hard-auto
Arena-Hard-Auto: An automatic LLM benchmark.
EmbeddedLLM/github-bot
EmbeddedLLM/infinity-executable
Infinity is a high-throughput, low-latency REST API for serving vector embeddings, supporting a wide range of text-embedding models and frameworks.
EmbeddedLLM/jamaibase-expressjs-vercel
EmbeddedLLM/nlux-jamai
The 𝗣𝗼𝘄𝗲𝗿𝗳𝘂𝗹 Conversational AI JavaScript Library
EmbeddedLLM/PowerToys
Windows system utilities to maximize productivity
EmbeddedLLM/unstructured-api-executable
EmbeddedLLM/unstructured-inference-executable
EmbeddedLLM/unstructured-python-client
A Python client for the Unstructured hosted API