fazlulkarimweb
A full-stack developer trying to enjoy his momentary journey in this pale blue dot 🎭 I speak Python, JavaScript and English. 🙌
@SuperDuperDBBangladesh
fazlulkarimweb's Stars
shadcn-ui/ui
Beautifully designed components that you can copy and paste into your apps. Accessible. Customizable. Open Source.
abi/screenshot-to-code
Drop in a screenshot and convert it to clean code (HTML/Tailwind/React/Vue)
pmndrs/zustand
🐻 Bear necessities for state management in React
recommenders-team/recommenders
Best Practices on Recommendation Systems
loft-sh/devpod
Codespaces but open-source, client-only and unopinionated: Works with any IDE and lets you use any cloud, kubernetes or just localhost docker.
OptimalScale/LMFlow
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.
NVIDIA/TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
zilliztech/GPTCache
Semantic cache for LLMs. Fully integrated with LangChain and llama_index.
TimDettmers/bitsandbytes
Accessible large language models via k-bit quantization for PyTorch.
a16z-infra/companion-app
AI companions with memory: a lightweight stack to create and host your own AI companions
SuperDuperDB/superduperdb
🔮 SuperDuperDB: Bring AI to your database! Build, deploy and manage any AI application directly with your existing data infrastructure, without moving your data. Including streaming inference, scalable model training and vector search.
llmware-ai/llmware
Unified framework for building enterprise RAG pipelines with small, specialized models
arcee-ai/mergekit
Tools for merging pretrained large language models.
turboderp/exllamav2
A fast inference library for running LLMs locally on modern consumer-class GPUs
neuralmagic/deepsparse
Sparsity-aware deep learning inference runtime for CPUs
turboderp/exllama
A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.
predibase/lorax
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
OpenLLMAI/OpenRLHF
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & Mixtral)
embeddings-benchmark/mteb
MTEB: Massive Text Embedding Benchmark
mlynch/nextjs-tailwind-ionic-capacitor-starter
A starting point for building an iOS, Android, and Progressive Web App with Tailwind CSS, React w/ Next.js, Ionic Framework, and Capacitor
pinecone-io/canopy
Retrieval Augmented Generation (RAG) framework and context engine powered by Pinecone
jakobhoeg/shadcn-chat
Customizable and re-usable chat component for you to use in your projects. Built on top of shadcn.
FalkorDB/FalkorDB
A super fast Graph Database uses GraphBLAS under the hood for its sparse adjacency matrix graph representation. Our goal is to provide the best Knowledge Graph for LLM (GraphRAG).
lancedb/vectordb-recipes
High quality resources & applications for LLMs, multi-modal models and VectorDBs
gitpod-io/workspace-images
Ready to use docker images for Gitpod workspaces
piraeusdatastore/piraeus
High Available Datastore for Kubernetes
theroyallab/tabbyAPI
An OAI compatible exllamav2 API that's both lightweight and fast
IST-DASLab/qmoe
Code for the paper "QMoE: Practical Sub-1-Bit Compression of Trillion-Parameter Models".
tyrchen/qdrant-lib
Extract core logic from qdrant and make it available as a library.
Schnuecks/frontail_AEM
📝 streaming logs to the browser