Pinned Repositories
100-gdb-tips
A collection of gdb tips. 100 maybe just mean many here.
Advanced-Lane-Line-Detection
AI_compiler_development_guide
Free resource for the book AI Compiler Development Guide
ailab
AITemplate
AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (NVIDIA GPU) and MatrixCore (AMD GPU) inference.
EnableGpuWithWSL2OnWin11
win11在WSL2虚拟环境中使用GPU开启训练任务
opencv_build
build for x86 or arm on linux
PaddleSeg
End-to-end image segmentation kit based on PaddlePaddle.
PP-LiteSeg
Pytorch implement of PP-LiteSeg
training_extensions
Trainable models and NN optimization tools
Amanda-Barbara's Repositories
Amanda-Barbara/autogen
Enable Next-Gen Large Language Model Applications. Join our Discord: https://discord.gg/pAbnFJrkgZ
Amanda-Barbara/blora-text-generation-inference
Batched LORA + Continuous Batching
Amanda-Barbara/BLoRA-TGI-with-python-server
Batched Lora + Continuous Batching
Amanda-Barbara/ComputeLibrary
The Compute Library is a set of computer vision and machine learning functions optimised for both Arm CPUs and GPUs using SIMD technologies.
Amanda-Barbara/CPlusPlus-Tutorial
C++ Tutorial
Amanda-Barbara/cutlass-flash-attention
flash attention tutorial written in python, triton, cuda, cutlass
Amanda-Barbara/flash-attention
Fast and memory-efficient exact attention
Amanda-Barbara/flux
A fast communication-overlapping library for tensor parallelism on GPUs.
Amanda-Barbara/generative-ai-for-beginners
12 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
Amanda-Barbara/gpu-profiling
GPU Profiling
Amanda-Barbara/InfiniGen
InfiniGen: Efficient Generative Inference of Large Language Models with Dynamic KV Cache Management (OSDI'24)
Amanda-Barbara/kohya_ss
train stable diffusion models
Amanda-Barbara/langchain
⚡ Building applications with LLMs through composability ⚡
Amanda-Barbara/Latte
The official implementation of Latte: Latent Diffusion Transformer for Video Generation.
Amanda-Barbara/leetcode-master
《代码随想录》LeetCode 刷题攻略:200道经典题目刷题顺序,共60w字的详细图解,视频难点剖析,50余张思维导图,支持C++,Java,Python,Go,JavaScript等多语言版本,从此算法学习不再迷茫!🔥🔥 来看看,你会发现相见恨晚!🚀
Amanda-Barbara/llm-benchmark-test
include LLM open framework measurement
Amanda-Barbara/llm_long_context_bench202405
Amanda-Barbara/long-context-attention
Sequence Parallel Attention for Long Context LLM Model Training and Inference
Amanda-Barbara/MetaGPT-agent
🌟 The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming
Amanda-Barbara/MiniGPT-4
Open-sourced codes for MiniGPT-4 and MiniGPT-v2 (https://minigpt-4.github.io, https://minigpt-v2.github.io/)
Amanda-Barbara/MiniGPT4-video
Amanda-Barbara/mistral-src
Reference implementation of Mistral AI 7B v0.1 model.
Amanda-Barbara/multimodal-ai-jina
☁️ Build multimodal AI applications with cloud-native stack
Amanda-Barbara/NvidiaTransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
Amanda-Barbara/text2video-generative-models
Generative Models by Stability AI
Amanda-Barbara/tgi-benchmarking
Benchmarking LLMs on GPUs
Amanda-Barbara/tvm-mlc-llm
Enable everyone to develop, optimize and deploy AI models natively on everyone's devices.
Amanda-Barbara/vAttention
Amanda-Barbara/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
Amanda-Barbara/yolov5-5.x-annotations
一个基于yolov5-5.0的中文注释版本!