glzbcrt
I am a solution architect at @microsoft helping partners develop intelligent and resilient solutions.
georgeluiz.comJoinville - SC - Brazil
glzbcrt's Stars
openai/whisper
Robust Speech Recognition via Large-Scale Weak Supervision
golang-standards/project-layout
Standard Go Project Layout
ManimCommunity/manim
A community-maintained Python framework for creating mathematical animations.
duckdb/duckdb
DuckDB is an analytical in-process SQL database management system
locustio/locust
Write scalable load tests in plain Python 🚗💨
encode/httpx
A next generation HTTP client for Python. 🦋
artidoro/qlora
QLoRA: Efficient Finetuning of Quantized LLMs
NirDiamant/RAG_Techniques
This repository showcases various advanced techniques for Retrieval-Augmented Generation (RAG) systems. RAG systems combine information retrieval with generative models to provide accurate and contextually rich responses.
huggingface/text-generation-inference
Large Language Model Text Generation Inference
huggingface/tokenizers
💥 Fast State-of-the-Art Tokenizers optimized for Research and Production
NVIDIA/TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
GoogleCloudPlatform/generative-ai
Sample code and notebooks for Generative AI on Google Cloud, with Gemini on Vertex AI
tpn/pdfs
Technically-oriented PDF Collection (Papers, Specs, Decks, Manuals, etc)
docker/docker-py
A Python library for the Docker Engine API
xorbitsai/inference
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
meta-llama/llama-models
Utilities intended for use with Llama models.
ray-project/llm-numbers
Numbers every LLM developer should know
microsoft/ProcMon-for-Linux
A Linux version of the Procmon Sysinternals tool
turboderp/exllamav2
A fast inference library for running LLMs locally on modern consumer-class GPUs
huggingface/text-embeddings-inference
A blazing fast inference solution for text embeddings models
predibase/lorax
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
chiphuyen/aie-book
[WIP] Resources for AI engineers. Also contains supporting materials for the book AI Engineering (Chip Huyen, 2025)
AIDC-AI/Marco-o1
An Open Large Reasoning Model for Real-World Solutions
IST-DASLab/marlin
FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.
fabioz/PyDev.Debugger
Sources for the debugger used in PyDev, PyCharm and VSCode Python
microsoft/SemanticKernelCookBook
This is a Semantic Kernel's book for beginners
Zefan-Cai/Awesome-LLM-KV-Cache
Awesome-LLM-KV-Cache: A curated list of 📙Awesome LLM KV Cache Papers with Codes.
blobfile/blobfile
Read Google Cloud Storage, Azure Blobs, and local paths with the same interface
microsoft/devex-unlocked
Content used in the DevEx Unlocked in LATAM Region
oaviles/hello_appconfiguration
Reference Implementation about Azure App Configuration. Build and Deploy Web App with feature flags support on Azure Kubernetes Services based on DevSecOps Practices