ilkerkesen's Stars
ollama/ollama
Get up and running with Llama 3.2, Mistral, Gemma 2, and other large language models.
hiyouga/LLaMA-Factory
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
unslothai/unsloth
Finetune Llama 3.2, Mistral, Phi, Qwen 2.5 & Gemma LLMs 2-5x faster with 80% less memory
meta-llama/llama-recipes
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama for WhatsApp & Messenger.
Lightning-AI/litgpt
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
RUCAIBox/LLMSurvey
The official GitHub page for the survey paper "A Survey of Large Language Models".
huggingface/text-generation-inference
Large Language Model Text Generation Inference
cxli233/FriendsDontLetFriends
Friends don't let friends make certain types of data visualization - What are they and why are they bad.
janikvonrotz/awesome-powershell
A curated list of delightful PowerShell modules and resources
allenai/OLMo
Modeling, training, eval, and inference code for OLMo
peterbrittain/asciimatics
A cross platform package to do curses-like operations, plus higher level APIs and widgets to create text UIs and ASCII art animations
stanford-futuredata/ColBERT
ColBERT: state-of-the-art neural search (SIGIR'20, TACL'21, NeurIPS'21, NAACL'22, CIKM'22, ACL'23, EMNLP'23)
NVlabs/VILA
VILA - a multi-image visual language model with training, inference and evaluation recipe, deployable from cloud to edge (Jetson Orin and laptops)
MiuLab/Taiwan-LLM
Traditional Mandarin LLMs for Taiwan
BatsResearch/bonito
A lightweight library for generating synthetic instruction tuning datasets for your data without GPT.
dome272/Wuerstchen
Official implementation of Würstchen: Efficient Pretraining of Text-to-Image Models
inseq-team/inseq
Interpretability for sequence generation models 🐛 🔍
tianyi-lab/HallusionBench
[CVPR'24] HallusionBench: You See What You Think? Or You Think What You See? An Image-Context Reasoning Benchmark Challenging for GPT-4V(ision), LLaVA-1.5, and Other Multi-modality Models
theislab/ehrapy
Electronic Health Record Analysis with Python.
DAMO-NLP-SG/VCD
[CVPR 2024 Highlight] Mitigating Object Hallucinations in Large Vision-Language Models through Visual Contrastive Decoding
IBM/SALMON
Self-Alignment with Principle-Following Reward Models
YiyangZhou/LURE
[ICLR 2024] Analyzing and Mitigating Object Hallucination in Large Vision-Language Models
zeyofu/BLINK_Benchmark
This repo contains evaluation code for the paper "BLINK: Multimodal Large Language Models Can See but Not Perceive". https://arxiv.org/abs/2404.12390 [ECCV 2024]
mbzuai-nlp/bactrian-x
A Multilingual Replicable Instruction-Following Model
neulab/langrank
A program to choose transfer languages for cross-lingual learning
Hritikbansal/videocon
ViralLab/TurkishBERTweet
TurkishBERTweet: Fast and Reliable Large Language Model for Social Media Analysis
Heidelberg-NLP/CC-SHAP-VLM
Official code implementation for the paper "Do Vision & Language Decoders use Images and Text equally? How Self-consistent are their Explanations?"
cyberiada/cartography-for-compositionality
The official repository for our EMNLP 2023 Findings paper, Harnessing Dataset Cartography for Better Compositional Generalization in Transformers
flipz357/overview-of-my-thesis-code
This is an overview over the works associated with my dissertation.