joelrorseth
PhD @ UWaterloo. Currently focused on explainability and interpretability for LLMs.
Waterloo, Ontario
joelrorseth's Stars
ggerganov/llama.cpp
LLM inference in C/C++
lm-sys/FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
karpathy/minGPT
A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
guidance-ai/guidance
A guidance language for controlling large language models.
BerriAI/litellm
Python SDK, Proxy Server (LLM Gateway) to call 100+ LLM APIs in OpenAI format - [Bedrock, Azure, OpenAI, VertexAI, Cohere, Anthropic, Sagemaker, HuggingFace, Replicate, Groq]
google/gemma_pytorch
The official PyTorch implementation of Google's Gemma models
allenai/OLMo
Modeling, training, eval, and inference code for OLMo
facebookresearch/lingua
Meta Lingua: a lean, efficient, and easy-to-hack codebase to research LLMs.
openai/transformer-debugger
PAIR-code/lit
The Learning Interpretability Tool: Interactively analyze ML models to understand their behavior in an extensible and framework agnostic interface.
spcl/graph-of-thoughts
Official Implementation of "Graph of Thoughts: Solving Elaborate Problems with Large Language Models"
run-llama/llama_deploy
Deploy your agentic worfklows to production
TransformerLensOrg/TransformerLens
A library for mechanistic interpretability of GPT-style language models
facebookresearch/DPR
Dense Passage Retriever - is a set of tools and models for open domain Q&A task.
google-research/bleurt
BLEURT is a metric for Natural Language Generation based on transfer learning.
likenneth/honest_llama
Inference-Time Intervention: Eliciting Truthful Answers from a Language Model
AlignmentResearch/tuned-lens
Tools for understanding how transformer predictions are built layer-by-layer
ndif-team/nnsight
The nnsight package enables interpreting and manipulating the internals of deep learned models.
kmeng01/memit
Mass-editing thousands of facts into a transformer memory (ICLR 2023)
epfml/landmark-attention
Landmark Attention: Random-Access Infinite Context Length for Transformers
nelson-liu/lost-in-the-middle
Code and data for "Lost in the Middle: How Language Models Use Long Contexts"
night-chen/ToolQA
ToolQA, a new dataset to evaluate the capabilities of LLMs in answering challenging questions with external tools. It offers two levels (easy/hard) across eight real-life scenarios.
davidbau/baukit
ADAPT-uiuc/dias
Dias: Dynamic Rewriting of Pandas Code
google/belief-localization
This repository includes code for the paper "Does Localization Inform Editing? Surprising Differences in Where Knowledge Is Stored vs. Can Be Injected in Language Models."
DigitalHarborFoundation/llm-math-education
Retrieval augmented generation for middle-school math question answering and hint generation.
AI4LIFE-GROUP/LLM_Explainer
Code for paper: Are Large Language Models Post Hoc Explainers?
seilna/CNN-Units-in-NLP
:scissors: Repository for our ICLR 2019 paper: Discovery of Natural Language Concepts in Individual Units of CNNs
stefan-grafberger/mlwhatif
Data-Centric What-If Analysis for Native Machine Learning Pipelines
DigitalHarborFoundation/rag-for-math-qa
Analysis code for a research paper