causal-language-modeling
There are 17 repositories under causal-language-modeling topic.
DunnBC22/NLP_Projects
Repository for My HuggingFace Natural Language Processing Projects
hogru/MolReactGen
Auto-regressive causal language model for molecule (SMILES) and reaction template (SMARTS) generation based on the Hugging Face implementation of OpenAI's GPT-2 transformer decoder model
rhubarbwu/linguistic-collapse
Codebase for arXiv:2405.17767, based on GPT-Neo and TinyStories.
Anonym0usWork1221/python-code-docstring-scraper
A multi-threaded GitHub scraper to collect Python code with docstrings from public repositories, creating a well-documented dataset for the JaraConverse LLM model.
SharathHebbar/Transformers
Transformers Intuition
Cyrilvallez/LLM_playground
A quick and easy way to interact with open-source LLMs.
Jayveersinh-Raj/LoRA_implementation
This is the implementation of low rank adaptation (LoRA) which is a subset of parameter efficient fine tuning (PEFT).
samyak24jain/gpt2-intent-classification
Causal language modeling and intent classification using GPT-2.
aneesh-aparajit/picturebook.ai
An AI generated picturebook.
JersonGB22/NaturalLanguageProcessing
Links to my repositories, where I implement a wide variety of Natural Language Processing models using TensorFlow and Hugging Face.
saagar-parikh/ASR_LLM_Rescoring
Rescoring Automatic Speech Recognition using Large Language Models
ShiningLab/PromptSub
This repository is for the paper Lexical Substitution as Causal Language Modeling. In Proceedings of the 13th Joint Conference on Lexical and Computational Semantics (*SEM 2024), Mexico City, Mexico. Association for Computational Linguistics.
thibaud-perrin/hibo-mistral-7b-fc
Dataset and model fine-tuning for function calling
tranquoctrinh/huggingface-transformers-examples
Fine-tuning (or training from scratch) the library models for language modeling on a text dataset for GPT, GPT-2, ALBERT, BERT, DitilBERT, RoBERTa, XLNet... GPT and GPT-2 are trained or fine-tuned using a causal language modeling (CLM) loss while ALBERT, BERT, DistilBERT and RoBERTa are trained or fine-tuned using a masked language modeling (MLM) loss.
PastelBelem8/ml4nlp-cogsci-summer22
Course materials for the Machine Learning for NLP course taught by Sameer Singh for the Cognitive Science summer school 2022.