Pinned Repositories
bloom_lora_ja
dedup_sentence
lit-llama-ja
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
makeSentens
quote_test
quote_test
RWKV-LM-LoRA-ja
if001's Repositories
if001/dedup_sentence
if001/lit-llama-ja
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
if001/alpaca-lora
Instruct-tuning LLaMA on consumer hardware
if001/arXiv_translate
if001/book_impressions
if001/chat_face
if001/FLAN
if001/gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
if001/grok
if001/harbor
Simple and minimal personal blog theme.
if001/HojiChar_OSCAR_sample
if001/if-blog-hugo
if001/litgpt
Pretrain, finetune, deploy 20+ LLMs on your own data. Uses state-of-the-art techniques: flash attention, FSDP, 4-bit, LoRA, and more.
if001/llm-jp-eval
if001/llm_bunpo
if001/lm-evaluation-harness
A framework for few-shot evaluation of autoregressive language models.
if001/matmulfreellm
Implementation for MatMul-free LM.
if001/Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
if001/mergekit
Tools for merging pretrained large language models.
if001/mergoo
A library for easily merging multiple LLM experts, and efficiently train the merged LLM.
if001/Pai-Megatron-Patch
The official repo of Pai-Megatron-Patch for LLM & VLM large scale training developed by Alibaba Cloud.
if001/qwen2_sft
if001/rinna_3b_instructions
if001/rinna_4b_multi_instructions
if001/spm_tokenizer_ja
if001/style-bert-vits_sample
if001/ucllm_nedo_prod
if001/wanda
A simple and effective LLM pruning approach.
if001/wiki_analysis
if001/wiki_classification