Pinned Repositories
run_sitter
binary_search
二分探索
calendar
chat_vector
ChunkLlama
Data and code for our paper "Training-Free Long-Context Scaling of Large Language Models"
CLEX
[ICLR 2024] CLEX: Continuous Length Extrapolation for Large Language Models
Cutout
2.56%, 15.20%, 1.30% on CIFAR10, CIFAR100, and SVHN https://arxiv.org/abs/1708.04552
dot_config
gcd
アルゴリズムは漸化式: ユークリッドの互除法
hpsc-2024
t0-0's Repositories
t0-0/dot_config
t0-0/parallel-assist
t0-0/swallow-tuning
t0-0/run_sitter
t0-0/hpsc-2024
t0-0/chat_vector
t0-0/CLEX
[ICLR 2024] CLEX: Continuous Length Extrapolation for Large Language Models
t0-0/calendar
t0-0/InfiniteBench
Codes for the paper "∞Bench: Extending Long Context Evaluation Beyond 100K Tokens": https://arxiv.org/abs/2402.13718
t0-0/LongBench
LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding
t0-0/InfLLM
The code of our paper "InfLLM: Unveiling the Intrinsic Capacity of LLMs for Understanding Extremely Long Sequences with Training-Free Memory"
t0-0/ChunkLlama
Data and code for our paper "Training-Free Long-Context Scaling of Large Language Models"
t0-0/react-2048
t0-0/ToDoApp
t0-0/Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
t0-0/mm_builder
t0-0/RedPajama-Data
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
t0-0/binary_search
二分探索
t0-0/gcd
アルゴリズムは漸化式: ユークリッドの互除法
t0-0/Cutout
2.56%, 15.20%, 1.30% on CIFAR10, CIFAR100, and SVHN https://arxiv.org/abs/1708.04552