okoge-kaz
Master (Computer Science) student at Tokyo Institute of Technology
Tokyo Institute of TechnologyTokyo Japan
Pinned Repositories
Compiler_Construction
Tokyo Institute of Technology 2022-2Q CSC. T372
Creative-Programming-Project
2021-1Q,2Q プログラミング創造演習 (Tokyo Tech)
document
Documentation for Computer Science and Blog for Software Engineer Career
golang-todo-application
システム設計演習 Tokyo Institute of Technology
llm-jp-sakura-ansible
llm-recipes
Ongoing Research Project for continaual pre-training LLM(dense mode)
megatron-deepspeed-turing-techblog
Turing Tech Blog Repository
moe-recipes
Ongoing research training Mixture of Expert models.
turing-techblog-megatron-deepspeed
環境構築方法の詳細は以下のLinkから
wandb_watcher
ABCI 大規模言語モデル構築支援にてwandbのジョブを監視するためのツール
okoge-kaz's Repositories
okoge-kaz/moe-recipes
Ongoing research training Mixture of Expert models.
okoge-kaz/llm-recipes
Ongoing Research Project for continaual pre-training LLM(dense mode)
okoge-kaz/llm-jp-sakura-ansible
okoge-kaz/wandb_watcher
ABCI 大規模言語モデル構築支援にてwandbのジョブを監視するためのツール
okoge-kaz/Megatron-LM
Ongoing research training transformer models at scale
okoge-kaz/swallow-tuning
okoge-kaz/accelerate
🚀 A simple way to train and use PyTorch models with multi-GPU, TPU, mixed-precision
okoge-kaz/axlearn
okoge-kaz/deploymentmanager-samples
Deployment Manager samples and templates.
okoge-kaz/grouped_gemm
PyTorch bindings for CUTLASS grouped GEMM.
okoge-kaz/hpsc-2024
okoge-kaz/levanter
Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax
okoge-kaz/llama-recipes
Examples and recipes for Llama 2 model
okoge-kaz/llm-jp-dpo
okoge-kaz/llm-jp-Megatron-DeepSpeed
okoge-kaz/llm-node-tests
okoge-kaz/megablocks
okoge-kaz/mistral-hackathon
okoge-kaz/ml-engineering
Machine Learning Engineering Open Book
okoge-kaz/MoEfication
okoge-kaz/nanotron
Minimalistic large language model 3D-parallelism training
okoge-kaz/NeMo
NeMo: a toolkit for conversational AI
okoge-kaz/NeMo-Aligner
Scalable toolkit for efficient model alignment
okoge-kaz/NeMo-Megatron-Launcher
NeMo Megatron launcher and tools
okoge-kaz/okoge-kaz
okoge-kaz/OLMo
Modeling, training, eval, and inference code for OLMo
okoge-kaz/ppcomp24
okoge-kaz/swallow-project-parper-graph
okoge-kaz/TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
okoge-kaz/TSUBAME-4.0-hands-on