Pinned Repositories
convai
deeplearning-papernotes
Краткое изложение статей по NLP и Deep Learning
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
grad-based-attention
build gradient-based attention maps for transformer models
hello-github-actions
LIBRA
Megatron-LM
Ongoing research training transformer models at scale
sentencepiece
Unsupervised text tokenizer for Neural Network-based text generation.
t5-experiments
Tools and scripts for experimenting with Transformers: Bert, T5...
yurakuratov's Repositories
yurakuratov/t5-experiments
Tools and scripts for experimenting with Transformers: Bert, T5...
yurakuratov/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
yurakuratov/convai
yurakuratov/deeplearning-papernotes
Краткое изложение статей по NLP и Deep Learning
yurakuratov/grad-based-attention
build gradient-based attention maps for transformer models
yurakuratov/hello-github-actions
yurakuratov/LIBRA
yurakuratov/Megatron-LM
Ongoing research training transformer models at scale
yurakuratov/sentencepiece
Unsupervised text tokenizer for Neural Network-based text generation.