Ab1992ao
Gubkin Russian State University (National Research Institute) of Oil and Gas (intellectual systems, computer science faculty), kaggle competitons master
SaluteDevicesMoscow
Pinned Repositories
ACM.14.Lab1
CFT-challenge-2018
CNN and GRU spellchecker
keras-tcn
Keras Temporal Convolutional Network.
RaiffHack
rusentiment
SiburAIHack
Simple-siamese-model-LSTM-with-attention-and-batch.norm
Simple_chat_bot
Simple bot with own knowledge database and simple QT-interface
tele2_hack
Code for tele2 hackathone. Team 2HACK4YOU
Toxic-Jigsaw-solving
My top 8% kernel (bronze medal) with pre-trained BERT and BiGRU network with attention and capsule layer
Ab1992ao's Repositories
Ab1992ao/LM-finetune
Код для файнтюна LM (rugpt, LLaMa, FRED T5) средствами transformers + deepspeed + LoRa
Ab1992ao/pigeonXT
🐦 Quickly annotate data from the comfort of your Jupyter notebook
Ab1992ao/vk_cloud_peft_gpt
Ab1992ao/albert
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations
Ab1992ao/awesome-RLHF
A curated list of reinforcement learning with human feedback resources (continually updated)
Ab1992ao/facebook-hateful-memes
Facebook hateful memes challenge using multi-modal learning. More info about it here: https://ai.facebook.com/blog/hateful-memes-challenge-and-data-set
Ab1992ao/FNet-TensorFlow-PyTorch
TensorFlow & PyTorch implementation of the paper "FNet: Mixing Tokens with Fourier Transforms".
Ab1992ao/FRED-T5-Finetuning
Скрипт для файнтюна FRED-T5
Ab1992ao/GPT-4-LLM
Ab1992ao/HateXplain
Can we use explanations to improve hate speech models? Our paper accepted at AAAI 2021 tries to explore that question.
Ab1992ao/JAX-in-Action
Notebooks for the "JAX in Action" book
Ab1992ao/LLaMA2
This repository contains an implementation of the LLaMA 2 (Large Language Model Meta AI) model, a Generative Pretrained Transformer (GPT) variant. The implementation focuses on the model architecture and the inference process. The code is restructured and heavily commented to facilitate easy understanding of the key parts of the architecture.
Ab1992ao/lm-human-preferences--tf
Code for the paper Fine-Tuning Language Models from Human Preferences
Ab1992ao/Machine-Learning-with-Python
Practice and tutorial-style notebooks covering wide variety of machine learning techniques
Ab1992ao/MergeLM
Codebase for Merging Language Models
Ab1992ao/mesh
Mesh TensorFlow: Model Parallelism Made Easier
Ab1992ao/metric-learning-in-nlp
Ab1992ao/minGPT
A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
Ab1992ao/model_merging_tf_transfomers
Ab1992ao/models
Models and examples built with TensorFlow
Ab1992ao/Prompt-Engineering-Guide
:octopus: Guides, papers, lecture, and resources for prompt engineering
Ab1992ao/rulm
Language modeling for Russian
Ab1992ao/ruTS
Библиотека для извлечения статистик из текстов на русском языке.
Ab1992ao/SafeNLP
Safety Score for Pre-Trained Language Models
Ab1992ao/t5x_retrieval
Ab1992ao/tensor-house
A collection of reference machine learning and optimization models for enterprise operations: marketing, pricing, supply chain
Ab1992ao/TF_JAX_tutorials
All about the fundamental blocks of TF and JAX!
Ab1992ao/torch-electra-pretrain
Ab1992ao/trl
Train transformer language models with reinforcement learning.
Ab1992ao/tuning_playbook
A playbook for systematically maximizing the performance of deep learning models.