attention-model
There are 237 repositories under attention-model topic.
JiahuiYu/generative_inpainting
DeepFill v1/v2 with Contextual Attention and Gated Convolution, CVPR 2018, and ICCV 2019 Oral
philipperemy/keras-attention
Keras Attention Layer (Luong and Bahdanau scores).
linto-ai/whisper-timestamped
Multilingual Automatic Speech Recognition with word-level timestamps and confidence
ozan-oktay/Attention-Gated-Networks
Use of Attention Gates in a Convolutional Neural Network / Medical Image Classification and Segmentation
awslabs/sockeye
Sequence-to-sequence framework with a focus on Neural Machine Translation based on PyTorch
pprp/awesome-attention-mechanism-in-cv
Awesome List of Attention Modules and Plug&Play Modules in Computer Vision
prakashpandey9/Text-Classification-Pytorch
Text classification using deep learning models in Pytorch
berniwal/swin-transformer-pytorch
Implementation of the Swin Transformer in PyTorch.
lorenmt/mtan
The implementation of "End-to-End Multi-Task Learning with Attention" [CVPR 2019].
Ha0Tang/AttentionGAN
AttentionGAN for Unpaired Image-to-Image Translation & Multi-Domain Image-to-Image Translation
lvapeab/nmt-keras
Neural Machine Translation with Keras
DengPingFan/SINet
Camouflaged Object Detection, CVPR 2020 (Oral)
kaushalshetty/Structured-Self-Attention
A Structured Self-attentive Sentence Embedding
daa233/generative-inpainting-pytorch
A PyTorch reimplementation for paper Generative Image Inpainting with Contextual Attention (https://arxiv.org/abs/1801.07892)
A-bone1/Attention-ocr-Chinese-Version
Attention OCR Based On Tensorflow
ai4co/rl4co
A PyTorch library for all things Reinforcement Learning (RL) for Combinatorial Optimization (CO)
chenjun2hao/Attention_ocr.pytorch
This repository implements the the encoder and decoder model with attention model for OCR
dabasajay/Image-Caption-Generator
A neural network to generate captions for an image using CNN and RNN with BEAM Search.
AuCson/PyTorch-Batch-Attention-Seq2seq
PyTorch implementation of batched bi-RNN encoder and attention-decoder.
RayanWang/Speech_emotion_recognition_BLSTM
Bidirectional LSTM network for speech emotion recognition.
rohitgirdhar/AttentionalPoolingAction
Code/Model release for NIPS 2017 paper "Attentional Pooling for Action Recognition"
hkproj/transformer-from-scratch-notes
Notes about "Attention is all you need" video (https://www.youtube.com/watch?v=bCz4OMemCcA)
zhongkaifu/Seq2SeqSharp
Seq2SeqSharp is a tensor based fast & flexible deep neural network framework written by .NET (C#). It has many highlighted features, such as automatic differentiation, different network types (Transformer, LSTM, BiLSTM and so on), multi-GPUs supported, cross-platforms (Windows, Linux, x86, x64, ARM), multimodal model for text and images and so on.
shyamupa/snli-entailment
attention model for entailment on SNLI corpus implemented in Tensorflow and Keras
dgai91/pytorch-acnn-model
code of Relation Classification via Multi-Level Attention CNNs
hugochan/BAMnet
Code & data accompanying the NAACL 2019 paper "Bidirectional Attentive Memory Networks for Question Answering over Knowledge Bases"
kyegomez/RT-X
Pytorch implementation of the models RT-1-X and RT-2-X from the paper: "Open X-Embodiment: Robotic Learning Datasets and RT-X Models"
tsenghungchen/SA-tensorflow
Soft attention mechanism for video caption generation
guillaume-chevalier/Linear-Attention-Recurrent-Neural-Network
A recurrent attention module consisting of an LSTM cell which can query its own past cell states by the means of windowed multi-head attention. The formulas are derived from the BN-LSTM and the Transformer Network. The LARNN cell with attention can be easily used inside a loop on the cell state, just like any other RNN. (LARNN)
LuZhenHuan/ECG-Classification-Demo
ECG Classification
monk1337/Various-Attention-mechanisms
This repository contain various types of attention mechanism like Bahdanau , Soft attention , Additive Attention , Hierarchical Attention etc in Pytorch, Tensorflow, Keras
pipidog/DeepTimeSeries
A Keras-based library for analysis of time series data using deep learning algorithms.
njchoma/transformer_image_caption
Image Captioning based on Bottom-Up and Top-Down Attention model
alokwhitewolf/Pytorch-Attention-Guided-CycleGAN
Pytorch implementation of Unsupervised Attention-guided Image-to-Image Translation.
14H034160212/HHH-An-Online-Question-Answering-System-for-Medical-Questions
HBAM: Hierarchical Bi-directional Word Attention Model
wenguanwang/deepattention
Deep Visual Attention Prediction (TIP18)