lsdefine/attention-is-all-you-need-keras
A Keras+TensorFlow Implementation of the Transformer: Attention Is All You Need
Python
Issues
- 0
- 8
startup error
#35 opened by rafaleo - 0
reshape may not match
#36 opened by pengxingang - 4
why get same output with different input?
#29 opened by maozezhong - 1
after embedding layer
#37 opened by lidongxing - 0
Time series forecasting?
#32 opened by salihgunduz - 0
layer norm end of the encoder?
#33 opened by salihgunduz - 0
- 1
- 0
Using the approach for video encoding.
#30 opened by kristosh - 2
Save model to json
#8 opened by jingyuanz - 0
the mask of attention
#28 opened by zjjzyl - 1
- 1
Skip-connection in Transformer
#17 opened by hoangcuong2011 - 2
- 1
seq2seq confused with shape
#26 opened by thomasyue - 0
ScaledDotProductAttention
#25 opened by t-kong - 0
the test demo
#24 opened by chenjun2hao - 0
dimension in GetSubMask
#23 opened by ichenjia - 0
Why wasn't K and V weren't passed from the top encoder to bottom decoder model?
#21 opened by ichenjia - 0
- 2
- 0
- 2
Issue with attention mask
#16 opened by LorrinWWW - 1
Reshape : Dimension mismatch
#15 opened by shashwattrivedi - 0
Decoding a sentence give same translation
#14 opened by mayurnewase - 2
maybe i find a point should be change
#12 opened by alphanlp - 4
How to perform translation?
#4 opened by lchunleo - 1
Keras and Tensorflow Versions
#9 opened by amirveyseh - 1
pure language model
#7 opened by XiaoLiuAI - 6
mask for decoder
#6 opened by XiaoLiuAI - 3
Issues with Keras Lambda Layers
#3 opened by wfmonster - 2
- 1
MultiHeadAttention
#2 opened by AMSakhnov - 1
LayerNormalization
#1 opened by AMSakhnov