Feature Request: Add Mish activation function
digantamisra98 opened this issue · 2 comments
Mish is a new novel activation function proposed in this paper.
It has shown promising results so far and has been adopted in several packages including:
- TensorFlow-Addons
- SpaCy (Tok2Vec Layer)
- Thinc - SpaCy's official NLP based ML library
- Echo AI
- Eclipse's deeplearning4j
- CNTKX - Extension of Microsoft's CNTK
- FastAI-Dev
- Darknet
- Yolov3
- BeeDNN - Library in C++
- Gen-EfficientNet-PyTorch
- dnet
All benchmarks, analysis and links to official package implementations can be found in this repository
It would be nice to have Mish as an option within the activation function group.
This is the comparison of Mish with other conventional activation functions in a SEResNet-50 for CIFAR-10: (Better accuracy and faster than GELU)
Hi,
thank you for your proposal.
We currently do not have a plan to implement Mish activation function, but you can easily implement it using our currently available functions:
mish_x = x * F.tanh(F.softplus(x))
@TE-andrewshin Thank you for the response. No worries. The implementation is more than helpful!