- Training language models to follow instructions with human feedback
- Zero-Resource Cross-Domain Named Entity Recognition
- Zero-Resource Cross-Lingual Named Entity Recognition
- ViT : An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
- CAM : Learning Deep Features for Discriminative Localization
- MoE : OUTRAGEOUSLY LARGE NEURAL NETWORKS: THE SPARSELY-GATED MIXTURE-OF-EXPERTS LAYER
- Swin Transformer : Swin Transformer: Hierarchical Vision Transformer using Shifted Windows