|
|
|
|
|
199 |
|
|
|
|
|
198 |
|
|
|
|
|
197 |
|
|
|
|
|
196 |
|
|
|
|
|
195 |
|
|
|
|
|
194 |
|
|
|
|
|
193 |
|
|
|
|
|
192 |
|
|
|
|
|
191 |
|
|
|
|
|
190 |
|
|
|
|
|
189 |
|
|
|
|
|
188 |
|
|
|
|
|
187 |
|
|
|
|
|
186 |
|
|
|
|
|
185 |
|
|
|
|
|
184 |
|
|
|
|
|
183 |
|
|
|
|
|
182 |
|
|
|
|
|
181 |
|
|
|
|
|
180 |
|
|
|
|
|
179 |
|
|
|
|
|
178 |
|
|
|
|
|
177 |
|
|
|
|
|
176 |
|
|
|
|
|
175 |
VAEBM: Variational Autoencoders and Energy-based Models |
|
|
|
|
174 |
On Separability of Self-Supervised Representations |
|
|
|
|
173 |
Revisiting Knowledge Distillation via Label Smoothing Regularization |
|
|
|
|
172 |
Regularizing Class-wise Predictions via Self-knowledge Distillation |
|
|
|
|
171 |
Be Your Own Teacher: Improve CNN via Self Distillation |
|
|
|
|
170 |
以下的B站链接全部失效, 被我删了 |
|
|
|
|
|
Bayesian Deep Learning and a Probabilistic Perspective of Generalization |
|
|
|
|
169 |
Rethink Image Mixture for Unsupervised Visual Representation Learning |
|
|
|
|
168 |
FixMatch: Semi-Supervised Learning with Consistency and Confidence |
|
|
|
|
167 |
ReMixMatch: Semi-Supervised Learning with Distribution Alignment and Augmentation Anchoring |
|
|
|
|
166 |
MiCE: Mixture of Contrastive Experts for Unsupervised Image Clustering |
|
|
|
|
165 |
Learn Representation via Information Maximizing Self-Augmented Training |
|
|
|
|
164 |
Supporting Clustering with Contrastive Learning |
|
|
|
|
163 |
Unsupervised Multi-hop Question Answering by Question Generation |
|
|
|
|
162 |
Perceiver: General Perception with Iterative Attention |
|
|
|
|
161 |
Joint EBM Training for Better Calibrated NLU Models |
|
|
|
|
160 |
A Unified Energy-Based Framework for Unsupervised Learning |
|
|
|
|
159 |
Energy-Based Models for Deep Probabilistic Regression |
|
|
|
|
158 |
Contrastive Learning Inverts the Data Generating Process |
|
|
|
|
157 |
Asymmetric Loss For Multi-Label Classification |
|
|
|
|
156 |
Computation-Efficient Knowledge Distillation by Uncertainty-Aware Mixup |
|
|
|
|
155 |
Knowledge Distillation Meets Self-Supervision |
|
|
|
|
154 |
Feature Projection for Improved Text Classification |
|
|
|
|
153 |
Improve Joint Train of Inference Net and Structure Predict EnergyNet |
|
|
|
|
152 |
BertGCN: Transductive Text Classification by Combining GCN and Bert |
|
|
|
|
151 |
The Authors Matter Understand Mitigate Implicit Bias in text classification |
|
|
|
|
150 |
Learning Approximate Inference Networks for Structured Prediction |
|
|
|
|
149 |
End-to-End Learning for Structured Prediction Energy Networks |
|
|
|
|
148 |
Revisiting Unsupervised Relation Extraction |
|
|
|
|
147 |
Sentence Meta-Embeddings for Unsupervised Semantic Textual Similarity |
|
|
|
|
146 |
X-Class: Text Classification with Extremely Weak Supervision |
|
|
|
|
145 |
Paint by Word |
|
|
|
|
144 |
Shape-Texture Debiased Neural Network Training |
|
|
|
|
143 |
Contrastive Learning through Alignment and Uniformity on the Hypersphere |
|
|
|
|
142 |
Deep INFOMAX representation mutual information estimation maximization |
|
|
|
|
141 |
SimCSE: Simple Contrastive Learning of Sentence Embeddings |
|
|
|
|
140 |
IMOJIE Iterative Memory-Based Joint Open Information Extraction |
|
|
|
|
139 |
Trash is Treasure Resisting Adversarial Examples by Adversarial Examples |
|
|
|
|
138 |
Enhancing Adversarial Defense by k-Winners-Take-All |
|
|
|
|
137 |
On Adaptive Attacks to Adversarial Example Defenses |
|
|
|
|
136 |
Knowledge distillation via softmax regression representation learning |
|
|
|
|
135 |
Revisiting Locally Supervised Learning Alternative to End-to-end Training |
|
|
|
|
134 |
Putting An End to End-to-End Gradient-Isolated Learning of Representations |
|
|
|
|
133 |
防御defense变分自编码器 |
|
|
|
|
132 |
Triple Wins Accuracy Robustness Efficiency by Input-adaptive Inference |
|
|
|
|
131 |
Using latent space regression to analyze leverage compositionality in GANs |
|
|
|
|
130 |
Theoretically(没看) Principled Trade-off between Robustness and Accuracy |
|
|
|
|
129 |
Representation learning with contrastive predictive coding |
|
|
|
|
128 |
Learning Representations for Time Series Clustering |
|
|
|
|
127 |
Stochastic Security: Adversarial Defense Using Long-Run Dynamics of EBM |
|
|
|
|
126 |
Improving Adversarial Robustness via Channel-wise Activation Suppressing |
|
|
|
|
125 |
Likelihood Landscapes: A Unifying Principle Behind Adversarial Defenses |
|
|
|
|
124 |
Barlow Twins: Self-Supervised Learning via Redundancy Reduction |
|
|
|
|
123 |
Geometry-Aware Instance-Reweighted Adversarial Training |
|
|
|
|
122 |
A Closer Look at Accuracy vs Robustness |
|
|
|
|
121 |
Unsupervised Clustering of Seismic Signals 地震波 using autoencoders |
|
|
|
|
120 |
Towards the first adversarially robust neural network model on MNIST |
|
|
|
|
119 |
PGD对抗训练 Towards Deep Learning Models Resistant to Adversarial Attacks |
|
|
|
|
118 |
Denoising Diffusion Probabilistic Models |
|
|
|
|
117 |
Deep Unsupervised Learning using Nonequilibrium Thermodynamics |
|
|
|
|
116 |
Variational Inference with Normalizing Flows |
|
|
|
|
115 |
CutMix Regularization Strategy with Localizable Features |
|
|
|
|
114 |
Clustering-friendly Representation Learning Feature Decorrelate |
|
|
|
|
113 |
Energy-based Out-of-distribution Detection |
|
|
|
|
112 |
High-Performance Large-Scale Image Recognition Without Normalization |
|
|
|
|
111 |
Characterizing signal propagation in unnormalized ResNets |
|
|
|
|
110 |
Concept Learners for Few-Shot Learning |
|
|
|
|
109 |
Image Generation by Minimize Frechet Distance in Discriminator feature space |
|
|
|
|
108 |
Learning Non-Convergent Non-Persistent Short-Run MCMC to EBM |
|
|
|
|
107 |
Concept Whitening for Interpretable Image Recognition |
|
|
|
|
106 |
Loss Landscape Sightseeing with Multi-Point Optimization |
|
|
|
|
105 |
Loss Surfaces, Mode Connectivity, and Fast Ensembling of DNNs |
|
|
|
|
104 |
Essentially No Barriers in Neural Network Energy Landscape |
|
|
|
|
103 |
Visualizing the Loss Landscape of Neural Nets |
|
|
|
|
102 |
Self-training for Few-shot Transfer Across Extreme Task Differences |
|
|
|
|
101 |
Darts: Differentiable architecture search |
|
|
|
|
100 |
Architecture Search Space in Neural Architecture Search(NAS) |
|
|
|
|
99 |
Free Lunch for Few-shot Learning: Distribution Calibration |
|
|
|
|
98 |
Online Deep Clustering for Unsupervised Representation Learning |
|
|
|
|
97 |
Coarse-to-Fine Pre-training for Named Entity Recognition |
|
|
|
|
96 |
Unsupervised Domain Adaptation with Variational Information Bottleneck |
|
|
|
|
95 |
A Unified MRC Framework for Named Entity Recognition |
|
|
|
|
94 |
Rethinking Softmax Cross-Entropy Loss for Adversarial Robustness |
|
|
|
|
93 |
Super-Convergence: Very Fast Training of NN use large LR |
|
|
|
|
92 |
Contrastive Clustering |
|
|
|
|
91 |
Graph Contrastive Learning with Augmentations NIPS 2020 / Graph Contrastive Learning with Adaptive Augmentation WWW 2021/ GCC: Graph Contrastive Coding for Graph Neural Network ... KDD 2021 |
GNN and Contrastive Learning |
|
|
|
90 |
Contrastive Representation Distillation |
|
|
|
|
89 |
Spectral Norm Regularization for Improving the Generalizability of NN |
|
|
|
|
88 |
UDA: Unsupervised Data Augmentation for Consistency Training |
|
|
|
|
87 |
Simplify the Usage of Lexicon in Chinese NER |
|
|
|
|
86 |
Chinese NER Using Lattice LSTM |
|
|
|
|
85 |
Uncertainty-aware Self-training for Few-shot Text Classification |
|
|
|
|
84 |
Concept Learning with Energy-Based Models |
|
|
|
|
83 |
Training data-efficient image transformers distillation through attention |
|
|
|
|
82 |
Adversarial Training Methods for Semi-Supervised Text Classification |
|
|
|
|
81 |
Delta-training Semi-Supervised Text Classification with word embedding |
|
|
|
|
80 |
On the Anatomy of MCMC-Based Maximum Likelihood Learning of EBMs |
|
|
|
|
79 |
Unsupervised Deep Embedding for Clustering Analysis |
|
|
|
|
78 |
Relation of Relation Learning Network for Sentence Semantic Matching |
|
|
|
|
77 |
Contextual Parameter Generation for Universal Neural Machine Translation |
|
|
|
|
76 |
Exploring Simple Siamese Representation Learning |
|
|
|
|
75 |
Contextual Parameter Generation for Knowledge Graph Link Prediction |
|
|
|
|
74 |
Robustness May Be at Odds with Accuracy |
|
|
|
|
73 |
Learning with Multiplicative Perturbations |
|
|
|
|
72 |
When Do Curricula Work? |
|
|
|
|
71 |
Self-Supervised Contrastive Learning with Adversarial Examples |
|
|
|
|
70 |
Supervised Contrastive Learning |
|
|
|
|
69 |
A Note on the Inception Score and FID |
|
|
|
|
68 |
Hierarchical Semantic Aggregation for Contrastive Representation Learning |
|
|
|
|
67 |
Syntactic and Semantic-driven Learning for Open Information Extraction |
|
|
|
|
66 |
Text Classification with Negative Supervision |
|
|
|
|
65 |
CESI Canonicalizing Open Knowledge Bases by Embeddings and Side Information |
|
|
|
|
64 |
CaRe: Open Knowledge Graph Embedding |
|
|
|
|
63 |
No MCMC for me, Amortized sampling for fast and stable training of EBMs |
|
|
|
|
62 |
Knowledge Graph Embedding Based Question Answering |
|
|
|
|
61 |
VAT Virtual Adversarial Training for regularization semi-supervised learn |
|
|
|
|
60 |
CNN-Generated Images Are Surprisingly Easy to Spot.. For Now |
|
|
|
|
59 |
Graph Agreement Models for Semi-Supervised Learning |
|
|
|
|
58 |
Be More with Less: Hypergraph Attention Networks for Inductive 文本分类 |
|
|
|
|
57 |
Text Level Graph Neural Network for Text Classification |
|
|
|
|
56 |
Graph Convolutional Networks for Text Classification |
|
|
|
|
55 |
Learning sparse neural networks through L0 regularization |
|
|
|
|
54 |
BigGAN: Large Scale GAN Training for High Fidelity Natural Image Synthesis |
|
|
|
|
53 |
On the steerability of generative adversarial networks |
|
|
|
|
52 |
What Makes for Good Views for Contrastive Learning |
|
|
|
|
51 |
Viewmaker Networks Learning Views for Unsupervised Representation Learning |
|
|
|
|
50 |
Auto-Encoding Variational Bayes |
|
|
|
|
49 |
Adversarial Examples Improve Image Recognition |
|
|
|
|
48 |
Stochastic Weight Averaging for Generalization |
|
|
|
|
47 |
There Are Many Consistent Explanations Of Unlabeled Data |
|
|
|
|
46 |
Interpretable Convolutional Neural Networks |
|
|
|
|
45 |
Understanding Black-box Predictions via Influence Functions |
|
|
|
|
44 |
Adversarial Examples Are Not Bugs, They Are Features |
|
|
|
|
43 |
You Only Propagate Once Accelerating AT via Maximal Principle |
|
|
|
|
42 |
Text Classification Using Label Names Only A LM self-training way |
|
|
|
|
41 |
10篇softmax,CrossEntropyLoss替代方法论文合集 |
|
|
|
|
40 |
Cyclical Stochastic Gradient MCMC and snapshot ensemble |
|
|
|
|
39 |
Unsupervised Feature Learning via Non-Parametric Instance Discrimination |
|
|
|
|
38 |
An Image is Worth 16x16 Words Transformers for Image Recognition at Scale |
|
|
|
|
37 |
Training independent subnetworks for robust prediction |
|
|
|
|
36 |
Active Learning for CNNs: A Core-Set Approach |
|
|
|
|
35 |
SimCLR A Simple Framework for Contrastive Learning of Visual Representation |
|
|
|
|
34 |
UNITER: UNiversal Image-TExt Representation Learning |
|
|
|
|
33 |
Image Synthesis with a Single (Robust) Classifier |
|
|
|
|
32 |
Set Transformer A Framework for Attention-based Permutation-Invariant NN |
|
|
|
|
31 |
Consistency Regularization in Semi-Supervised Learning |
|
|
|
|
30 |
Did the model understand the question |
|
|
|
|
29 |
Rethinking Feature Distribution for Loss Functions in Image Classification |
|
|
|
|
28 |
Bootstrap your own latent: A new way to self supervised learning |
|
|
|
|
27 |
Hybrid Discriminative-Generative Training via Contrastive Learning(EBMs) |
|
|
|
|
26 |
A Multimodal Translation-Based Approach for Knowledge Graph Representation (ACL 2018) |
|
|
|
|
25 |
Deep Bayesian Active Learning with Image Data (ICML 2017) The power of ensembles for active learning in image classification (CVPR 2018) |
Bilibili |
|
|
|
24 |
SCAN Learnrnning to Classify Images without Labels (ECCV 2020) |
|
|
|
|
23 |
Unsupervised Question Answering by Cloze Translation (ACL 2019) |
|
|
|
|
22 |
Phrase-Based & Neural Unsupervised Machine Translation (EMNLP 2018) |
|
|
|
|
21 |
MixUp as Locally Linear Out-Of-Manifold Regularization (AAAI 2019) |
|
|
|
|
20 |
Manifold Mixup: Better Representations by Interpolating Hidden States ICML2019 |
|
|
|
|
19 |
Bag of Tricks for Image Classification with CNN |
|
|
|
|
18 |
On Mixup Training Improved Calibration for DNN |
|
|
|
|
17 |
BERT: Pre-training of Deep Bidirectional Transformers |
|
|
|
|
16 |
Rationalizing Neural Predictions (EMNLP2016) |
|
|
|
|
15 |
Attention is all you need, Transformer (NIPS 2017) |
|
|
|
|
14 |
Learn To Pay Attention (ICLR 2018) |
|
|
|
|
13 |
A Self-Training Method for MRC with Soft Evidence Extraction(ACL 2019) |
Bilibili |
Youtube |
|
|
12 |
Deep Fool(CVPR2016) 和 Deep Defense(NIPS 2018) |
Bilibili |
youtube |
|
|
11 |
R-Trans RNN Transformer Network for 中文机器阅理解(IEEE-Access) |
Bilibili |
youtube |
|
|
10 |
一系列Energy-based models 能量模型论文摘要简介 |
Bilibili |
Youtube |
|
|
9 |
Implicit Generation and Modeling with EBM(NIPS 2019) |
Bilibili |
Youtube |
|
|
8 |
MixMatch A Holistic Approach to Semi-supervised Learning(NIPS 2019) |
Bilibili |
Youtube |
Arxiv |
|
7 |
Obfuscated Gradients Give a False Sense of Security(ICML2017 best reward) |
Bilibili |
Youtube |
Arxiv |
|
6 |
Explaining and Harnessing Adversarial Examples(ICLR 2015) |
Bilibili |
Youtube |
Arxiv |
|
5 |
Imagenet-Trained CNNS are Biased Towards Texture(ICLR2018) |
Bilibili |
Youtube |
Arxiv |
|
4 |
Momentum Contrast for Unsupervised Visual Representation Learning(CVPR2020) |
Bilibili |
Youtube |
Arxiv |
|
3 |
Mixup: Beyond Empirical Risk Minimization(ICLR2018) |
Bilibili |
Youtube |
Arxiv |
|
2 |
Your Classifier is secretely an Energy Based Model(ICLR 2019) |
Bilibili |
Youtube |
Arxiv |
|
1(2020-08-19) |