(Bi-)Weekly NLP Research Paper Series

Direct Submission ARR Commit Author Response Notification Conference Notice
SIGDIAL 05/11 06/18 - 07/02 09/07 - 09/09 Edinburgh
COLING 05/17 - - 08/15 10/12 - 10/15 Gyeongju, Korea
EMNLP 06/24 07/24 08/23 - 08/29 10/06 12/07 - 12/11 Abu Dhabi, (ARR Withdraw: 05/24)
AACL 07/15 08/21 08/15 - 08/21 09/20 11/21 - 11/24 Taiwan
ACL Rolling Review 06/01, 07/15, 09/01, 10/15, 12/01, 01/15/2023

(Conference deadlines: https://aideadlin.es/?sub=ML,CV,NLP,RO,SP or https://ccfddl.github.io/)

Goals:

  • Primary for sharing knowledge across different domains and catching up on recent updates.
  • Contents:
    • Mainly and only collect interesting papers.
    • Summarize the approaches and frameworks.
    • Write strengths and weaknesses, and share potential applications to other domains.
    • Highlight some exciting papers. Template :
      • Title: the paper title
      • Summary: strengths and weaknesses
      • Deserve to note: specific paragraphs or designs deserve to be noted or further reading

🤖 Schedule:

  • This document will keep updating and release (bi-)weekly every Friday.

❤️ Welcome:

  • You are more than welcome to invite any people and edit any parts of the documents, including but not limited to deleting, adding, and modifying any parts.

🚀 Week 03 05/09/2022

Dialogue & Multi-modal

Deepmind

Microsoft

OpenAI

Question Answering & Retrieval

  1. RETRO (Deepmind): Borgeaud, Sebastian, Arthur Mensch, Jordan Hoffmann, Trevor Cai, Eliza Rutherford, Katie Millican, George van den Driessche et al. "Improving language models by retrieving from trillions of tokens." arXiv preprint arXiv:2112.04426 (2021). [pdf]

🚀 Week 02 04/29/2022

Dialogue Related Papers

RL for Dialog (NAACL 2022) - BY Sergey Levine

Seeker and it relevant papers

Conversational Recommendation:

ARR April 2022

WSDM 2022

CRS Lab

Question Answering

A Memory Efficient Baseline for Open Domain Question Answering

🚀 Week 01 04/22/2022

Dialogue Related Papers

Question Answering

Towards Unsupervised Dense Information Retrieval with Contrastive Learning

  • It evaluates the models on the BEIR benchmark, where the benchmark contains 18 retrieval datasets with a focus on diversity. Most datasets do not contain a training set and the focus of the benchmark is zero-shot retrieval.
  • It shows SOTA performances on unsupervised learning and few-shot learning. The unsupervised pre-training alone outperforms BERT with intermediateMS-MARCOfine-tuning.
  • Deserve to note:
    • It explores the limits of contrastive learning as a way to train unsupervised dense retrievers, and show that it leads to strong retrieval performance.
    • The ways to build positive pairs and negative pairs are interesting.
      • Building positive pairs from a single document: (1) Inverse Cloze Task: it uses the tokens of the span as the query and the rest of the tokens as the document (or key); (2) Independent cropping: It samples independently two spans from a document to form a positive pair.
      • Building large set of negative pairs: (1) Negatively pairs within a batch based on SimCLR. (2) Negative pairs across batches where queries are generated from the elements of the current batch and keys are the elements stored in the queue. The technique is proposed by MoCO.

Improving Passage Retrieval with Zero-Shot Question Generation LOOPITR: Combining Dual and Cross Encoder Architectures for Image-Text Retrieval RocketQA: An Optimized Training Approach to Dense Passage Retrieval for Open-Domain Question Answering Augmented SBERT: Data Augmentation Method for Improving Bi-Encoders for Pairwise Sentence Scoring Tasks Improving Efficient Neural Ranking Models with Cross-Architecture Knowledge Distillation In-Batch Negatives for Knowledge Distillation with Tightly-Coupled Teachers for Dense Retrieval Improving Bi-encoder Document Ranking Models with Two Rankers and Multi-teacher Distillation

Conversational Recommendation Systems

Two tutorials:

WSDM 2022

ACL ARR April

Recommendation Systems

CV & Multi-modal