GithubHelp home page GithubHelp logo

acl2022's Introduction

Non-Autoregressive Sequence Generation

Tutorial @ ACL 2022, May 22, 2022

Speakers

Jiatao Gu, Facebook AI Research, [email protected]
Xu Tan, Microsoft Research Asia, [email protected]

Abstract

Non-autoregressive sequence generation (NAR) attempts to generate the entire or partial output sequences in parallel to speed up the generation process and avoid potential issues (e.g., label bias, exposure bias) in autoregressive generation. While it has received much research attention and has been applied in many sequence generation tasks in natural language and speech, naive NAR models still face many challenges to close the performance gap between state-of-the-art autoregressive models because of a lack of modeling power. In this tutorial, we will provide a thorough introduction and review of non-autoregressive sequence generation, in four sections: 1) Background, which covers the motivation of NAR generation, the problem definition, the evaluation protocol, and the comparison with standard autoregressive generation approaches. 2) Method, which includes different aspects: model architecture, objective function, training data, learning paradigm, and additional inference tricks. 3) Application, which covers different tasks in text and speech generation, and some advanced topics in applications. 4) Conclusion, in which we describe several research challenges and discuss the potential future research directions. We hope this tutorial can serve both academic researchers and industry practitioners working on non-autoregressive sequence generation.

Outline

PART 1  Introduction (~ 20 minutes) 
  1.1 Problem definition 
  1.2 Evaluation protocol 
  1.3 Multi-modality problem 
  
PART 2  Methods  (~ 80 minutes) 
  2.1 Model architectures 
      2.1.1  Fully NAR models 
      2.1.2  Iteration-based NAR models 
      2.1.3  Partially NAR models 
      2.1.4  Locally AR models 
      2.1.5  NAR models with latent variables 
  2.2 Objective functions 
      2.2.1  Loss with latent variables 
      2.2.2  Loss beyond token-level 
  2.3 Training data 
  2.4 Learning paradigms 
      2.4.1  Curriculum learning 
      2.4.2  Self-supervised pre-training 
  2.5 Inference methods and tricks 
  
PART 3  Applications  (~ 60 minutes) 
  3.1 Task overview in text/speech/image generation
  3.2 NAR generation tasks     
      3.2.1  Neural machine translation   
      3.2.2  Text error correction
      3.2.3  Automatic speech recognition  
      3.2.4  Text to speech / singing voice synthesis
      3.2.5  Image (pixel/token) generation 
  3.3 Summary of NAR Applications  
      3.3.1  Benefits of NAR for different tasks
      3.3.2  Addressing target-target/source dependency 
      3.3.3  Data difficulty vs model capacity
      3.3.4  Streaming vs NAR, AR vs iterative NAR
      
PART 4  Open problems, future directions, Q\&A  (~ 20 minutes)

Materials

Slides

acl2022's People

Contributors

tan-xu avatar

Watchers

 avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.