Conversational AI Design and Practice
GitHubAuthor
  • Preface
    • Syllabus
    • Schedule
  • 0. Getting Started
    • 0.1. Environment Setup
    • 0.2. Quiz
  • 1. Exploration
    • 1.1. Overview
    • 1.2. Project Ideas
    • 1.3. Quiz
  • 2. Dialogue Graph
    • 2.1. Emora STDM
    • 2.2. State Transition
    • 2.3. Matching Strategy
    • 2.4. Multi-turn Dialogue
    • 2.5. Quiz
  • 3. Contextual Understanding
    • 3.1. Natex
    • 3.2. Ontology
    • 3.4. Regular Expression
    • 3.5. Macro
    • 3.5. Quiz
  • 4. Interaction Design
    • 4.1. State Referencing
    • 4.2. Advanced Interaction
    • 4.3. Compound States
    • 4.4. Global Transition
    • 4.5. Saving and Loading
    • 4.6. Quiz
  • 5. LM-based Matching
    • 5.1. Language Models
    • 5.2. Quickstart with GPT
    • 5.3. Information Extraction
    • 5.4. Quiz
  • 6. Conversational Analysis
    • 6.1. H2H vs. H2M
    • 6.2. Team Evaluation
    • 6.3. Quiz
  • Project
    • Projects
    • Proposal Guidelines
    • Final Report Guidelines
  • Supplements
    • LINC Course
    • Page 1
Powered by GitBook

©2023 Emory University - All rights reserved

On this page
  • Statistical
  • Neural-based
  • Transformers
  • Tokenization
  • GPT (Generative Pre-trained Transformer)

Was this helpful?

Export as PDF
  1. 5. LM-based Matching

5.1. Language Models

Previous5. LM-based MatchingNext5.2. Quickstart with GPT

Last updated 2 years ago

Was this helpful?

Statistical

  • , Jurafsky and Martin, Chapter 3 in Speech and Language Processing (3rd ed.), 2023.

Neural-based

  • , Mikolov et al., ICLR, 2013. <- Word2Vec

  • , Pennington et al., EMNLP, 2014.

  • , Ppeters et al., NAACL, 2018. <- ELMo

Transformers

  • , Vaswani et al., NIPS, 2017. <- Transformer

  • , Liu et al., ICLR, 2018.

  • , Devlin et al., NAACL, 2018.

Tokenization

  • , Sennrich et al., ACL, 2016. <- Byte-Pair Encoding (BPE)

  • , Wu et al., arXiv, 2016. <- WordPiece

  • , Kudo and Richardson, EMNLP, 2018.

GPT (Generative Pre-trained Transformer)

, Radford et al., OpenAI, 2018. <- GPT-1

, Radford et al., OpenAI, 2019. <- GPT-2

, Brown et al., NeurIPS, 2020. <- GPT-3

N-gram Language Models
Efficient Estimation of Word Representations in Vector Space
GloVe: Global Vectors for Word Representation
Deep Contextualized Word Representations
Attention is All You Need
Generating Wikipedia by Summarizing Long Sequences
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Neural Machine Translation of Rare Words with Subword Units
Google’s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation
SentencePiece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing
Improving Language Understanding by Generative Pre-Training
Language Models are Unsupervised Multitask Learners
Language Models are Few-Shot Learners