/
 Are End-to-end Systems the Ultimate Solutions for NLP?  Are End-to-end Systems the Ultimate Solutions for NLP?

Are End-to-end Systems the Ultimate Solutions for NLP? - PowerPoint Presentation

kittie-lecroy
kittie-lecroy . @kittie-lecroy
Follow
343 views
Uploaded On 2020-04-05

Are End-to-end Systems the Ultimate Solutions for NLP? - PPT Presentation

Jing Jiang March 20 2018 CICLing Background Recent years have witnessed a fastgrowing trend of using deep learning solutions oftentimes endtoend for NLP tasks Machine translation Information extraction ID: 775660

systems analysis parsing model systems analysis parsing model feature question nlp training answering interpret amount large engineering deep solutions

Share:

Link:

Embed:

Download Presentation from below link

Download Presentation The PPT/PDF document " Are End-to-end Systems the Ultimate Sol..." is the property of its rightful owner. Permission is granted to download and print the materials on this web site for personal, non-commercial use only, and to display it on your personal computer provided you do not modify the materials and that you retain all copyright notices contained in the materials. By downloading content from our website, you accept the terms of this agreement.


Presentation Transcript

Slide1

Are End-to-end Systems the Ultimate Solutions for NLP?

Jing Jiang

March 20, 2018

CICLing

Slide2

Background

Recent years have witnessed a fast-growing trend of using deep learning solutions, oftentimes end-to-end, for NLP tasks.Machine translationInformation extractionQuestion answering Abstractive summarizationGood performanceNo feature engineeringRequires a large amount of training dataHard to interpret

2

Slide3

Example: Question Answering

3

Slide4

Question Answering

4

named entity recognition

question type analysis

syntactic parsing

semantic parsing

Slide5

End-to-end Question Answering

Starts from passages and questions as word sequencesUses deep neural networks for encoding, matching and predictionDoes not need named entity recognition, question type analysis, syntactic parsing, etc.On some benchmark dataset, best performance is close to human performance

5

Slide6

SQuAD Leaderboard

6

Slide7

Other Examples

Relation extractionTraditionally feature engineering involves POS tagging, constituency parsing, dependency parsing, etc.Recent work uses LSTM or CNN and position embedding, without feature engineeringHeadline generationRecent work uses sequence-to-sequence model trained on large amount of automatically obtained training dataNeural machine translation

7

Slide8

End-to-end Systems

Advantages:Eliminate the need to design subcomponents and featuresReduce error propagationResults are good when sufficient training data is used

8

Slide9

End-to-end Systems

Problems:Require a large amount of training data, which may not always be readily availableRequire careful tuningMay not be adaptable to a different dataset or domain / overfittingHard to interpret

9

Slide10

End-to-end Systems

Do you believe end-to-end systems will become the ultimate solutions to all NLP applications?Many intermediate steps such as morphological analysis, syntactic analysis or even discourse analysis would not be useful

10

Slide11

End-to-end Systems

Or do you believe end-to-end systems have their limitations?E.g, how do we share knowledge across different tasks?

11

Slide12

What Are Your Thoughts?

12

Slide13

How do you typically build your systems?

Feature engineering + traditional ML method (e.g., SVM)Mixture of traditional method and NN (e.g., incorporate POS tagging and parsing features into a neural network)End-to-end (i.e., no feature engineering)

13

Slide14

Is NN model useful for your task?

Have not tried it yetTried but not usefulUseful through word embeddings onlyUseful through models such as CNN and LSTM

14

Slide15

Interpretability

Do you think interpretability is important?Do you find it hard to interpret your NN model?Does error analysis help you come up with ways to improve your NN model?

15

Slide16

Tuning

Is a model considered good if it requires heavy tuning?Parameter sensitivity study?

16

Slide17

Benchmark datasets

Are we just chasing the numbers?Statistical significance tests?

17

Slide18

Challenges

What challenges do you face when adopting deep learning models for your NLP problem?

18