ΜΗΠΔTV

Skip to main content Skip to navigation

syllabus__text_data_in_economics_warwick

QUAPEC

Syllabus

Text Data in Economics

ΜΗΠΔTV Summer School 2022

 

Teaching Team

Instructor: , ashe@ethz.ch
TA:ΜύΜύΜύΜύΜύ ΜύΜύΜύΜύΜύΜύ Claudia Marangon, claudia.marangon@gess.ethz.ch

 

Schedule

Lectures: XX

TA office hrs: XX

 

Important Links

(bit.ly/NLP-git)

Problem Set

 

Learning Objectives:

LO1. Implement and evaluate text-as-data methods.

LO2. Evaluate the use of text-analysis tools in economics research.

LO3. Plan a research project using text data.

 

Course Format

●     8 lectures on zoom (12 hours)

●     In-person workshopping of student project papers

 

Assignments:

●     Problem set

●     Referee report on one of the course readings

●     Research proposal on a text-data project (first and second draft)

 

 

 


 

 

Topics Outline and Main Economics Papers Readings

 

  1. Overview
    1. Gentzkow, Kelly, and Taddy, β€œ.”
    2. Ash and Hansen, Text Algorithms in Economics
  2. Dictionaries (Macro)
    1. Baker, Bloom, and Davis (2016),
    2. Hassan, Hollander, Van Lent, and Tahoun,
  3. Dictionaries (Micro)
    1. Michalopoulous and Xue (2019),
    2. Enke 2020,
    3. Djourelova, Media persuasion through slanted language: Evidence from the coverage of immigration
    4. Truffa and Wong (2021),
    5. Advani, Ash, Cai, and Rasul (2022), Race-Related Research in Economics.
  4. Document Distance
    1. Kelly, Papanikolau, Seru, and Taddy, .
    2. Cage, Herve, and Viaud,
  5. Topic Models
    1. Hansen, McMahon, and Prat, .
    2. Ash, Morelli, and Vannoni, β€œβ€
  6. Supervised Learning
    1. Gentzkow and Shapiro (2010), .
    2. Gentzkow, Shapiro, and Taddy (2019)
    3. Widmer, Galletta, and Ash, Media Slant is Contagious
  7. Word Embeddings (2 classes)
    1. Ash, Chen, and Ornaghi (2022)
    2. Gennaro and Ash, Emotion and Reason in Political Language (2021) and Transparency and Emotionality in Politics: Evidence from C-SPANΜύ (2022).
    3. Ash, Gennaro, Hangartner, and Stampi-Bombelli (2022)
    4. Kozlowsky et al
  8. Syntactic and Semantic Parsing
    1. Ash, Gauthier, and Widmer, Text semantics capture political and economic narratives

 

Learning Materials

Books

●     Natural Language Processing in Python, Third Edition (β€œNLTK Book”).

β—‹     Available at .

β—‹     Classic treatments of traditional NLP tools.

●     Aurelien Geron, Hands-on Machine Learning with Scikit-Learn, Keras, and TensorFlow (2019)

β—‹     , should be available with an academic account using ETH email.

β—‹     A great practical book for machine learning and deep learning in Python, but not NLP-focused. We will use material from Chapters 2-4, 7-11, 13, and 15-17.

β—‹     The deep learning chapters use Keras + TensorFlow.

β—‹    

●     Yoav Goldberg, Neural Network Methods for Natural Language Processing (2017)

β—‹     (email me if this doesn’t work)

β—‹     A more advanced theoretical treatment of neural networks with an NLP focus, but already somewhat dated. We will use material from Chapters 1-17 and 19.

●     Jurafsky and Martin, Speech and Language Processing (3d Ed. 2019).

β—‹     .

β—‹     The standard theory text on computational linguistics.

 

Programming

Python is probably the best option for NLP, used by most data scientists. All the sample code is in Python. You are welcome to use another programming language.

●     New to Python?

β—‹    

β—‹    

β—‹    

β—‹    

β—‹    

β—‹    

β—‹    

β—‹    

β—‹    

●     New to Machine Learning?

β—‹    

β—‹     Read the Geron Book, Chapters 1-7

β—‹    

●     New to Text Mining / NLP?

β—‹    

β—‹     Read the , Chapters 1-5

β—‹    

●    

β—‹     Lists of papers with replication repos.

●     Want to use R instead?

β—‹     is popular for text analysis among political scientists.

●     Other resources:

β—‹    

β—‹    

 

Python Libraries

pip install pandas seaborn scikit-learn tensorflow nltk gensim flair spacy transformers

●     Basics:

β—‹     : data loading and management

β—‹     : visualization

β—‹     : general purpose Python ML library

β—‹     : deep learning library

●     NLP Necessities:

β—‹     : standard NLP tools

β—‹     : topic models and embeddings

β—‹     : tokenization, NER, syntactic parsing, word vectors

β—‹     : sentiment analysis and some other tools ()

β—‹     : transformer architectures

●     Specialized tools:

β—‹     : library of models for semantic role labeling, entailment, question answering, etc

β—‹     : library of embeddings

β—‹     : interface from spaCy to huggingface

 


 

References

Yellow highlighting indicates required reading
Blue highlighting indicates recommended methods reading

 

Reference (Overview):

●     Gentzkow, Kelly, and Taddy, β€œ.”

●     Goldberg, Ch. 1

●     NLTK book, Chapters 1, 2, 4

●     Grimmer and Stewart, β€œ.”

●     Raschka, β€œβ€.

 

Reference (Dictionary Methods):

●    

 

Reference: ΜύTokenization

●    

●     Goldberg, Ch. 6

●     NLTK book, Chapter 3, 5, 7, 8

●    

●     Denny and Spirling, β€œ.”

 

 

Reference (Dimensionality Reduction):

●     Geron, Chapters 8-9

●     Gilis, .

Methods (Document Distance):

●     Lee et al, .

●     Brandon Rose, β€œ.”

Methods (Topic Models)

●     Prabhakaran,

●     Quinn, et al, .

●     Roberts et al, β€œβ€.

●     Christian Fong and Justin Grimmer, β€œ.”

 

Reference (Machine Learning):

●     Goldberg Ch. 2, 7

●    

●     NLTK book, chapter 6

●     Geron, Chapters 2-4, 7

 

 

Overview (Deep Learning for NLP)

●     Sebastian Ruder,

Reference (Neural Nets):

●    

●     Goldberg, Ch. 3-5

●     Geron, Chapters 10-11

●     Leslie Smith,

●     Chris Olah,

●     Baldi and Sadowski,

Reference (Embedding Layers):

●     Goldberg, Ch. 8

●    

References (RNNs):

●     Geron Ch. 15-17

●     Goldberg, Ch. 14-17

●     Sutskever, Vinyals, and Le,

●     Michael Nguyen,

●     Andrej Karpathy,

●     Chang and Masteron, .

Reference (Model Interpretation)

●     Ribeiro, Singh, and Guestrub, .

●    

Applications (MLP):

●     Vamossy,

●     Meursault,Μύ

Applications (RNN):

●     [short] Iyyer et al, .ΜύΜύΜύΜύΜύ

●     Ash et al,

 

Reference (Word Embeddings):

●     Spirling and Rodriguez, .

●     Goldberg, Ch. 10-11

●     Chapter Yoav Goldberg and Omer Levy, β€œβ€.

●     Piero Molino, β€œβ€.

●     Matt Kusner, Yu Sun, Nicholas Kolkin, and Killian Weinberger, β€œβ€.

●     Sanjeev Arora, Yuanzhi Li, Yingyu Liang, Tengyu Ma, Andrej Risteski,

●     Allen and Hospedales,

●     Ruder,

●     Peters, Ruder, and Smith, .

●    

●     Bojanowski et al, .

●     Antoniak and Mimno, .

●     Ash, Chen, and Ornaghi,

●     Hamilton, Clark, Leskovec, and Jurafsky, 2016,Μύ .

Tools (Word Embeddings)

●    

Contextualized Word Embeddings:

●     Peters et al, .

●    

 

Reference (Syntactic Parsing):

●     NLTK Book,

●     , Chapters 12-15, 20

●    

Reference (Semantic Role Labeling):

●     Jurafsky and Martin,

●    

 

Tools (Syntactic Parsing):

●    

Tools (Semantic Role Labeling):

●    

 

Tools for Document Embeddings

●     Ruder,

●    

●    

References (Document Embeddings):

●     Arora, Liang, and Ma, β€œ.”

●     Doc2Vec:

β—‹     Le and Mikolov, β€œ.”

β—‹    

β—‹    

β—‹    

●     Wu et al,

●     Bhatia, Lau, and Baldwin, β€œβ€

●     InferSent

●     USE:

β—‹     Cer et al, , ()

β—‹     Yang et al, .

●     Clark, Celikyilmaz, and Smith, .

References (Attention / Transformers)

●     Bloem,

●     Ruder,

●    

●     Geron, Chapter 16

●     Goldberg, Ch. 17

●     Vaswani et al,

 

●     BERT:

β—‹     Devlin et al, .

β—‹    

β—‹    

β—‹     Reimers and Guryevich,

β—‹     Nie et al,

β—‹     Hassan et al,

 

Reference (Language Models):

●     Goldberg, Ch. 9, 17

Reference (Transformers):

●    

Reference (Autoregressive Models)

●     Shree,

●     Radford et al, .

β—‹    

●     .

●     .

●     Brown et al, (GPT-3)

●     (open-sourced GPT-3)

Reference: Conditioned Text Generation

●    

●     Dathathri et al,

●     CTRL

 

Reference (Sequence-to-Sequence Transformers)

●    

●     Lewis et al,

●     Raffel et al,

●     Narang et al,

 

Reference (Coreference Resolution):

●     Jurafsky and Martin,

Reference (Textual Entailment)

●      

Reference (Siamese Neural Nets)

●    

Reference (Discourse):

●     Jurafsky and Martin,

Reference (Dialogue):

●     Li et al, .

●    

β—‹    

●     Luo et al,

 

Big Bird

●    

Reference (Information Extraction):

●     NLTK Book,

●     Jurafsky and Martin,

●     Angeli et al, .

●    

●     Qin et al,

Reference (Knowledge Graphs)

●    

●     Nickel,

●     Joulin et al,

●     Lehmann et al, .

●     ()

●    

●     Yao et al,

Reference (Summarization):

●     Gabriel et al,

●     See et al,

●    

●     Stiennon et al,

Reference (Question Answering):

●     Jurafsky and Martin,

●    

●    

Reference (Claim Checking):

●     Vlachos, e-Fever

Language Model Interpretation:

●      

 

Reference (Legal AI)

●     Zhong et al,

Methods (Causal Inference with Text):

●     Keith et al,

●     Wood-Doughty et al,

●     Egami, Fong, Grimmer, Roberts, and Stewart,

 


 

Additional Applications

 

Complexity in Text):

●     Katz and Bommarito, β€œ.”

●     Katz et al,

●     Benoit, Munger, and Spirling (2017), β€œβ€.

●     [short] Louis and Nenkova (2013),

 

Dictionary Methods:

●     Michalopoulous and Xue (2019),

●     Baker, Bloom, and Davis (2016),

●     Djourelova,

●     Enke 2020,

●     Cao et al,

Tokens/N-grams:

●     Gentzkow and Shapiro (2010), .

●     Ash, Morelli, and Van Weelden, .

 

Document Distance:

●     Kelly, Papanikolau, Seru, and Taddy, .

●     Hoberg and Phillips, .

 

Topic Models

●     Barron, Huang, Spang, and DeDeo, . [has appendix]

●     Hansen, McMahon, and Prat, .

●     Ash, Morelli, and Vannoni, β€œβ€

 

 

Text Classification:

●     Osnabrugge, Ash, and Morelli,

●     Widmer, Galletta, and Ash,

●     Gentzkow, Shapiro, and Taddy (2019), β€œ.”

●     Kelly et al,

●     Peterson and Spirling,

 

Word Embeddings:

●     Ash, Chen, and Ornaghi,

●     Gennaro and Ash,

●     Caliskan et al, β€œβ€

●     Bolukbasi et al, .

●     Kozlowski, Taddy, and Evans 2019,

●     Stoltz and Taylor,

●     [short] Gillani and Levy, .

●     Garg et al 2018, [includes appendix]

●     [short] Lucy et al,

●     Thompson et al, [includes appendix]

●     Rheault and Cochrane, .

●     Nyarko and Sanga, .

 

Syntactic Parsing:

●     Hoyle et al, .

●     [short] Ash, Jacobs, MacLeod, Naidu, and Stammbach, .

●     Vannoni, Ash, and Morelli,

●     Michael Webb,

 

Semantic Role Labeling:

●     Ash, Gauthier, and Widmer, Mining narratives from large text corpora

●     Fetzer,

 

Information Extraction:

●     [short] Surdeanu et al 2011, .

●     [short] Jurafsky and Chambers, .

●     [short] Clark, Ji, and Smith, .

●     [short] Bamman and Smith, .

●     [short] Wyner and Peters, .

●     [short] Xia and Ding,

 

Document Embeddings:

●     [short] Demzsky et al, 2019,

●     [short] Dai, Olah, and Le, .

●     Ash and Chen 2018, .

●     Galletta, Ash, and Chen,

●     Ash, Chen, and Naidu (2020), β€œ.”

●     [short] Tong et al, .

Transformer Classification:

●     Bingler et al,

●     Pei and Jurgens,

 

Language Models:

●     [short] Peric, Mijic, Stammbach, and Ash,

●     Kreps, McCain, and Brundage, .

●     [short] Peng et al,

●     [short] Adeem, Bethky, Reddy,

 

Local Semantics

●     Ross et al, [short]

●     Prabhakaran et al, .

 

Global Semantics

●     [short] Stammbach and Ash,

●     Chen et al,

●     Vold and Conrad, .

 

Causal Inference with Text:

●     Margaret Roberts, Brandon Stewart, and Richard Nielsen, β€œβ€

●     [short] Veitch et al,

●     All the papers in Table 1 .

●     Zeng et al, .

 

Argument Mining

●     [short] Subramanian et al,

 

Quote Extraction

●     Newell et al,

 

Let us know you agree to cookies