research
∙
05/17/2023
Token-wise Decomposition of Autoregressive Language Model Hidden States for Analyzing Model Predictions
While there is much recent interest in studying why Transformer-based la...
research
∙
04/22/2023
Transformer-Based LM Surprisal Predicts Human Reading Times Best with About Two Billion Training Tokens
Recent psycholinguistic studies have drawn conflicting conclusions about...
research
∙
12/23/2022
Why Does Surprisal From Larger Transformer-Based Language Models Provide a Poorer Fit to Human Reading Times?
This work presents a detailed linguistic analysis into why larger Transf...
research
∙
12/21/2022