I am also affiliated with the Center for Language and Speech Processing.
Tal Linzen (2019). What can linguistics and deep learning contribute to each other? Language: Perspectives section. [arxiv]
Tal Linzen, Emmanuel Dupoux & Yoav Goldberg (2016). Assessing the ability of LSTMs to learn syntax-sensitive dependencies. Transactions of the Association for Computational Linguistics 4, 521–535. [link] [pdf] [bib]
Here's a video of a talk I gave in December 2018 at the Allen Institute for Artificial Intelligence, titled "Using cognitive science to evaluate and interpret neural language models".
Before joining JHU, I was a postdoctoral researcher in LSCP and IJN in the Cognitive Science Department at the École Normale Supérieure in Paris, where I worked with Emmanuel Dupoux and Benjamin Spector.
I obtained my Ph.D. in Linguistics in September 2015 from New York University, under the supervision of Alec Marantz. During my Ph.D., I also collaborated with Gillian Gallagher, Maria Gouskova and Liina Pylkkänen at NYU, as well as with Florian Jaeger at the University of Rochester.
Before that, I obtained a B.Sc. in Mathematics and Linguistics and an M.A. in Linguistics (with Mira Ariel), both from Tel Aviv University, and worked as a data scientist and software engineer.
February 8: LTI colloquium talk at Carnegie Mellon University.
December: Paper on interpreting neural network internal representations using tensor products accepted to ICLR.
Winter break talks: Yale Linguistics (Dec 10), Microsoft Research Redmond (Dec 13), Allen AI Institute (Dec 14), Google New York (Jan 7).
November 1: Co-organized the workshop on analyzing and interpreting neural networks for NLP (at EMNLP).
October 19: Colloquium talk at Georgetown University.
September 3-4: Visiting the University of Potsdam.
August: Two papers accepted to EMNLP.