


default search action
6th BlackboxNLP@EMNLP 2023: Singapore
- Yonatan Belinkov, Sophie Hao, Jaap Jumelet, Najoung Kim, Arya McCarthy, Hosein Mohebbi:

Proceedings of the 6th BlackboxNLP Workshop: Analyzing and Interpreting Neural Networks for NLP, BlackboxNLP@EMNLP 2023, Singapore, December 7, 2023. Association for Computational Linguistics 2023, ISBN 979-8-89176-052-3 - Anthony M. Colas, Jun Araki, Zhengyu Zhou, Bingqing Wang, Zhe Feng:

Knowledge-Grounded Natural Language Recommendation Explanation. 1-15 - Neel Nanda, Andrew Lee, Martin Wattenberg:

Emergent Linear Representations in World Models of Self-Supervised Sequence Models. 16-30 - Chandan Singh, John X. Morris, Jyoti Aneja, Alexander M. Rush, Jianfeng Gao:

Explaining Data Patterns in Natural Language with Language Models. 31-55 - Akshat Gupta:

Probing Quantifier Comprehension in Large Language Models: Another Example of Inverse Scaling. 56-64 - Stefan Arnold, Nils Kemmerzell, Annika Schreiner:

Disentangling the Linguistic Competence of Privacy-Preserving BERT. 65-75 - Antoine Chaffin, Julien Delaunay:

"Honey, Tell Me What's Wrong", Global Explanation of Textual Discriminative Models through Cooperative Generation. 76-88 - Henning Bartsch, Ole Jorgensen, Domenic Rosati, Jason Hoelscher-Obermaier, Jacob Pfau:

Self-Consistency of Large Language Models under Ambiguity. 89-105 - Hao Sun, John Hewitt:

Character-Level Chinese Backpack Language Models. 106-119 - Sunit Bhattacharya, Ondrej Bojar:

Unveiling Multilinguality in Transformer Models: Exploring Language Specificity in Feed-Forward Networks. 120-126 - Timothee Mickus

, Raúl Vázquez
:
Why Bother with Geometry? On the Relevance of Linear Decompositions of Transformer Embeddings. 127-141 - Dmitry Nikolaev

, Sebastian Padó:
Investigating Semantic Subspaces of Transformer Sentence Embeddings through Linear Structural Probing. 142-154 - Juanhe (TJ) Tan:

Causal Abstraction for Chain-of-Thought Reasoning in Arithmetic Word Problems. 155-168 - Natalia Flechas Manrique, Wanqian Bao, Aurélie Herbelot

, Uri Hasson
:
Enhancing Interpretability Using Human Similarity Judgements to Prune Word Embeddings. 169-179 - Judith Sieker, Sina Zarrieß:

When Your Language Model Cannot Even Do Determiners Right: Probing for Anti-Presuppositions and the Maximize Presupposition! Principle. 180-198 - Jonas Groschwitz:

Introducing VULCAN: A Visualization Tool for Understanding Our Models and Data by Example. 199-211 - David Kletz, Pascal Amsili, Marie Candito:

The Self-Contained Negation Test Set. 212-221 - Yan Cong

, Emmanuele Chersoni, Yu-Yin Hsu, Philippe Blache:
Investigating the Effect of Discourse Connectives on Transformer Surprisal: Language Models Understand Connectives, Even So They Are Surprised. 222-232 - Yichu Zhou, Vivek Srikumar:

METAPROBE: A Representation- and Task-Agnostic Probe. 233-249 - Jacob K. Johnson, Ana Marasovic:

How Much Consistency Is Your Accuracy Worth? 250-260 - Tanja Baeumel, Soniya Vijayakumar, Josef van Genabith, Guenter Neumann, Simon Ostermann:

Investigating the Encoding of Words in BERT's Neurons Using Feature Textualization. 261-270 - Shunjie Wang, Shane Steinert-Threlkeld:

Evaluating Transformer's Ability to Learn Mildly Context-Sensitive Languages. 271-283 - Nirmalendu Prakash, Roy Ka-Wei Lee:

Layered Bias: Interpreting Bias in Pretrained Large Language Models. 284-295 - Isabelle Lorge, Janet B. Pierrehumbert:

Not Wacky vs. Definitely Wacky: A Study of Scalar Adverbs in Pretrained Language Models. 296-316 - Jing Huang

, Atticus Geiger, Karel D'Oosterlinck, Zhengxuan Wu, Christopher Potts:
Rigorously Assessing Natural Language Explanations of Neurons. 317-331 - Deanna DeCarlo, William Palmer, Michael Wilson, Bob Frank:

NPIs Aren't Exactly Easy: Variation in Licensing across Large Language Models. 332-341 - Mansi Sakarvadia, Aswathy Ajith, Arham Khan, Daniel Grzenda, Nathaniel Hudson, André Bauer

, Kyle Chard, Ian T. Foster:
Memory Injections: Correcting Multi-Hop Reasoning Failures During Inference in Transformer-Based Language Models. 342-356 - Aishik Chakraborty, Jackie C. K. Cheung, Timothy J. O'Donnell:

Systematic Generalization by Finetuning? Analyzing Pretrained Language Models Using Constituency Tests. 357-366 - Chenxin Liu, Emmanuele Chersoni:

On Quick Kisses and How to Make Them Count: A Study on Event Construal in Light Verb Constructions with BERT. 367-378 - Abhijith Chintam, Rahel Beloch, Willem H. Zuidema, Michael Hanna

, Oskar van der Wal:
Identifying and Adapting Transformer-Components Responsible for Gender Bias in an English Language Model. 379-394

manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.


Google
Google Scholar
Semantic Scholar
Internet Archive Scholar
CiteSeerX
ORCID














