| Title |
Authors |
Topic |
| Can Language Models Be Tricked by Language Illusions? Easier with Syntax, Harder with Semantics |
Yuhan Zhang, Edward Gibson and Forrest Davis |
Computational Psycholinguistics, Cognition and Linguistics |
| ToMChallenges: A Principle-Guided Dataset and Diverse Evaluation Tasks for Exploring Theory of Mind |
Xiaomeng Ma, Lingyu Gao and Qihui Xu |
Computational Psycholinguistics, Cognition and Linguistics |
| The Zipfian Challenge: Learning the statistical fingerprint of natural languages |
Christian Bentz |
Computational Psycholinguistics, Cognition and Linguistics |
| On the Effects of Structural Modeling for Neural Semantic Parsing |
Xiang Zhang, Shizhu He, Kang Liu and Jun Zhao |
Lexical, Compositional and Discourse Semantics |
| Title |
Authors |
Topic |
| The Validity of Evaluation Results: Assessing Concurrence Across Compositionality Benchmarks |
Kaiser Sun, Adina Williams and Dieuwke Hupkes |
Theoretical Analysis and Interpretation of ML Models for NLP |
| Mind the instructions: a holistic evaluation of consistency and interactions in prompt-based learning |
Lucas Weber, Elia Bruni and Dieuwke Hupkes |
Theoretical Analysis and Interpretation of ML Models for NLP |
| Med-HALT: Medical Domain Hallucination Test for Large Language Models |
Ankit pal, Logesh Kumar Umapathi and Malaikannan Sankarasubbu |
Resources and Tools for Scientifically Motivated Research |
| Revising with a Backward Glance: Regressions and Skips during Reading as Cognitive Signals for Revision Policies in Incremental Processing |
Brielen Madureira, Pelin Çelikkol and David Schlangen |
Theoretical Analysis and Interpretation of ML Models for NLP |
| Title |
Authors |
Topic |
| ChiSCor: A Corpus of Freely-Told Fantasy Stories by Dutch Children for Computational Linguistics and Cognitive Science |
Bram van Dijk, Max van Duijn, Suzan Verberne and Marco Spruit |
Resources and Tools for Scientifically Motivated Research |
| HNC: Leveraging Hard Negative Captions towards Models with Fine-Grained Visual-Linguistic Comprehension Capabilities |
Esra Dönmez, Pascal Tilli, Hsiu-Yu Yang, Ngoc Thang Vu and Carina Silberer |
Interaction and Grounded Language Learning |
| Theory of Mind in Large Language Models: Examining Performance of 11 State-of-the-Art models vs. Children Aged 7-10 on Advanced Tests |
Max van Duijn, Bram van Dijk, Tom Kouwenhoven, Werner de Valk, Marco Spruit and Peter vander Putten |
Computational Psycholinguistics, Cognition and Linguistics |
| A Block Metropolis-Hastings Sampler for Controllable Energy-based Text Generation |
Jarad Forristal, Fatemehsadat Mireshghallah, Greg Durrett and Taylor Berg-Kirkpatrick |
Natural Language Generation |
| Title |
Authors |
| Not all layers are equally as important: Every Layer Counts BERT |
Lucas Georges Gabriel Charpentier and David Samuel |
| Towards more Human-like Language Models based on Contextualizer Pretraining Strategy |
Chenghao Xiao, G Thomas Hudson and Noura Al Moubayed |
| Large GPT-like Models are Bad Babies: A Closer Look at the Relationship between Linguistic Competence and Psycholinguistic Measures |
Julius Steuer, Marius Mosbach and Dietrich Klakow |
| CLIMB – Curriculum Learning for Infant-inspired Model Building |
Richard Diehl Martinez, Hope McGovern, Zebulon Goriely, Christopher Davis, Andrew Caines, Paula Buttery and Lisa Beinborn |
| Title |
Authors |
| Humans and language models diverge when predicting repeating text |
Aditya Vaidya, Javier Turek and Alexander Huth |
| Investigating the Nature of Disagreements on Mid-Scale Ratings: A Case Study on the Abstractness-Concreteness Continuum |
Urban Knuples, Diego Frassinelli and Sabine Schulte im Walde |
| ArchBERT: Bi-Modal Understanding of Neural Architectures and Natural Languages |
Mohammad Akbari, Saeed Ranjbar Alvar, Behnam Kamranian, Amin Banitalebi-Dehkordi and Yong Zhang |
| A Comparative Study on Textual Saliency of Styles from Eye Tracking, Annotations, and Language Models |
Karin de Langis and Dongyeop Kang |
| PROPRES: Investigating the Projectivity of Presupposition with Various Triggers and Environments |
Daiki Asami and Saku Sugawara |
| A Minimal Approach for Natural Language Action Space in Text-based Games |
Dongwon Ryu, Meng Fang, Gholamreza Haffari, Shirui Pan and Ehsan Shareghi |
| Structural Ambiguity and its Disambiguation in Language Model Based Parsers: the Case of Dutch Clause Relativization |
Gijs Wijnholds and Michael Moortgat |
| Quirk or Palmer: A Comparative Study of Modal Verb Frameworks with Annotated Datasets |
Risako Owan, Maria Gini and Dongyeop Kang |
| Quantifying Information of Tokens for Simple and Flexible Simultaneous Machine Translation |
DongHyun Lee, Minkyung Park and Byung-Jun Lee |
| Enhancing Code-mixed Text Generation Using Synthetic Data Filtering in Neural Machine Translation |
Dama Sravani and Radhika Mamidi |
| Towards Better Evaluation of Instruction-Following: A Case-Study in Summarization |
Ondrej Skopek, Rahul Aralikatte, Sian Gooding and Victor Carbune |
| Syntactic Inductive Bias in Transformer Language Models: Especially Helpful for Low-Resource Languages? |
Luke Gessler and Nathan Schneider |
| Attribution and Alignment: Effects of Local Context Repetition on Utterance Production and Comprehension in Dialogue |
Aron Molnar, Jaap Jumelet, Mario Giulianelli and Arabella Sinclair |
| On the utility of enhancing BERT syntactic bias with Token Reordering Pretraining |
Yassir El Mesbahi, Atif Mahmud, Abbas Ghaddar, Mehdi Rezagholizadeh, Phillippe Langlais and Prasanna Parthasarathi |
| Baby Llama: knowledge distillation from an ensemble of teachers trained on a small dataset with no performance penalty |
Inar Timiryasov and Jean-Loup Tastet |
| BabyLM Challenge: Curriculum learning based on sentence complexity approximating language acquisition |
Miyu Oba, Akari Haga, Akiyo Fukatsu andYohei Oseki |
| Can training neural language models on a curriculum with developmentally plausible data improve alignment with human reading behavior?
|
Aryaman Chobey, Oliver Smith, Anzi Wang and Grusha Prasad |
| CogMemLM: Human-Like Memory Mechanisms Improve Performance and Cognitive Plausibility of LLMs |
Lukas Thoma, Ivonne Weyers, Erion Çano, Stefan Schweter, Jutta L Mueller and Benjamin Roth |
| McGill BabyLM Shared Task Submission: The Effects of Data Formatting and Structural Biases |
Ziling Cheng, Rahul Aralikatte, Ian Porada, Cesare Spinoso-Di Piano and Jackie CK Cheung |
| On the effect of curriculum learning with developmental data for grammar acquisition |
Mattia Opper, J Morrison and Siddharth N |
| ToddlerBERTa: Exploiting BabyBERTa for Grammar Learning and Language Understanding |
Ömer Veysel Çağatan |
| Title |
Authors |
| How Fragile is Relation Extraction under Entity Replacements? |
Yiwei Wang, Bryan Hooi, Fei Wang, Yujun Cai, Yuxuan Liang, Wenxuan Zhou, Jing Tang, Manjuan Duan and Muhao Chen |
| JaSPICE: Automatic Evaluation Metric Using Predicate-Argument Structures for Image Captioning Models |
Yuiga Wada, Kanta Kaneda and Komei Sugiura |
| MuLER: Detailed and Scalable Reference-based Evaluation |
Taelin Karidi, Leshem Choshen, Gal Patel and Omri Abend |
| The Impact of Familiarity on Naming Variation: A Study on Object Naming in Mandarin Chinese |
Yunke He, Xixian Liao, Jialing Liang and Gemma Boleda |
| PSST! Prosodic Speech Segmentation with Transformers |
Nathan Roll, Calbert Graham and Simon Todd |
| Alignment via Mutual Information |
Shinjini Ghosh, Yoon Kim, Ramon Fernandez Astudillo, Tahira Naseem and Jacob Andreas |
| Challenging the "One Single Vector per Token" Assumption |
Mathieu Dehouck |
| Strategies to Improve Low-Resource Agglutinative Languages Morphological Inflection |
Gulinigeer Abudouwaili, Wayit Ablez, Kahaerjiang Abiderexiti, Aishan Wumaier and Nian Yi |
| Exploring Transformers as Compact, Data-efficient Language Models |
Clayton Fields and Casey Kennington |
| Tree-shape Uncertainty for Analyzing the Inherent Branching Bias of Unsupervised Parsing Models |
Taiga Ishii and Yusuke Miyao |
| Future Lens: Anticipating Subsequent Tokens from a Single Hidden State |
Koyena Pal, Jiuding Sun, Andrew Yuan, Byron Wallace and David Bau |
| Cross-Document Event Coreference Resolution: Instruct Humans or Instruct GPT? |
Jin Zhao, Nianwen Xue and Bonan Min |
| Implications of Annotation Artifacts in Edge Probing Test Datasets |
Sagnik Ray Choudhury and Jushaan Kalra |
| REFER: An End-to-end Rationale Extraction Framework for Explanation Regularization |
MohammadReza GhasemiMadani and Pasquale Minervini |
| A surprisal oracle for active curriculum language modeling |
Xudong Hong, Sharid Loáiciga and Asad B. Sayeed |
| Baby’s CoThought: Leveraging Large Language Models for Enhanced Reasoning in Compact Models |
Zheyu Zhang, Han Yang, Bolei Ma, David Rügamer and Ercong Nie |
| Byte-ranked Curriculum Learning for BabyLM Strict-small Shared Task 2023 |
Justin DeBenedetto |
| ChapGTP, ILLC’s Attempt at Raising a BabyLM: Improving Data Efficiency by Automatic Task Formation |
Jaap Jumelet, Michael Hanna, Marianne De Heer Kloots, Anna Langedijk, Charlotte Pouw and Oskar van der Wal |
| GPT-wee: How Small Can a Small Language Model Really Get? |
Bastian Bunzeck and Sina Zarrieß |
| Mean BERTs make erratic language teachers: the effectiveness of latent bootstrapping in low-resource settings |
David Samuel |
| Tiny Language Models Enriched with Multimodal Knowledge from Multiplex Networks |
Clayton Fields, Osama Natouf, Andrew McMains, Catherine Henry and Casey Kennington |