Leshem Choshen

Profile

I'm an Open and Collaborative Natural Language Processing researcher MIT & IBM, currently working on making large language model research more efficient, collaborative and achievable by anyone. I work a lot on evaluation (check out Unitxt), co-created model merging, ZipNN for compressing models (no not quantization, compression :-)). My work focuses on democratizing AI through open science initiatives like BabyLM challenge which I co-organize to promote sample-efficient language model training. I am passionate about collaborative and accessible research. My recent projects include ComPEFT for compressing fine-tuned models, ShareLM for sharing human-model conversations with the community, and tinyBenchmarks for efficient model evaluation. I've also worked extensively on model merging techniques like TIES-Merging and ColD Fusion to enable model recycling.

I believe some technologies are more beneficial to the world than others and that science can be fun.
My research emphasizes making AI systems more accessible to broader communities to use, build, tweak and understand.

Publications

Sloth: scaling laws for LLM skills to predict multi-benchmark performance across families

Felipe Maia Polo, Seamus Somerstep, Leshem Choshen, Yuekai Sun, M. Yurochkin

arXiv.org 2024

Findings of the Second BabyLM Challenge: Sample-Efficient Pretraining on Developmentally Plausible Corpora

Findings of the Second BabyLM Challenge: Sample-Efficient Pretraining on Developmentally Plausible Corpora

Michael Y. Hu, Aaron Mueller, Candace Ross, Adina Williams, Tal Linzen, Chengxu Zhuang, Ryan Cotterell, Leshem Choshen, A. Warstadt, E. Wilcox

arXiv.org 2024

Global MMLU: Understanding and Addressing Cultural and Linguistic Biases in Multilingual Evaluation

Global MMLU: Understanding and Addressing Cultural and Linguistic Biases in Multilingual Evaluation

Shivalika Singh, Angelika Romanou, Clémentine Fourrier, D. Adelani, Jian Gang Ngui, Daniel Vila-Suero, Peerat Limkonchotiwat, Kelly Marchisio, Wei Qi Leong, Yosephine Susanto, Raymond Ng, Shayne Longpre, Wei-Yin Ko, Madeline Smith, Antoine Bosselut, Alice Oh, Andre F. T. Martins, Leshem Choshen, Daphne Ippolito, Enzo Ferrante, Marzieh Fadaee, B. Ermiş, Sara Hooker

arXiv.org 2024

Holmes ⌕ A Benchmark to Assess the Linguistic Competence of Language Models

Andreas Waldis, Yotam Perlitz, Leshem Choshen, Yufang Hou, Iryna Gurevych

Transactions of the Association for Computational Linguistics 2024

ZipNN: Lossless Compression for AI Models

ZipNN: Lossless Compression for AI Models

Moshik Hershcovitch, Andrew Wood, Leshem Choshen, Guy Girmonsky, Roy Leibovitz, Ilias Ennmouri, Michal Malka, Peter Chin, S. Sundararaman, Danny Harnik

arXiv.org 2024

Model merging with SVD to tie the Knots

Model merging with SVD to tie the Knots

George Stoica, Pratik Ramesh, B. Ecsedi, Leshem Choshen, Judy Hoffman

arXiv.org 2024

A Hitchhiker's Guide to Scaling Law Estimation

A Hitchhiker's Guide to Scaling Law Estimation

Leshem Choshen, Yang Zhang, Jacob Andreas

arXiv.org 2024

LiveXiv - A Multi-Modal Live Benchmark Based on Arxiv Papers Content

LiveXiv - A Multi-Modal Live Benchmark Based on Arxiv Papers Content

Nimrod Shabtay, Felipe Maia Polo, Sivan Doveh, Wei Lin, M. Mirza, Leshem Choshen, M. Yurochkin, Yuekai Sun, Assaf Arbelle, Leonid Karlinsky, Raja Giryes

arXiv.org 2024

Unforgettable Generalization in Language Models

Eric Zhang, Leshem Choshen, Jacob Andreas

arXiv.org 2024

Can You Trust Your Metric? Automatic Concatenation-Based Tests for Metric Validity

Can You Trust Your Metric? Automatic Concatenation-Based Tests for Metric Validity

Ora Nova Fandina, Leshem Choshen, E. Farchi, George Kour, Yotam Perlitz, Orna Raz

arXiv.org 2024

Beneath the Surface of Consistency: Exploring Cross-lingual Knowledge Representation Sharing in LLMs

Beneath the Surface of Consistency: Exploring Cross-lingual Knowledge Representation Sharing in LLMs

Maxim Ifergan, Leshem Choshen, Roee Aharoni, Idan Szpektor, Omri Abend

arXiv.org 2024

The ShareLM Collection and Plugin: Contributing Human-Model Chats for the Benefit of the Community

The ShareLM Collection and Plugin: Contributing Human-Model Chats for the Benefit of the Community

Shachar Don-Yehiya, Leshem Choshen, Omri Abend

arXiv.org 2024

The Future of Open Human Feedback

The Future of Open Human Feedback

Shachar Don-Yehiya, Ben Burtenshaw, Ramon Fernandez Astudillo, Cailean Osborne, Mimansa Jaiswal, Tzu-Sheng Kuo, Wenting Zhao, Idan Shenfeld, Andi Peng, Mikhail Yurochkin, Atoosa Kasirzadeh, Yangsibo Huang, Tatsunori Hashimoto, Yacine Jernite, Daniel Vila-Suero, Omri Abend, Jennifer Ding, Sara Hooker, Hannah Rose Kirk, Leshem Choshen

arXiv.org 2024

A Survey on Model MoErging: Recycling and Routing Among Specialized Experts for Collaborative Learning

A Survey on Model MoErging: Recycling and Routing Among Specialized Experts for Collaborative Learning

Prateek Yadav, Colin Raffel, Mohammed Muqeeth, Lucas Caccia, Haokun Liu, Tian-Xiang Chen, Mohit Bansal, Leshem Choshen, Alessandro Sordoni

arXiv.org 2024

Data Contamination Report from the 2024 CONDA Shared Task

Data Contamination Report from the 2024 CONDA Shared Task

Oscar Sainz, Iker Garc'ia-Ferrero, Alon Jacovi, Jon Ander Campos, Yanai Elazar, Eneko Agirre, Yoav Goldberg, Wei-Lin Chen, Jenny Chim, Leshem Choshen, Luca D'Amico-Wong, Melissa Dell, Run-Ze Fan, Shahriar Golchin, Yucheng Li, Pengfei Liu, Bhavish Pahwa, Ameya Prabhu, Suryansh Sharma, Emily Silcock, Kateryna Solonko, David Stap, M. Surdeanu, Yu-Min Tseng, Vishaal Udandarao, Zengzhi Wang, Ruijie Xu, Jinglin Yang

CONDA 2024

Do These LLM Benchmarks Agree? Fixing Benchmark Evaluation with BenchBench

Do These LLM Benchmarks Agree? Fixing Benchmark Evaluation with BenchBench

Yotam Perlitz, Ariel Gera, Ofir Arviv, Asaf Yehudai, Elron Bandel, Eyal Shnarch, Michal Shmueli-Scheuer, Leshem Choshen

Learning from Naturally Occurring Feedback

Learning from Naturally Occurring Feedback

Shachar Don-Yehiya, Leshem Choshen, Omri Abend

arXiv.org 2024

Compress then Serve: Serving Thousands of LoRA Adapters with Little Overhead

Compress then Serve: Serving Thousands of LoRA Adapters with Little Overhead

Rickard Brüel Gabrielsson, Jiacheng Zhu, Onkar Bhardwaj, Leshem Choshen, K. Greenewald, M. Yurochkin, Justin Solomon

arXiv.org 2024

Efficient multi-prompt evaluation of LLMs

Efficient multi-prompt evaluation of LLMs

Felipe Maia Polo, Ronald Xu, Lucas Weber, M'irian Silva, Onkar Bhardwaj, Leshem Choshen, Allysson Flavio Melo de Oliveira, Yuekai Sun, M. Yurochkin

arXiv.org 2024

Elements of World Knowledge (EWOK): A cognition-inspired framework for evaluating basic world knowledge in language models

Elements of World Knowledge (EWOK): A cognition-inspired framework for evaluating basic world knowledge in language models

Anna A. Ivanova, Aalok Sathe, Benjamin Lipkin, Unnathi Kumar, S. Radkani, T. H. Clark, Carina Kauf, Jennifer Hu, R. T. Pramod, Gabriel Grand, Vivian C. Paulun, Maria Ryskina, Ekin Akyürek, E. Wilcox, Nafisa Rashid, Leshem Choshen, Roger Levy, Evelina Fedorenko, Josh Tenenbaum, Jacob Andreas

arXiv.org 2024

Holmes: A Benchmark to Assess the Linguistic Competence of Language Models

Holmes: A Benchmark to Assess the Linguistic Competence of Language Models

Andreas Waldis, Yotam Perlitz, Leshem Choshen, Yufang Hou, Iryna Gurevych

[Call for Papers] The 2nd BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus

[Call for Papers] The 2nd BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus

Leshem Choshen, Ryan Cotterell, Michael Y. Hu, Tal Linzen, Aaron Mueller, Candace Ross, Alex Warstadt, E. Wilcox, Adina Williams, Chengxu Zhuang

arXiv.org 2024

Lossless and Near-Lossless Compression for Foundation Models

Lossless and Near-Lossless Compression for Foundation Models

Moshik Hershcovitch, Leshem Choshen, Andrew Wood, Ilias Enmouri, Peter Chin, S. Sundararaman, Danny Harnik

arXiv.org 2024

NumeroLogic: Number Encoding for Enhanced LLMs’ Numerical Reasoning

NumeroLogic: Number Encoding for Enhanced LLMs’ Numerical Reasoning

Eli Schwartz, Leshem Choshen, J. Shtok, Sivan Doveh, Leonid Karlinsky, Assaf Arbelle

Conference on Empirical Methods in Natural Language Processing 2024

Asymmetry in Low-Rank Adapters of Foundation Models

Asymmetry in Low-Rank Adapters of Foundation Models

Jiacheng Zhu, K. Greenewald, Kimia Nadjahi, Haitz S'aez de Oc'ariz Borde, Rickard Brüel Gabrielsson, Leshem Choshen, Marzyeh Ghassemi, M. Yurochkin, Justin Solomon

International Conference on Machine Learning 2024

tinyBenchmarks: evaluating LLMs with fewer examples

tinyBenchmarks: evaluating LLMs with fewer examples

Felipe Maia Polo, Lucas Weber, Leshem Choshen, Yuekai Sun, Gongjun Xu, M. Yurochkin

International Conference on Machine Learning 2024

Label-Efficient Model Selection for Text Generation

Label-Efficient Model Selection for Text Generation

Shir Ashury-Tahan, B. Sznajder, Leshem Choshen, L. Ein-Dor, Eyal Shnarch, Ariel Gera

Annual Meeting of the Association for Computational Linguistics 2024

Unitxt: Flexible, Shareable and Reusable Data Preparation and Evaluation for Generative AI

Unitxt: Flexible, Shareable and Reusable Data Preparation and Evaluation for Generative AI

Elron Bandel, Yotam Perlitz, Elad Venezian, Roni Friedman-Melamed, Ofir Arviv, Matan Orbach, Shachar Don-Yehiya, D. Sheinwald, Ariel Gera, Leshem Choshen, Michal Shmueli-Scheuer, Yoav Katz

North American Chapter of the Association for Computational Linguistics 2024

Genie: Achieving Human Parity in Content-Grounded Datasets Generation

Genie: Achieving Human Parity in Content-Grounded Datasets Generation

Asaf Yehudai, Boaz Carmeli, Y. Mass, Ofir Arviv, Nathaniel Mills, Assaf Toledo, Eyal Shnarch, Leshem Choshen

arXiv.org 2024

Deductive Closure Training of Language Models for Coherence, Accuracy, and Updatability

Deductive Closure Training of Language Models for Coherence, Accuracy, and Updatability

Afra Feyza Akyürek, Ekin Akyürek, Leshem Choshen, Derry Tanti Wijaya, Jacob Andreas

Annual Meeting of the Association for Computational Linguistics 2024

ComPEFT: Compression for Communicating Parameter Efficient Updates via Sparsification and Quantization

ComPEFT: Compression for Communicating Parameter Efficient Updates via Sparsification and Quantization

Prateek Yadav, Leshem Choshen, Colin Raffel, Mohit Bansal

arXiv.org 2023

Human Learning by Model Feedback: The Dynamics of Iterative Prompting with Midjourney

Human Learning by Model Feedback: The Dynamics of Iterative Prompting with Midjourney

Shachar Don-Yehiya, Leshem Choshen, Omri Abend

Conference on Empirical Methods in Natural Language Processing 2023

Fuse to Forget: Bias Reduction and Selective Memorization through Model Fusion

Fuse to Forget: Bias Reduction and Selective Memorization through Model Fusion

Kerem Zaman, Leshem Choshen, Shashank Srivastava

Conference on Empirical Methods in Natural Language Processing 2023

Efficient Benchmarking (of Language Models)

Efficient Benchmarking (of Language Models)

Yotam Perlitz, Elron Bandel, Ariel Gera, Ofir Arviv, L. Ein-Dor, Eyal Shnarch, N. Slonim, Michal Shmueli-Scheuer, Leshem Choshen

North American Chapter of the Association for Computational Linguistics 2023

TIES-Merging: Resolving Interference When Merging Models

TIES-Merging: Resolving Interference When Merging Models

Prateek Yadav, Derek Tam, Leshem Choshen, Colin Raffel, Mohit Bansal

Neural Information Processing Systems 2023

MuLER: Detailed and Scalable Reference-based Evaluation

MuLER: Detailed and Scalable Reference-based Evaluation

Taelin Karidi, Leshem Choshen, Gal Patel, Omri Abend

Conference on Computational Natural Language Learning 2023

Jump to Conclusions: Short-Cutting Transformers with Linear Transformations

Jump to Conclusions: Short-Cutting Transformers with Linear Transformations

Alexander Yom Din, Taelin Karidi, Leshem Choshen, Mor Geva

International Conference on Language Resources and Evaluation 2023

Knowledge is a Region in Weight Space for Fine-tuned Language Models

Knowledge is a Region in Weight Space for Fine-tuned Language Models

Almog Gueta, Elad Venezian, Colin Raffel, N. Slonim, Yoav Katz, Leshem Choshen

Conference on Empirical Methods in Natural Language Processing 2023

Call for Papers - The BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus

Call for Papers - The BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus

Alex Warstadt, Leshem Choshen, Aaron Mueller, Adina Williams, Ethan Gotlieb Wilcox, Chengxu Zhuang

arXiv.org 2023

ColD Fusion: Collaborative Descent for Distributed Multitask Finetuning

ColD Fusion: Collaborative Descent for Distributed Multitask Finetuning

Shachar Don-Yehiya, Elad Venezian, Colin Raffel, N. Slonim, Yoav Katz, Leshem Choshen

Annual Meeting of the Association for Computational Linguistics 2022

DisentQA: Disentangling Parametric and Contextual Knowledge with Counterfactual Question Answering

DisentQA: Disentangling Parametric and Contextual Knowledge with Counterfactual Question Answering

Ella Neeman, Roee Aharoni, Or Honovich, Leshem Choshen, Idan Szpektor, Omri Abend

Annual Meeting of the Association for Computational Linguistics 2022

Where to start? Analyzing the potential value of intermediate models

Where to start? Analyzing the potential value of intermediate models

Leshem Choshen, Elad Venezian, Shachar Don-Yehiya, N. Slonim, Yoav Katz

Conference on Empirical Methods in Natural Language Processing 2022

Reinforcement Learning with Large Action Spaces for Neural Machine Translation

Reinforcement Learning with Large Action Spaces for Neural Machine Translation

Asaf Yehudai, Leshem Choshen, Lior Fox, Omri Abend

International Conference on Computational Linguistics 2022

Label Sleuth: From Unlabeled Text to a Classifier in a Few Hours

Eyal Shnarch, Alon Halfon, Ariel Gera, Marina Danilevsky, Yannis Katsis, Leshem Choshen, M. Cooper, Dina Epelboim, Zheng Zhang, Dakuo Wang, Lucy Yip, L. Ein-Dor, Lena Dankin, Ilya Shnayderman, R. Aharonov, Yunyao Li, Naftali Liberman, Philip Levin Slesarev, Gwilym Newton, Shila Ofek-Koifman, N. Slonim, Yoav Katz

Conference on Empirical Methods in Natural Language Processing 2022

PreQuEL: Quality Estimation of Machine Translation Outputs in Advance

PreQuEL: Quality Estimation of Machine Translation Outputs in Advance

Shachar Don-Yehiya, Leshem Choshen, Omri Abend

Conference on Empirical Methods in Natural Language Processing 2022

Some Grammatical Errors are Frequent, Others are Important

Some Grammatical Errors are Frequent, Others are Important

Leshem Choshen, Ofir Shifman, Omri Abend

arXiv.org 2022

Fusing finetuned models for better pretraining

Fusing finetuned models for better pretraining

Leshem Choshen, Elad Venezian, N. Slonim, Yoav Katz

arXiv.org 2022

Cluster & Tune: Boost Cold Start Performance in Text Classification

Cluster & Tune: Boost Cold Start Performance in Text Classification

Eyal Shnarch, Ariel Gera, Alon Halfon, Lena Dankin, Leshem Choshen, R. Aharonov, N. Slonim

Annual Meeting of the Association for Computational Linguistics 2022

Semantics-aware Attention Improves Neural Machine Translation

Semantics-aware Attention Improves Neural Machine Translation

Aviv Slobodkin, Leshem Choshen, Omri Abend

STARSEM 2021

On Neurons Invariant to Sentence Structural Changes in Neural Machine Translation

On Neurons Invariant to Sentence Structural Changes in Neural Machine Translation

Gal Patel, Leshem Choshen, Omri Abend

Conference on Computational Natural Language Learning 2021

The Grammar-Learning Trajectories of Neural Language Models

The Grammar-Learning Trajectories of Neural Language Models

Leshem Choshen, Guy Hacohen, D. Weinshall, Omri Abend

Annual Meeting of the Association for Computational Linguistics 2021

ComSum: Commit Messages Summarization and Meaning Preservation

ComSum: Commit Messages Summarization and Meaning Preservation

Leshem Choshen, Idan Amit

arXiv.org 2021

Part of Speech and Universal Dependency effects on English Arabic Machine Translation

Part of Speech and Universal Dependency effects on English Arabic Machine Translation

Ofek Rafaeli, Omri Abend, Leshem Choshen, D. Nikolaev

arXiv.org 2021

Q^{2}: Evaluating Factual Consistency in Knowledge-Grounded Dialogues via Question Generation and Question Answering

Q^{2}: Evaluating Factual Consistency in Knowledge-Grounded Dialogues via Question Generation and Question Answering

Or Honovich, Leshem Choshen, Roee Aharoni, Ella Neeman, Idan Szpektor, Omri Abend

Conference on Empirical Methods in Natural Language Processing 2021

Mediators in Determining what Processing BERT Performs First

Aviv Slobodkin, Leshem Choshen, Omri Abend

North American Chapter of the Association for Computational Linguistics 2021

GrASP: A Library for Extracting and Exploring Human-Interpretable Textual Patterns

GrASP: A Library for Extracting and Exploring Human-Interpretable Textual Patterns

Piyawat Lertvittayakumjorn, Leshem Choshen, Eyal Shnarch, Francesca Toni

International Conference on Language Resources and Evaluation 2021

SERRANT: a syntactic classifier for English Grammatical Error Types

Leshem Choshen, Matanel Orenm Dmitry Nikolaev, Omri Abend

arXiv.org 2021

An autonomous debating system

N. Slonim, Yonatan Bilu, Carlos Alzate, Roy Bar-Haim, Ben Bogin, Francesca Bonin, Leshem Choshen, Edo Cohen-Karlik, Lena Dankin, Lilach Edelstein, L. Ein-Dor, Roni Friedman-Melamed, A. Gavron, Ariel Gera, Martin Gleize, Shai Gretz, Dan Gutfreund, Alon Halfon, Daniel Hershcovich, R. Hoory, Yufang Hou, S. Hummel, Michal Jacovi, Charles Jochim, Yoav Kantor, Yoav Katz, D. Konopnicki, Zvi Kons, Lili Kotlerman, Dalia Krieger, Dan Lahav, Tamar Lavee, Ran Levy, Naftali Liberman, Y. Mass, Amir Menczel, Shachar Mirkin, Guy Moshkowich, Shila Ofek-Koifman, Matan Orbach, Ella Rabinovich, Ruty Rinott, Slava Shechtman, D. Sheinwald, Eyal Shnarch, Ilya Shnayderman, A. Soffer, Artem Spector, B. Sznajder, Assaf Toledo, Orith Toledo-Ronen, Elad Venezian, R. Aharonov

Nature 2021

Transition based Graph Decoder for Neural Machine Translation

Transition based Graph Decoder for Neural Machine Translation

Leshem Choshen, Omri Abend

arXiv.org 2021

Enhancing the Transformer Decoder with Transition-based Syntax

Enhancing the Transformer Decoder with Transition-based Syntax

Leshem Choshen, Omri Abend

Conference on Computational Natural Language Learning 2021

Active Learning for BERT: An Empirical Study

Active Learning for BERT: An Empirical Study

L. Ein-Dor, Alon Halfon, Ariel Gera, Eyal Shnarch, Lena Dankin, Leshem Choshen, Marina Danilevsky, R. Aharonov, Yoav Katz, N. Slonim

Conference on Empirical Methods in Natural Language Processing 2020

Classifying Syntactic Errors in Learner Language

Classifying Syntactic Errors in Learner Language

Leshem Choshen, D. Nikolaev, Yevgeni Berzak, Omri Abend

Conference on Computational Natural Language Learning 2020

Unsupervised Expressive Rules Provide Explainability and Assist Human Experts Grasping New Domains

Unsupervised Expressive Rules Provide Explainability and Assist Human Experts Grasping New Domains

Eyal Shnarch, Leshem Choshen, Guy Moshkowich, N. Slonim, R. Aharonov

Findings 2020

Corpus Wide Argument Mining - a Working Solution

Corpus Wide Argument Mining - a Working Solution

L. Ein-Dor, Eyal Shnarch, Lena Dankin, Alon Halfon, B. Sznajder, Ariel Gera, Carlos Alzate, Martin Gleize, Leshem Choshen, Yufang Hou, Yonatan Bilu, R. Aharonov, N. Slonim

AAAI Conference on Artificial Intelligence 2019

All Neural Networks are Created Equal

All Neural Networks are Created Equal

Guy Hacohen, Leshem Choshen, D. Weinshall

arXiv.org 2019

Automatically Extracting Challenge Sets for Non-Local Phenomena in Neural Machine Translation

Automatically Extracting Challenge Sets for Non-Local Phenomena in Neural Machine Translation

Leshem Choshen, Omri Abend

Conference on Computational Natural Language Learning 2019

On the Weaknesses of Reinforcement Learning for Neural Machine Translation

On the Weaknesses of Reinforcement Learning for Neural Machine Translation

Leshem Choshen, Lior Fox, Zohar Aizenbud, Omri Abend

International Conference on Learning Representations 2019

Are You Convinced? Choosing the More Convincing Evidence with a Siamese Network

Are You Convinced? Choosing the More Convincing Evidence with a Siamese Network

Martin Gleize, Eyal Shnarch, Leshem Choshen, Lena Dankin, Guy Moshkowich, R. Aharonov, N. Slonim

Annual Meeting of the Association for Computational Linguistics 2019

Learning to combine Grammatical Error Corrections

Learning to combine Grammatical Error Corrections

Yoav Kantor, Yoav Katz, Leshem Choshen, Edo Cohen-Karlik, Naftali Liberman, Assaf Toledo, Amir Menczel, N. Slonim

BEA@ACL 2019

Let's Agree to Agree: Neural Networks Share Classification Order on Real Datasets

Let's Agree to Agree: Neural Networks Share Classification Order on Real Datasets

Guy Hacohen, Leshem Choshen, D. Weinshall

International Conference on Machine Learning 2019

The Language of Legal and Illegal Activity on the Darknet

The Language of Legal and Illegal Activity on the Darknet

Leshem Choshen, D. Eldad, Daniel Hershcovich, Elior Sulem, Omri Abend

Annual Meeting of the Association for Computational Linguistics 2019

SemEval-2019 Task 1: Cross-lingual Semantic Parsing with UCCA

SemEval-2019 Task 1: Cross-lingual Semantic Parsing with UCCA

Daniel Hershcovich, Zohar Aizenbud, Leshem Choshen, Elior Sulem, A. Rappoport, Omri Abend

International Workshop on Semantic Evaluation 2019

Will it Blend? Blending Weak and Strong Labeled Data in a Neural Network for Argumentation Mining

Will it Blend? Blending Weak and Strong Labeled Data in a Neural Network for Argumentation Mining

Eyal Shnarch, Carlos Alzate, Lena Dankin, Martin Gleize, Yufang Hou, Leshem Choshen, R. Aharonov, N. Slonim

Annual Meeting of the Association for Computational Linguistics 2018

SemEval 2019 Shared Task: Cross-lingual Semantic Parsing with UCCA - Call for Participation

SemEval 2019 Shared Task: Cross-lingual Semantic Parsing with UCCA - Call for Participation

Daniel Hershcovich, Leshem Choshen, Elior Sulem, Zohar Aizenbud, A. Rappoport, Omri Abend

arXiv.org 2018

Inherent Biases in Reference-based Evaluation for Grammatical Error Correction

Inherent Biases in Reference-based Evaluation for Grammatical Error Correction

Leshem Choshen, Omri Abend

Annual Meeting of the Association for Computational Linguistics 2018

Automatic Metric Validation for Grammatical Error Correction

Automatic Metric Validation for Grammatical Error Correction

Leshem Choshen, Omri Abend

Annual Meeting of the Association for Computational Linguistics 2018

Reference-less Measure of Faithfulness for Grammatical Error Correction

Reference-less Measure of Faithfulness for Grammatical Error Correction

Leshem Choshen, Omri Abend

North American Chapter of the Association for Computational Linguistics 2018

DORA The Explorer: Directed Outreaching Reinforcement Action-Selection

DORA The Explorer: Directed Outreaching Reinforcement Action-Selection

Leshem Choshen, Lior Fox, Y. Loewenstein

International Conference on Learning Representations 2018

Findings of the BabyLM Challenge: Sample-Efficient Pretraining on Developmentally Plausible Corpora

Alex Warstadt, Aaron Mueller, Leshem Choshen, E. Wilcox, Chengxu Zhuang, Juan Ciro, Rafael Mosquera, Bhargavi Paranjabe, Adina Williams, Tal Linzen, Ryan Cotterell

Proceedings of the BabyLM Challenge at the 27th Conference on Computational Natural Language Learning 2023

On the Weaknesses or Reinforcement Learning for Neural Machine Translation

Leshem Choshen, Lior Fox, Zohar Aizenbud, Omri Abend

Achieving Human Parity in Content-Grounded Datasets Generation

Asaf Yehudai, Boaz Carmeli, Y. Mass, Ofir Arviv, Nathaniel Mills, Eyal Shnarch, Leshem Choshen

International Conference on Learning Representations 2024

Navigating the Modern Evaluation Landscape: Considerations in Benchmarks and Frameworks for Large Language Models (LLMs)

Leshem Choshen, Ariel Gera, Yotam Perlitz, Michal Shmueli-Scheuer, Gabriel Stanovsky

International Conference on Language Resources and Evaluation 2024