Natural Language Processing (NLP) is an established and dynamic field. Despite this, what constitutes NLP research remains debated. In this work, we address the question by quantitatively examining NLP research papers. We propose a taxonomy of research contributions and introduce NLPContributions, a dataset of nearly 2k NLP research paper abstracts, carefully annotated to identify scientific contributions and classify their types according to this taxonomy. We also introduce a novel task of automatically identifying contribution statements and classifying their types from research papers. We present experimental results for this task and apply our model to ∼29k NLP research papers to analyze their contributions, aiding in the understanding of the nature of NLP research. We show that NLP research has taken a winding path — with the focus on language and human-centric studies being prominent in the 1970s and 80s, tapering off in the 1990s and 2000s, and starting to rise again since the late 2010s. Alongside this revival, we observe a steady rise in dataset and methodological contributions since the 1990s, such that today, on average, individual NLP papers contribute in more ways than ever before. Our dataset and analyses offer a powerful lens for tracing research trends and focus on language and human-centric studies being prominent in the 1970s and 80s, tapering off in the 1990s and 2000s, and starting to rise again since the late 2010s. Alongside this revival, we observe a steady rise in dataset and methodological contributions since the 1990s, such that today, on average, individual NLP papers contribute in more ways than ever before. Our dataset and analyses offer a powerful lens for tracing research trends and offer potential for generating informed, data-driven literature surveys.
Type | Sub-type | Description | Example |
---|---|---|---|
Knowledge | k-dataset | Describes new knowledge about datasets, such as their new properties or characteristics. | "Furthermore, our thorough analysis demonstrates the average distance between aspect and opinion words are shortened by at least 19% on the standard SemEval Restaurant14 dataset." - Zhou et al., 2021 |
k-language | Presents new knowledge about language, such as a new property or characteristic of language. | "In modern Chinese articles or conversations, it is very popular to involve a few English words, especially in emails and Internet literature." - Zhao et al., 2012 | |
k-method | Describes new knowledge or analysis about NLP models or methods (which predominantly draw from Machine Learning). | “Different generative processes identify specific failure modes of the underlying model.” – Deng et al. (2022) | |
k-people | Presents new knowledge about people, humankind, society, or human civilization. | “Combating the outcomes of this infodemic is not only a question of identifying false claims, but also reasoning about the decisions individuals make.” – Pacheco et al. (2022) | |
k-task | Describes new knowledge about NLP tasks. | “We show that these bilingual features outperform the monolingual features used in prior work for the task of classifying translation direction.” – Eetemadi and Toutanova (2014) | |
Artifact | a-dataset | Introduces a new NLP dataset (i.e., textual resources such as corpora or lexicon). | “We present a new corpus of Weibo messages annotated for both name and nominal mentions.” – Peng and Dredze (2015) |
a-method | Introduces or proposes a new or novel NLP method or model (primarily to solve NLP task(s)) | “The paper also describes a novel method, EXEMPLAR, which adapts ideas from SRL to less costly NLP machinery, resulting in substantial gains both in efficiency and effectiveness, over binary and n-ary relation extraction tasks.” – Mesquita et al. (2013) | |
a-task | Introduces or proposes a new or novel NLP task (i.e., well-defined NLP problem). | “We formulate a task that represents a hybrid of slot-filling information extraction and named entity recognition and annotate data from four different forums.” – Durrett et al. (2017) | |
Fine-tuned SciBERT (macro-F1 ≈ 0.80) → tag 28,937 ACL papers.
Now we can ask field-level questions without months of manual coding.
“Individual NLP papers contribute in more ways than ever before.”
— Let’s keep that diversity alive, measure it, and learn from it.
@inproceedings{pramanick-etal-2025-nlpcontributions,
title={The Nature of NLP: Analyzing Contributions in NLP Papers},
author={Pramanick, Aniket and Hou, Yufang and Mohammad, Saif and Gurevych, Iryna},
booktitle={The 63rd Annual Meeting of the Association for Computational Linguistics},
year={2025},
url={https://arxiv.org/abs/2409.19505}
}