Translating Natural Language to Planning Goals with Large-Language Models [PDF]
Recent large language models (LLMs) have demonstrated remarkable performance on a variety of natural language processing (NLP) tasks, leading to intense excitement about their applicability across various domains.
Yaqi Xie+5 more
semanticscholar +1 more source
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension [PDF]
We present BART, a denoising autoencoder for pretraining sequence-to-sequence models. BART is trained by (1) corrupting text with an arbitrary noising function, and (2) learning a model to reconstruct the original text.
M. Lewis+7 more
semanticscholar +1 more source
Making Sense of Language Signals for Monitoring Radicalization
Understanding radicalization pathways, drivers, and factors is essential for the effective design of prevention and counter-radicalization programs. Traditionally, the primary methods used by social scientists to detect these drivers and factors include ...
Óscar Araque+7 more
doaj +1 more source
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding [PDF]
Human ability to understand language is general, flexible, and robust. In contrast, most NLU models above the word level are designed for a specific task and struggle with out-of-domain data.
Alex Wang+5 more
semanticscholar +1 more source
CLAP Learning Audio Concepts from Natural Language Supervision
Mainstream machine listening models are trained to learn audio concepts under the paradigm of one class label to many recordings focusing on one task.
Benjamin Elizalde+3 more
semanticscholar +1 more source
Domain-Specific Language Model Pretraining for Biomedical Natural Language Processing [PDF]
Pretraining large neural language models, such as BERT, has led to impressive gains on many natural language processing (NLP) tasks. However, most pretraining efforts focus on general domain corpora, such as newswire and Web.
Yu Gu+8 more
semanticscholar +1 more source
A large annotated corpus for learning natural language inference [PDF]
Understanding entailment and contradiction is fundamental to understanding natural language, and inference about entailment and contradiction is a valuable testing ground for the development of semantic representations. However, machine learning research
Samuel R. Bowman+3 more
semanticscholar +1 more source
Stanza: A Python Natural Language Processing Toolkit for Many Human Languages [PDF]
We introduce Stanza, an open-source Python natural language processing toolkit supporting 66 human languages. Compared to existing widely used toolkits, Stanza features a language-agnostic fully neural pipeline for text analysis, including tokenization ...
Peng Qi+4 more
semanticscholar +1 more source
Translation between Molecules and Natural Language [PDF]
We present MolT5 - a self-supervised learning framework for pretraining models on a vast amount of unlabeled natural language text and molecule strings.
Carl N. Edwards+4 more
semanticscholar +1 more source
Natural Language Reasoning, A Survey [PDF]
This survey article proposes a clearer view of Natural Language Reasoning (NLR) in the field of Natural Language Processing (NLP), both conceptually and practically.
Fei Yu, Hongbo Zhang, Benyou Wang
semanticscholar +1 more source