Titles
stringlengths 6
220
| Abstracts
stringlengths 37
3.26k
| Years
int64 1.99k
2.02k
| Categories
stringclasses 1
value |
---|---|---|---|
A General and Flexible Multi-concept Parsing Framework for Multilingual
Semantic Matching | Sentence semantic matching is a research hotspot in natural language
processing, which is considerably significant in various key scenarios, such as
community question answering, searching, chatbot, and recommendation. Since
most of the advanced models directly model the semantic relevance among words
between two sentences while neglecting the \textit{keywords} and
\textit{intents} concepts of them, DC-Match is proposed to disentangle keywords
from intents and utilizes them to optimize the matching performance. Although
DC-Match is a simple yet effective method for semantic matching, it highly
depends on the external NER techniques to identify the keywords of sentences,
which limits the performance of semantic matching for minor languages since
satisfactory NER tools are usually hard to obtain. In this paper, we propose to
generally and flexibly resolve the text into multi concepts for multilingual
semantic matching to liberate the model from the reliance on NER models. To
this end, we devise a \underline{M}ulti-\underline{C}oncept \underline{P}arsed
\underline{S}emantic \underline{M}atching framework based on the pre-trained
language models, abbreviated as \textbf{MCP-SM}, to extract various concepts
and infuse them into the classification tokens. We conduct comprehensive
experiments on English datasets QQP and MRPC, and Chinese dataset Medical-SM.
Besides, we experiment on Arabic datasets MQ2Q and XNLI, the outstanding
performance further prove MCP-SM's applicability in low-resource languages.
| 2,024 | Computation and Language |
Data Augmentation using LLMs: Data Perspectives, Learning Paradigms and
Challenges | In the rapidly evolving field of machine learning (ML), data augmentation
(DA) has emerged as a pivotal technique for enhancing model performance by
diversifying training examples without the need for additional data collection.
This survey explores the transformative impact of Large Language Models (LLMs)
on DA, particularly addressing the unique challenges and opportunities they
present in the context of natural language processing (NLP) and beyond. From a
data perspective and a learning perspective, we examine various strategies that
utilize Large Language Models for data augmentation, including a novel
exploration of learning paradigms where LLM-generated data is used for further
training. Additionally, this paper delineates the primary challenges faced in
this domain, ranging from controllable data augmentation to multi modal data
augmentation. This survey highlights the paradigm shift introduced by LLMs in
DA, aims to serve as a foundational guide for researchers and practitioners in
this field.
| 2,024 | Computation and Language |
The Case for Evaluating Multimodal Translation Models on Text Datasets | A good evaluation framework should evaluate multimodal machine translation
(MMT) models by measuring 1) their use of visual information to aid in the
translation task and 2) their ability to translate complex sentences such as
done for text-only machine translation. However, most current work in MMT is
evaluated against the Multi30k testing sets, which do not measure these
properties. Namely, the use of visual information by the MMT model cannot be
shown directly from the Multi30k test set results and the sentences in Multi30k
are are image captions, i.e., short, descriptive sentences, as opposed to
complex sentences that typical text-only machine translation models are
evaluated against.
Therefore, we propose that MMT models be evaluated using 1) the CoMMuTE
evaluation framework, which measures the use of visual information by MMT
models, 2) the text-only WMT news translation task test sets, which evaluates
translation performance against complex sentences, and 3) the Multi30k test
sets, for measuring MMT model performance against a real MMT dataset. Finally,
we evaluate recent MMT models trained solely against the Multi30k dataset
against our proposed evaluation framework and demonstrate the dramatic drop
performance against text-only testing sets compared to recent text-only MT
models.
| 2,024 | Computation and Language |
Socratic Reasoning Improves Positive Text Rewriting | Reframing a negative into a positive thought is at the crux of several
cognitive approaches to mental health and psychotherapy that could be made more
accessible by large language model-based solutions. Such reframing is typically
non-trivial and requires multiple rationalization steps to uncover the
underlying issue of a negative thought and transform it to be more positive.
However, this rationalization process is currently neglected by both datasets
and models which reframe thoughts in one step. In this work, we address this
gap by augmenting open-source datasets for positive text rewriting with
synthetically-generated Socratic rationales using a novel framework called
\textsc{SocraticReframe}. \textsc{SocraticReframe} uses a sequence of
question-answer pairs to rationalize the thought rewriting process. We show
that such Socratic rationales significantly improve positive text rewriting for
different open-source LLMs according to both automatic and human evaluations
guided by criteria from psychotherapy research.
| 2,024 | Computation and Language |
Learning to Use Tools via Cooperative and Interactive Agents | Tool learning empowers large language models (LLMs) as agents to use external
tools to extend their capability. Existing methods employ one single LLM-based
agent to iteratively select and execute tools, thereafter incorporating the
result into the next action prediction. However, they still suffer from
potential performance degradation when addressing complex tasks due to: (1) the
limitation of the inherent capability of a single LLM to perform diverse
actions, and (2) the struggle to adaptively correct mistakes when the task
fails. To mitigate these problems, we propose the ConAgents, a Cooperative and
interactive Agents framework, which modularizes the workflow of tool learning
into Grounding, Execution, and Observing agents. We also introduce an iterative
calibration (IterCali) method, enabling the agents to adapt themselves based on
the feedback from the tool environment. Experiments conducted on three datasets
demonstrate the superiority of our ConAgents (e.g., 6 point improvement over
the SOTA baseline). We further provide fine-granularity analysis for the
efficiency and consistency of our framework.
| 2,024 | Computation and Language |
Adding Multimodal Capabilities to a Text-only Translation Model | While most current work in multimodal machine translation (MMT) uses the
Multi30k dataset for training and evaluation, we find that the resulting models
overfit to the Multi30k dataset to an extreme degree. Consequently, these
models perform very badly when evaluated against typical text-only testing sets
such as the WMT newstest datasets. In order to perform well on both Multi30k
and typical text-only datasets, we use a performant text-only machine
translation (MT) model as the starting point of our MMT model. We add
vision-text adapter layers connected via gating mechanisms to the MT model, and
incrementally transform the MT model into an MMT model by 1) pre-training using
vision-based masking of the source text and 2) fine-tuning on Multi30k.
| 2,024 | Computation and Language |
Detecting Concrete Visual Tokens for Multimodal Machine Translation | The challenge of visual grounding and masking in multimodal machine
translation (MMT) systems has encouraged varying approaches to the detection
and selection of visually-grounded text tokens for masking. We introduce new
methods for detection of visually and contextually relevant (concrete) tokens
from source sentences, including detection with natural language processing
(NLP), detection with object detection, and a joint detection-verification
technique. We also introduce new methods for selection of detected tokens,
including shortest $n$ tokens, longest $n$ tokens, and all detected concrete
tokens. We utilize the GRAM MMT architecture to train models against
synthetically collated multimodal datasets of source images with masked
sentences, showing performance improvements and improved usage of visual
context during translation tasks over the baseline model.
| 2,024 | Computation and Language |
KnowAgent: Knowledge-Augmented Planning for LLM-Based Agents | Large Language Models (LLMs) have demonstrated great potential in complex
reasoning tasks, yet they fall short when tackling more sophisticated
challenges, especially when interacting with environments through generating
executable actions. This inadequacy primarily stems from the lack of built-in
action knowledge in language agents, which fails to effectively guide the
planning trajectories during task solving and results in planning
hallucination. To address this issue, we introduce KnowAgent, a novel approach
designed to enhance the planning capabilities of LLMs by incorporating explicit
action knowledge. Specifically, KnowAgent employs an action knowledge base and
a knowledgeable self-learning strategy to constrain the action path during
planning, enabling more reasonable trajectory synthesis, and thereby enhancing
the planning performance of language agents. Experimental results on HotpotQA
and ALFWorld based on various backbone models demonstrate that KnowAgent can
achieve comparable or superior performance to existing baselines. Further
analysis indicates the effectiveness of KnowAgent in terms of planning
hallucinations mitigation. Code is available in
https://github.com/zjunlp/KnowAgent.
| 2,024 | Computation and Language |
"In Dialogues We Learn": Towards Personalized Dialogue Without
Pre-defined Profiles through In-Dialogue Learning | Personalized dialogue systems have gained significant attention in recent
years for their ability to generate responses in alignment with different
personas. However, most existing approaches rely on pre-defined personal
profiles, which are not only time-consuming and labor-intensive to create but
also lack flexibility. We propose In-Dialogue Learning (IDL), a fine-tuning
framework that enhances the ability of pre-trained large language models to
leverage dialogue history to characterize persona for completing personalized
dialogue generation tasks without pre-defined profiles. Our experiments on
three datasets demonstrate that IDL brings substantial improvements, with BLEU
and ROUGE scores increasing by up to 200% and 247%, respectively. Additionally,
the results of human evaluations further validate the efficacy of our proposed
method.
| 2,024 | Computation and Language |
Angry Men, Sad Women: Large Language Models Reflect Gendered Stereotypes
in Emotion Attribution | Large language models (LLMs) reflect societal norms and biases, especially
about gender. While societal biases and stereotypes have been extensively
researched in various NLP applications, there is a surprising gap for emotion
analysis. However, emotion and gender are closely linked in societal discourse.
E.g., women are often thought of as more empathetic, while men's anger is more
socially accepted. To fill this gap, we present the first comprehensive study
of gendered emotion attribution in five state-of-the-art LLMs (open- and
closed-source). We investigate whether emotions are gendered, and whether these
variations are based on societal stereotypes. We prompt the models to adopt a
gendered persona and attribute emotions to an event like 'When I had a serious
argument with a dear person'. We then analyze the emotions generated by the
models in relation to the gender-event pairs. We find that all models
consistently exhibit gendered emotions, influenced by gender stereotypes. These
findings are in line with established research in psychology and gender
studies. Our study sheds light on the complex societal interplay between
language, gender, and emotion. The reproduction of emotion stereotypes in LLMs
allows us to use those models to study the topic in detail, but raises
questions about the predictive use of those same LLMs for emotion applications.
| 2,024 | Computation and Language |
CoGenesis: A Framework Collaborating Large and Small Language Models for
Secure Context-Aware Instruction Following | With the advancement of language models (LMs), their exposure to private data
is increasingly inevitable, and their deployment (especially for smaller ones)
on personal devices, such as PCs and smartphones, has become a prevailing
trend. In contexts laden with user information, enabling models to both
safeguard user privacy and execute commands efficiently emerges as an essential
research imperative. In this paper, we propose CoGenesis, a collaborative
generation framework integrating large (hosted on cloud infrastructure) and
small models (deployed on local devices) to address privacy concerns logically.
Initially, we design a pipeline to create personalized writing instruction
datasets enriched with extensive context details as the testbed of this
research issue. Subsequently, we introduce two variants of CoGenesis based on
sketch and logits respectively. Our experimental findings, based on our
synthesized dataset and two additional open-source datasets, indicate that: 1)
Large-scale models perform well when provided with user context but struggle in
the absence of such context. 2) While specialized smaller models fine-tuned on
the synthetic dataset show promise, they still lag behind their larger
counterparts. 3) Our CoGenesis framework, utilizing mixed-scale models,
showcases competitive performance, providing a feasible solution to privacy
issues.
| 2,024 | Computation and Language |
Language Guided Exploration for RL Agents in Text Environments | Real-world sequential decision making is characterized by sparse rewards and
large decision spaces, posing significant difficulty for experiential learning
systems like $\textit{tabula rasa}$ reinforcement learning (RL) agents. Large
Language Models (LLMs), with a wealth of world knowledge, can help RL agents
learn quickly and adapt to distribution shifts. In this work, we introduce
Language Guided Exploration (LGE) framework, which uses a pre-trained language
model (called GUIDE ) to provide decision-level guidance to an RL agent (called
EXPLORER). We observe that on ScienceWorld (Wang et al.,2022), a challenging
text environment, LGE outperforms vanilla RL agents significantly and also
outperforms other sophisticated methods like Behaviour Cloning and Text
Decision Transformer.
| 2,024 | Computation and Language |
Design2Code: How Far Are We From Automating Front-End Engineering? | Generative AI has made rapid advancements in recent years, achieving
unprecedented capabilities in multimodal understanding and code generation.
This can enable a new paradigm of front-end development, in which multimodal
LLMs might directly convert visual designs into code implementations. In this
work, we formalize this as a Design2Code task and conduct comprehensive
benchmarking. Specifically, we manually curate a benchmark of 484 diverse
real-world webpages as test cases and develop a set of automatic evaluation
metrics to assess how well current multimodal LLMs can generate the code
implementations that directly render into the given reference webpages, given
the screenshots as input. We also complement automatic metrics with
comprehensive human evaluations. We develop a suite of multimodal prompting
methods and show their effectiveness on GPT-4V and Gemini Pro Vision. We
further finetune an open-source Design2Code-18B model that successfully matches
the performance of Gemini Pro Vision. Both human evaluation and automatic
metrics show that GPT-4V performs the best on this task compared to other
models. Moreover, annotators think GPT-4V generated webpages can replace the
original reference webpages in 49% of cases in terms of visual appearance and
content; and perhaps surprisingly, in 64% of cases GPT-4V generated webpages
are considered better than the original reference webpages. Our fine-grained
break-down metrics indicate that open-source models mostly lag in recalling
visual elements from the input webpages and in generating correct layout
designs, while aspects like text content and coloring can be drastically
improved with proper finetuning.
| 2,024 | Computation and Language |
PARADISE: Evaluating Implicit Planning Skills of Language Models with
Procedural Warnings and Tips Dataset | Recently, there has been growing interest within the community regarding
whether large language models are capable of planning or executing plans.
However, most prior studies use LLMs to generate high-level plans for
simplified scenarios lacking linguistic complexity and domain diversity,
limiting analysis of their planning abilities. These setups constrain
evaluation methods (e.g., predefined action space), architectural choices
(e.g., only generative models), and overlook the linguistic nuances essential
for realistic analysis. To tackle this, we present PARADISE, an abductive
reasoning task using Q\&A format on practical procedural text sourced from
wikiHow. It involves warning and tip inference tasks directly associated with
goals, excluding intermediary steps, with the aim of testing the ability of the
models to infer implicit knowledge of the plan solely from the given goal. Our
experiments, utilizing fine-tuned language models and zero-shot prompting,
reveal the effectiveness of task-specific small models over large language
models in most scenarios. Despite advancements, all models fall short of human
performance. Notably, our analysis uncovers intriguing insights, such as
variations in model behavior with dropped keywords, struggles of BERT-family
and GPT-4 with physical and abstract goals, and the proposed tasks offering
valuable prior knowledge for other unseen procedural tasks. The PARADISE
dataset and associated resources are publicly available for further research
exploration with https://github.com/GGLAB-KU/paradise.
| 2,024 | Computation and Language |
Reliable, Adaptable, and Attributable Language Models with Retrieval | Parametric language models (LMs), which are trained on vast amounts of web
data, exhibit remarkable flexibility and capability. However, they still face
practical challenges such as hallucinations, difficulty in adapting to new data
distributions, and a lack of verifiability. In this position paper, we advocate
for retrieval-augmented LMs to replace parametric LMs as the next generation of
LMs. By incorporating large-scale datastores during inference,
retrieval-augmented LMs can be more reliable, adaptable, and attributable.
Despite their potential, retrieval-augmented LMs have yet to be widely adopted
due to several obstacles: specifically, current retrieval-augmented LMs
struggle to leverage helpful text beyond knowledge-intensive tasks such as
question answering, have limited interaction between retrieval and LM
components, and lack the infrastructure for scaling. To address these, we
propose a roadmap for developing general-purpose retrieval-augmented LMs. This
involves a reconsideration of datastores and retrievers, the exploration of
pipelines with improved retriever-LM interaction, and significant investment in
infrastructure for efficient training and inference.
| 2,024 | Computation and Language |
MAGID: An Automated Pipeline for Generating Synthetic Multi-modal
Datasets | Development of multimodal interactive systems is hindered by the lack of
rich, multimodal (text, images) conversational data, which is needed in large
quantities for LLMs. Previous approaches augment textual dialogues with
retrieved images, posing privacy, diversity, and quality constraints. In this
work, we introduce \textbf{M}ultimodal \textbf{A}ugmented \textbf{G}enerative
\textbf{I}mages \textbf{D}ialogues (MAGID), a framework to augment text-only
dialogues with diverse and high-quality images. Subsequently, a diffusion model
is applied to craft corresponding images, ensuring alignment with the
identified text. Finally, MAGID incorporates an innovative feedback loop
between an image description generation module (textual LLM) and image quality
modules (addressing aesthetics, image-text matching, and safety), that work in
tandem to generate high-quality and multi-modal dialogues. We compare MAGID to
other SOTA baselines on three dialogue datasets, using automated and human
evaluation. Our results show that MAGID is comparable to or better than
baselines, with significant improvements in human evaluation, especially
against retrieval baselines where the image database is small.
| 2,024 | Computation and Language |
Mad Libs Are All You Need: Augmenting Cross-Domain Document-Level Event
Argument Data | Document-Level Event Argument Extraction (DocEAE) is an extremely difficult
information extraction problem -- with significant limitations in low-resource
cross-domain settings. To address this problem, we introduce Mad Lib Aug (MLA),
a novel generative DocEAE data augmentation framework. Our approach leverages
the intuition that Mad Libs, which are categorically masked documents used as a
part of a popular game, can be generated and solved by LLMs to produce data for
DocEAE. Using MLA, we achieve a 2.6-point average improvement in overall F1
score. Moreover, this approach achieves a 3.9 and 5.2 point average increase in
zero and few-shot event roles compared to augmentation-free baselines across
all experiments.
To better facilitate analysis of cross-domain DocEAE, we additionally
introduce a new metric, Role-Depth F1 (RDF1), which uses statistical depth to
identify roles in the target domain which are semantic outliers with respect to
roles observed in the source domain. Our experiments show that MLA augmentation
can boost RDF1 performance by an average of 5.85 points compared to
non-augmented datasets.
| 2,024 | Computation and Language |
Best of Both Worlds: A Pliable and Generalizable Neuro-Symbolic Approach
for Relation Classification | This paper introduces a novel neuro-symbolic architecture for relation
classification (RC) that combines rule-based methods with contemporary deep
learning techniques. This approach capitalizes on the strengths of both
paradigms: the adaptability of rule-based systems and the generalization power
of neural networks. Our architecture consists of two components: a declarative
rule-based model for transparent classification and a neural component to
enhance rule generalizability through semantic text matching. Notably, our
semantic matcher is trained in an unsupervised domain-agnostic way, solely with
synthetic data. Further, these components are loosely coupled, allowing for
rule modifications without retraining the semantic matcher. In our evaluation,
we focused on two few-shot relation classification datasets: Few-Shot TACRED
and a Few-Shot version of NYT29. We show that our proposed method outperforms
previous state-of-the-art models in three out of four settings, despite not
seeing any human-annotated training data. Further, we show that our approach
remains modular and pliable, i.e., the corresponding rules can be locally
modified to improve the overall model. Human interventions to the rules for the
TACRED relation \texttt{org:parents} boost the performance on that relation by
as much as 26\% relative improvement, without negatively impacting the other
relations, and without retraining the semantic matching component.
| 2,024 | Computation and Language |
Book2Dial: Generating Teacher-Student Interactions from Textbooks for
Cost-Effective Development of Educational Chatbots | Educational chatbots are a promising tool for assisting student learning.
However, the development of effective chatbots in education has been
challenging, as high-quality data is seldom available in this domain. In this
paper, we propose a framework for generating synthetic teacher-student
interactions grounded in a set of textbooks. Our approaches capture one aspect
of learning interactions where curious students with partial knowledge
interactively ask a teacher questions about the material in the textbook. We
highlight various quality criteria that such dialogues should fulfill and
compare several approaches relying on either prompting or fine-tuning large
language models. We use synthetic dialogues to train educational chatbots and
show benefits of further fine-tuning in different educational domains. However,
human evaluation shows that our best data synthesis method still suffers from
hallucinations and tends to reiterate information from previous conversations.
Our findings offer insights for future efforts in synthesizing conversational
data that strikes a balance between size and quality. We will open-source our
data and code.
| 2,024 | Computation and Language |
Guardrail Baselines for Unlearning in LLMs | Recent work has demonstrated that fine-tuning is a promising approach to
`unlearn' concepts from large language models. However, fine-tuning can be
expensive, as it requires both generating a set of examples and running
iterations of fine-tuning to update the model. In this work, we show that
simple guardrail-based approaches such as prompting and filtering can achieve
unlearning results comparable to fine-tuning. We recommend that researchers
investigate these lightweight baselines when evaluating the performance of more
computationally intensive fine-tuning methods. While we do not claim that
methods such as prompting or filtering are universal solutions to the problem
of unlearning, our work suggests the need for evaluation metrics that can
better separate the power of guardrails vs. fine-tuning, and highlights
scenarios where guardrails themselves may be advantageous for unlearning, such
as in generating examples for fine-tuning or unlearning when only API access is
available.
| 2,024 | Computation and Language |
DIVERSE: Deciphering Internet Views on the U.S. Military Through Video
Comment Stance Analysis, A Novel Benchmark Dataset for Stance Classification | Stance detection of social media text is a key component of downstream tasks
involving the identification of groups of users with opposing opinions on
contested topics such as vaccination and within arguments. In particular,
stance provides an indication of an opinion towards an entity. This paper
introduces DIVERSE, a dataset of over 173,000 YouTube video comments annotated
for their stance towards videos of the U.S. military. The stance is annotated
through a human-guided, machine-assisted labeling methodology that makes use of
weak signals of tone within the sentence as supporting indicators, as opposed
to using manual annotations by humans. These weak signals consist of the
presence of hate speech and sarcasm, the presence of specific keywords, the
sentiment of the text, and the stance inference from two Large Language Models.
The weak signals are then consolidated using a data programming model before
each comment is annotated with a final stance label. On average, the videos
have 200 comments each, and the stance of the comments skews slightly towards
the "against" characterization for both the U.S. Army and the videos posted on
the channel.
| 2,024 | Computation and Language |
Scope of Large Language Models for Mining Emerging Opinions in Online
Health Discourse | In this paper, we develop an LLM-powered framework for the curation and
evaluation of emerging opinion mining in online health communities. We
formulate emerging opinion mining as a pairwise stance detection problem
between (title, comment) pairs sourced from Reddit, where post titles contain
emerging health-related claims on a topic that is not predefined. The claims
are either explicitly or implicitly expressed by the user. We detail (i) a
method of claim identification -- the task of identifying if a post title
contains a claim and (ii) an opinion mining-driven evaluation framework for
stance detection using LLMs.
We facilitate our exploration by releasing a novel test dataset, Long
COVID-Stance, or LC-stance, which can be used to evaluate LLMs on the tasks of
claim identification and stance detection in online health communities. Long
Covid is an emerging post-COVID disorder with uncertain and complex treatment
guidelines, thus making it a suitable use case for our task. LC-Stance contains
long COVID treatment related discourse sourced from a Reddit community. Our
evaluation shows that GPT-4 significantly outperforms prior works on zero-shot
stance detection. We then perform thorough LLM model diagnostics, identifying
the role of claim type (i.e. implicit vs explicit claims) and comment length as
sources of model error.
| 2,024 | Computation and Language |
Learning to Maximize Mutual Information for Chain-of-Thought
Distillation | Knowledge distillation, the technique of transferring knowledge from large,
complex models to smaller ones, marks a pivotal step towards efficient AI
deployment. Distilling Step-by-Step (DSS), a novel method utilizing
chain-of-thought (CoT) distillation, has demonstrated promise by imbuing
smaller models with the superior reasoning capabilities of their larger
counterparts. In DSS, the distilled model acquires the ability to generate
rationales and predict labels concurrently through a multi-task learning
framework. However, DSS overlooks the intrinsic relationship between the two
training tasks, leading to ineffective integration of CoT knowledge with the
task of label prediction. To this end, we investigate the mutual relationship
of the two tasks from Information Bottleneck perspective and formulate it as
maximizing the mutual information of the representation features of the two
tasks. We propose a variational approach to solve this optimization problem
using a learning-based method. Our experimental results across four datasets
demonstrate that our method outperforms the state-of-the-art DSS. Our findings
offer insightful guidance for future research on language model distillation as
well as applications involving CoT. Code and models will be released soon.
| 2,024 | Computation and Language |
Japanese-English Sentence Translation Exercises Dataset for Automatic
Grading | This paper proposes the task of automatic assessment of Sentence Translation
Exercises (STEs), that have been used in the early stage of L2 language
learning. We formalize the task as grading student responses for each rubric
criterion pre-specified by the educators. We then create a dataset for STE
between Japanese and English including 21 questions, along with a total of 3,
498 student responses (167 on average). The answer responses were collected
from students and crowd workers. Using this dataset, we demonstrate the
performance of baselines including finetuned BERT and GPT models with few-shot
in-context learning. Experimental results show that the baseline model with
finetuned BERT was able to classify correct responses with approximately 90% in
F1, but only less than 80% for incorrect responses. Furthermore, the GPT models
with few-shot learning show poorer results than finetuned BERT, indicating that
our newly proposed task presents a challenging issue, even for the
stateof-the-art large language models.
| 2,024 | Computation and Language |
Negating Negatives: Alignment without Human Positive Samples via
Distributional Dispreference Optimization | Large language models (LLMs) have revolutionized the role of AI, yet also
pose potential risks of propagating unethical content. Alignment technologies
have been introduced to steer LLMs towards human preference, gaining increasing
attention. Despite notable breakthroughs in this direction, existing methods
heavily rely on high-quality positive-negative training pairs, suffering from
noisy labels and the marginal distinction between preferred and dispreferred
response data. Given recent LLMs' proficiency in generating helpful responses,
this work pivots towards a new research focus: achieving alignment using solely
human-annotated negative samples, preserving helpfulness while reducing
harmfulness. For this purpose, we propose Distributional Dispreference
Optimization (D$^2$O), which maximizes the discrepancy between the generated
responses and the dispreferred ones to effectively eschew harmful information.
We theoretically demonstrate that D$^2$O is equivalent to learning a
distributional instead of instance-level preference model reflecting human
dispreference against the distribution of negative responses. Besides, D$^2$O
integrates an implicit Jeffrey Divergence regularization to balance the
exploitation and exploration of reference policies and converges to a
non-negative one during training. Extensive experiments demonstrate that our
method achieves comparable generation quality and surpasses the latest
baselines in producing less harmful and more informative responses with better
training stability and faster convergence.
| 2,024 | Computation and Language |
Mixture-of-LoRAs: An Efficient Multitask Tuning for Large Language
Models | Instruction Tuning has the potential to stimulate or enhance specific
capabilities of large language models (LLMs). However, achieving the right
balance of data is crucial to prevent catastrophic forgetting and interference
between tasks. To address these limitations and enhance training flexibility,
we propose the Mixture-of-LoRAs (MoA) architecture which is a novel and
parameter-efficient tuning method designed for multi-task learning with LLMs.
In this paper, we start by individually training multiple domain-specific LoRA
modules using corresponding supervised corpus data. These LoRA modules can be
aligned with the expert design principles observed in Mixture-of-Experts (MoE).
Subsequently, we combine the multiple LoRAs using an explicit routing strategy
and introduce domain labels to facilitate multi-task learning, which help
prevent interference between tasks and ultimately enhances the performance of
each individual task. Furthermore, each LoRA model can be iteratively adapted
to a new domain, allowing for quick domain-specific adaptation. Experiments on
diverse tasks demonstrate superior and robust performance, which can further
promote the wide application of domain-specific LLMs.
| 2,024 | Computation and Language |
VLSP 2023 -- LTER: A Summary of the Challenge on Legal Textual
Entailment Recognition | In this new era of rapid AI development, especially in language processing,
the demand for AI in the legal domain is increasingly critical. In the context
where research in other languages such as English, Japanese, and Chinese has
been well-established, we introduce the first fundamental research for the
Vietnamese language in the legal domain: legal textual entailment recognition
through the Vietnamese Language and Speech Processing workshop. In analyzing
participants' results, we discuss certain linguistic aspects critical in the
legal domain that pose challenges that need to be addressed.
| 2,024 | Computation and Language |
Magic Markup: Maintaining Document-External Markup with an LLM | Text documents, including programs, typically have human-readable semantic
structure. Historically, programmatic access to these semantics has required
explicit in-document tagging. Especially in systems where the text has an
execution semantics, this means it is an opt-in feature that is hard to support
properly. Today, language models offer a new method: metadata can be bound to
entities in changing text using a model's human-like understanding of
semantics, with no requirements on the document structure. This method expands
the applications of document annotation, a fundamental operation in program
writing, debugging, maintenance, and presentation. We contribute a system that
employs an intelligent agent to re-tag modified programs, enabling rich
annotations to automatically follow code as it evolves. We also contribute a
formal problem definition, an empirical synthetic benchmark suite, and our
benchmark generator. Our system achieves an accuracy of 90% on our benchmarks
and can replace a document's tags in parallel at a rate of 5 seconds per tag.
While there remains significant room for improvement, we find performance
reliable enough to justify further exploration of applications.
| 2,024 | Computation and Language |
A Knowledge Plug-and-Play Test Bed for Open-domain Dialogue Generation | Knowledge-based, open-domain dialogue generation aims to build chit-chat
systems that talk to humans using mined support knowledge. Many types and
sources of knowledge have previously been shown to be useful as support
knowledge. Even in the era of large language models, response generation
grounded in knowledge retrieved from additional up-to-date sources remains a
practically important approach. While prior work using single-source knowledge
has shown a clear positive correlation between the performances of knowledge
selection and response generation, there are no existing multi-source datasets
for evaluating support knowledge retrieval. Further, prior work has assumed
that the knowledge sources available at test time are the same as during
training. This unrealistic assumption unnecessarily handicaps models, as new
knowledge sources can become available after a model is trained. In this paper,
we present a high-quality benchmark named multi-source Wizard of Wikipedia
(Ms.WoW) for evaluating multi-source dialogue knowledge selection and response
generation. Unlike existing datasets, it contains clean support knowledge,
grounded at the utterance level and partitioned into multiple knowledge
sources. We further propose a new challenge, dialogue knowledge plug-and-play,
which aims to test an already trained dialogue model on using new support
knowledge from previously unseen sources in a zero-shot fashion.
| 2,024 | Computation and Language |
Towards Detecting AI-Generated Text within Human-AI Collaborative Hybrid
Texts | This study explores the challenge of sentence-level AI-generated text
detection within human-AI collaborative hybrid texts. Existing studies of
AI-generated text detection for hybrid texts often rely on synthetic datasets.
These typically involve hybrid texts with a limited number of boundaries. We
contend that studies of detecting AI-generated content within hybrid texts
should cover different types of hybrid texts generated in realistic settings to
better inform real-world applications. Therefore, our study utilizes the
CoAuthor dataset, which includes diverse, realistic hybrid texts generated
through the collaboration between human writers and an intelligent writing
system in multi-turn interactions. We adopt a two-step, segmentation-based
pipeline: (i) detect segments within a given hybrid text where each segment
contains sentences of consistent authorship, and (ii) classify the authorship
of each identified segment. Our empirical findings highlight (1) detecting
AI-generated sentences in hybrid texts is overall a challenging task because
(1.1) human writers' selecting and even editing AI-generated sentences based on
personal preferences adds difficulty in identifying the authorship of segments;
(1.2) the frequent change of authorship between neighboring sentences within
the hybrid text creates difficulties for segment detectors in identifying
authorship-consistent segments; (1.3) the short length of text segments within
hybrid texts provides limited stylistic cues for reliable authorship
determination; (2) before embarking on the detection process, it is beneficial
to assess the average length of segments within the hybrid text. This
assessment aids in deciding whether (2.1) to employ a text segmentation-based
strategy for hybrid texts with longer segments, or (2.2) to adopt a direct
sentence-by-sentence classification strategy for those with shorter segments.
| 2,024 | Computation and Language |
CLongEval: A Chinese Benchmark for Evaluating Long-Context Large
Language Models | Developing Large Language Models (LLMs) with robust long-context capabilities
has been the recent research focus, resulting in the emergence of long-context
LLMs proficient in Chinese. However, the evaluation of these models remains
underdeveloped due to a lack of benchmarks. To address this gap, we present
CLongEval, a comprehensive Chinese benchmark for evaluating long-context LLMs.
CLongEval is characterized by three key features: (1) Sufficient data volume,
comprising 7 distinct tasks and 7,267 examples; (2) Broad applicability,
accommodating to models with context windows size from 1K to 100K; (3) High
quality, with over 2,000 manually annotated question-answer pairs in addition
to the automatically constructed labels. With CLongEval, we undertake a
comprehensive assessment of 6 open-source long-context LLMs and 2 leading
commercial counterparts that feature both long-context abilities and
proficiency in Chinese. We also provide in-depth analysis based on the
empirical results, trying to shed light on the critical capabilities that
present challenges in long-context settings. The dataset, evaluation scripts,
and model outputs will be released.
| 2,024 | Computation and Language |
Unsupervised Multilingual Dense Retrieval via Generative Pseudo Labeling | Dense retrieval methods have demonstrated promising performance in
multilingual information retrieval, where queries and documents can be in
different languages. However, dense retrievers typically require a substantial
amount of paired data, which poses even greater challenges in multilingual
scenarios. This paper introduces UMR, an Unsupervised Multilingual dense
Retriever trained without any paired data. Our approach leverages the sequence
likelihood estimation capabilities of multilingual language models to acquire
pseudo labels for training dense retrievers. We propose a two-stage framework
which iteratively improves the performance of multilingual dense retrievers.
Experimental results on two benchmark datasets show that UMR outperforms
supervised baselines, showcasing the potential of training multilingual
retrievers without paired data, thereby enhancing their practicality. Our
source code, data, and models are publicly available at
https://github.com/MiuLab/UMR
| 2,024 | Computation and Language |
BiVert: Bidirectional Vocabulary Evaluation using Relations for Machine
Translation | Neural machine translation (NMT) has progressed rapidly in the past few
years, promising improvements and quality translations for different languages.
Evaluation of this task is crucial to determine the quality of the translation.
Overall, insufficient emphasis is placed on the actual sense of the translation
in traditional methods. We propose a bidirectional semantic-based evaluation
method designed to assess the sense distance of the translation from the source
text. This approach employs the comprehensive multilingual encyclopedic
dictionary BabelNet. Through the calculation of the semantic distance between
the source and its back translation of the output, our method introduces a
quantifiable approach that empowers sentence comparison on the same linguistic
level. Factual analysis shows a strong correlation between the average
evaluation scores generated by our method and the human assessments across
various machine translation systems for English-German language pair. Finally,
our method proposes a new multilingual approach to rank MT systems without the
need for parallel corpora.
| 2,024 | Computation and Language |
Benchmarking Hallucination in Large Language Models based on
Unanswerable Math Word Problem | Large language models (LLMs) are highly effective in various natural language
processing (NLP) tasks. However, they are susceptible to producing unreliable
conjectures in ambiguous contexts called hallucination. This paper presents a
new method for evaluating LLM hallucination in Question Answering (QA) based on
the unanswerable math word problem (MWP). To support this approach, we
innovatively develop a dataset called Unanswerable Math Word Problem (UMWP)
which comprises 5200 questions across five categories. We developed an
evaluation methodology combining text similarity and mathematical expression
detection to determine whether LLM considers the question unanswerable. The
results of extensive experiments conducted on 31 LLMs, including GPT-3,
InstructGPT, LLaMA, and Claude, demonstrate that in-context learning and
reinforcement learning with human feedback (RLHF) training significantly
enhance the model's ability to avoid hallucination. We show that utilizing MWP
is a reliable and effective approach to assess hallucination. Our code and data
are available at https://github.com/Yuki-Asuuna/UMWP.
| 2,024 | Computation and Language |
gaHealth: An English-Irish Bilingual Corpus of Health Data | Machine Translation is a mature technology for many high-resource language
pairs. However in the context of low-resource languages, there is a paucity of
parallel data datasets available for developing translation models.
Furthermore, the development of datasets for low-resource languages often
focuses on simply creating the largest possible dataset for generic
translation. The benefits and development of smaller in-domain datasets can
easily be overlooked. To assess the merits of using in-domain data, a dataset
for the specific domain of health was developed for the low-resource English to
Irish language pair. Our study outlines the process used in developing the
corpus and empirically demonstrates the benefits of using an in-domain dataset
for the health domain. In the context of translating health-related data,
models developed using the gaHealth corpus demonstrated a maximum BLEU score
improvement of 22.2 points (40%) when compared with top performing models from
the LoResMT2021 Shared Task. Furthermore, we define linguistic guidelines for
developing gaHealth, the first bilingual corpus of health data for the Irish
language, which we hope will be of use to other creators of low-resource data
sets. gaHealth is now freely available online and is ready to be explored for
further research.
| 2,022 | Computation and Language |
Enhancing ASD detection accuracy: a combined approach of machine
learning and deep learning models with natural language processing | Purpose: Our study explored the use of artificial intelligence (AI) to
diagnose autism spectrum disorder (ASD). It focused on machine learning (ML)
and deep learning (DL) to detect ASD from text inputs on social media,
addressing challenges in traditional ASD diagnosis.
Methods: We used natural language processing (NLP), ML, and DL models
(including decision trees, XGB, KNN, RNN, LSTM, Bi-LSTM, BERT, and BERTweet) to
analyze 404,627 tweets, classifying them based on ASD or non-ASD authors. A
subset of 90,000 tweets was used for model training and testing.
Results: Our AI models showed high accuracy, with an 88% success rate in
identifying texts from individuals with ASD.
Conclusion: The study demonstrates AI's potential in improving ASD diagnosis,
especially in children, highlighting the importance of early detection.
| 2,024 | Computation and Language |
Design of an Open-Source Architecture for Neural Machine Translation | adaptNMT is an open-source application that offers a streamlined approach to
the development and deployment of Recurrent Neural Networks and Transformer
models. This application is built upon the widely-adopted OpenNMT ecosystem,
and is particularly useful for new entrants to the field, as it simplifies the
setup of the development environment and creation of train, validation, and
test splits. The application offers a graphing feature that illustrates the
progress of model training, and employs SentencePiece for creating subword
segmentation models. Furthermore, the application provides an intuitive user
interface that facilitates hyperparameter customization. Notably, a
single-click model development approach has been implemented, and models
developed by adaptNMT can be evaluated using a range of metrics. To encourage
eco-friendly research, adaptNMT incorporates a green report that flags the
power consumption and kgCO${_2}$ emissions generated during model development.
The application is freely available.
| 2,023 | Computation and Language |
Multimodal Large Language Models to Support Real-World Fact-Checking | Multimodal large language models (MLLMs) carry the potential to support
humans in processing vast amounts of information. While MLLMs are already being
used as a fact-checking tool, their abilities and limitations in this regard
are understudied. Here is aim to bridge this gap. In particular, we propose a
framework for systematically assessing the capacity of current multimodal
models to facilitate real-world fact-checking. Our methodology is
evidence-free, leveraging only these models' intrinsic knowledge and reasoning
capabilities. By designing prompts that extract models' predictions,
explanations, and confidence levels, we delve into research questions
concerning model accuracy, robustness, and reasons for failure. We empirically
find that (1) GPT-4V exhibits superior performance in identifying malicious and
misleading multimodal claims, with the ability to explain the unreasonable
aspects and underlying motives, and (2) existing open-source models exhibit
strong biases and are highly sensitive to the prompt. Our study offers insights
into combating false multimodal information and building secure, trustworthy
multimodal models. To the best of our knowledge, we are the first to evaluate
MLLMs for real-world fact-checking.
| 2,024 | Computation and Language |
GPTopic: Dynamic and Interactive Topic Representations | Topic modeling seems to be almost synonymous with generating lists of top
words to represent topics within large text corpora. However, deducing a topic
from such list of individual terms can require substantial expertise and
experience, making topic modelling less accessible to people unfamiliar with
the particularities and pitfalls of top-word interpretation. A topic
representation limited to top-words might further fall short of offering a
comprehensive and easily accessible characterization of the various aspects,
facets and nuances a topic might have. To address these challenges, we
introduce GPTopic, a software package that leverages Large Language Models
(LLMs) to create dynamic, interactive topic representations. GPTopic provides
an intuitive chat interface for users to explore, analyze, and refine topics
interactively, making topic modeling more accessible and comprehensive. The
corresponding code is available here: https://github. com/05ec6602be/GPTopic.
| 2,024 | Computation and Language |
Apollo: Lightweight Multilingual Medical LLMs towards Democratizing
Medical AI to 6B People | Despite the vast repository of global medical knowledge predominantly being
in English, local languages are crucial for delivering tailored healthcare
services, particularly in areas with limited medical resources. To extend the
reach of medical AI advancements to a broader population, we aim to develop
medical LLMs across the six most widely spoken languages, encompassing a global
population of 6.1 billion. This effort culminates in the creation of the
ApolloCorpora multilingual medical dataset and the XMedBench benchmark. In the
multilingual medical benchmark, the released Apollo models, at various
relatively-small sizes (i.e., 0.5B, 1.8B, 2B, 6B, and 7B), achieve the best
performance among models of equivalent size. Especially, Apollo-7B is the
state-of-the-art multilingual medical LLMs up to 70B. Additionally, these lite
models could be used to improve the multi-lingual medical capabilities of
larger models without fine-tuning in a proxy-tuning fashion. We will
open-source training corpora, code, model weights and evaluation benchmark.
| 2,024 | Computation and Language |
General2Specialized LLMs Translation for E-commerce | Existing Neural Machine Translation (NMT) models mainly handle translation in
the general domain, while overlooking domains with special writing formulas,
such as e-commerce and legal documents. Taking e-commerce as an example, the
texts usually include amounts of domain-related words and have more grammar
problems, which leads to inferior performances of current NMT methods. To
address these problems, we collect two domain-related resources, including a
set of term pairs (aligned Chinese-English bilingual terms) and a parallel
corpus annotated for the e-commerce domain. Furthermore, we propose a two-step
fine-tuning paradigm (named G2ST) with self-contrastive semantic enhancement to
transfer one general NMT model to the specialized NMT model for e-commerce. The
paradigm can be used for the NMT models based on Large language models (LLMs).
Extensive evaluations on real e-commerce titles demonstrate the superior
translation quality and robustness of our G2ST approach, as compared with
state-of-the-art NMT models such as LLaMA, Qwen, GPT-3.5, and even GPT-4.
| 2,024 | Computation and Language |
Rapidly Developing High-quality Instruction Data and Evaluation
Benchmark for Large Language Models with Minimal Human Effort: A Case Study
on Japanese | The creation of instruction data and evaluation benchmarks for serving Large
language models often involves enormous human annotation. This issue becomes
particularly pronounced when rapidly developing such resources for a
non-English language like Japanese. Instead of following the popular practice
of directly translating existing English resources into Japanese (e.g.,
Japanese-Alpaca), we propose an efficient self-instruct method based on GPT-4.
We first translate a small amount of English instructions into Japanese and
post-edit them to obtain native-level quality. GPT-4 then utilizes them as
demonstrations to automatically generate Japanese instruction data. We also
construct an evaluation benchmark containing 80 questions across 8 categories,
using GPT-4 to automatically assess the response quality of LLMs without human
references. The empirical results suggest that the models fine-tuned on our
GPT-4 self-instruct data significantly outperformed the Japanese-Alpaca across
all three base pre-trained models. Our GPT-4 self-instruct data allowed the
LLaMA 13B model to defeat GPT-3.5 (Davinci-003) with a 54.37\% win-rate. The
human evaluation exhibits the consistency between GPT-4's assessments and human
preference. Our high-quality instruction data and evaluation benchmark have
been released here.
| 2,024 | Computation and Language |
German also Hallucinates! Inconsistency Detection in News Summaries with
the Absinth Dataset | The advent of Large Language Models (LLMs) has led to remarkable progress on
a wide range of natural language processing tasks. Despite the advances, these
large-sized models still suffer from hallucinating information in their output,
which poses a major issue in automatic text summarization, as we must guarantee
that the generated summary is consistent with the content of the source
document. Previous research addresses the challenging task of detecting
hallucinations in the output (i.e. inconsistency detection) in order to
evaluate the faithfulness of the generated summaries. However, these works
primarily focus on English and recent multilingual approaches lack German data.
This work presents absinth, a manually annotated dataset for hallucination
detection in German news summarization and explores the capabilities of novel
open-source LLMs on this task in both fine-tuning and in-context learning
settings. We open-source and release the absinth dataset to foster further
research on hallucination detection in German.
| 2,024 | Computation and Language |
PPTC-R benchmark: Towards Evaluating the Robustness of Large Language
Models for PowerPoint Task Completion | The growing dependence on Large Language Models (LLMs) for finishing user
instructions necessitates a comprehensive understanding of their robustness to
complex task completion in real-world situations. To address this critical
need, we propose the PowerPoint Task Completion Robustness benchmark (PPTC-R)
to measure LLMs' robustness to the user PPT task instruction and software
version. Specifically, we construct adversarial user instructions by attacking
user instructions at sentence, semantic, and multi-language levels. To assess
the robustness of Language Models to software versions, we vary the number of
provided APIs to simulate both the newest version and earlier version settings.
Subsequently, we test 3 closed-source and 4 open-source LLMs using a benchmark
that incorporates these robustness settings, aiming to evaluate how deviations
impact LLMs' API calls for task completion. We find that GPT-4 exhibits the
highest performance and strong robustness in our benchmark, particularly in the
version update and the multilingual settings. However, we find that all LLMs
lose their robustness when confronted with multiple challenges (e.g.,
multi-turn) simultaneously, leading to significant performance drops. We
further analyze the robustness behavior and error reasons of LLMs in our
benchmark, which provide valuable insights for researchers to understand the
LLM's robustness in task completion and develop more robust LLMs and agents. We
release the code and data at \url{https://github.com/ZekaiGalaxy/PPTCR}.
| 2,024 | Computation and Language |
Evaluating the Elementary Multilingual Capabilities of Large Language
Models with MultiQ | Large language models (LLMs) need to serve everyone, including a global
majority of non-English speakers. However, most LLMs today, and open LLMs in
particular, are often intended for use in just English (e.g. Llama2, Mistral)
or a small handful of high-resource languages (e.g. Mixtral, Qwen). Recent
research shows that, despite limits in their intended use, people prompt LLMs
in many different languages. Therefore, in this paper, we investigate the basic
multilingual capabilities of state-of-the-art open LLMs beyond their intended
use. For this purpose, we introduce MultiQ, a new silver standard benchmark for
basic open-ended question answering with 27.4k test questions across a
typologically diverse set of 137 languages. With MultiQ, we evaluate language
fidelity, i.e.\ whether models respond in the prompted language, and question
answering accuracy. All LLMs we test respond faithfully and/or accurately for
at least some languages beyond their intended use. Most models are more
accurate when they respond faithfully. However, differences across models are
large, and there is a long tail of languages where models are neither accurate
nor faithful. We explore differences in tokenization as a potential explanation
for our findings, identifying possible correlations that warrant further
investigation.
| 2,024 | Computation and Language |
A Modular Approach for Multimodal Summarization of TV Shows | In this paper we address the task of summarizing television shows, which
touches key areas in AI research: complex reasoning, multiple modalities, and
long narratives. We present a modular approach where separate components
perform specialized sub-tasks which we argue affords greater flexibility
compared to end-to-end methods. Our modules involve detecting scene boundaries,
reordering scenes so as to minimize the number of cuts between different
events, converting visual information to text, summarizing the dialogue in each
scene, and fusing the scene summaries into a final summary for the entire
episode. We also present a new metric, PREFS (Precision and Recall Evaluation
of Summary FactS), to measure both precision and recall of generated summaries,
which we decompose into atomic facts. Tested on the recently released
SummScreen3D dataset Papalampidi and Lapata (2023), our method produces higher
quality summaries than comparison models, as measured with ROUGE and our new
fact-based metric.
| 2,024 | Computation and Language |
ShortGPT: Layers in Large Language Models are More Redundant Than You
Expect | As Large Language Models (LLMs) continue to advance in performance, their
size has escalated significantly, with current LLMs containing billions or even
trillions of parameters. However, in this study, we discovered that many layers
of LLMs exhibit high similarity, and some layers play a negligible role in
network functionality. Based on this observation, we define a metric called
Block Influence (BI) to gauge the significance of each layer in LLMs. We then
propose a straightforward pruning approach: layer removal, in which we directly
delete the redundant layers in LLMs based on their BI scores. Experiments
demonstrate that our method, which we call ShortGPT, significantly outperforms
previous state-of-the-art (SOTA) methods in model pruning. Moreover, ShortGPT
is orthogonal to quantization-like methods, enabling further reduction in
parameters and computation. The ability to achieve better results through
simple layer removal, as opposed to more complex pruning techniques, suggests a
high degree of redundancy in the model architecture.
| 2,024 | Computation and Language |
Emojinize: Enriching Any Text with Emoji Translations | Emoji have become ubiquitous in written communication, on the Web and beyond.
They can emphasize or clarify emotions, add details to conversations, or simply
serve decorative purposes. This casual use, however, barely scratches the
surface of the expressive power of emoji. To further unleash this power, we
present Emojinize, a method for translating arbitrary text phrases into
sequences of one or more emoji without requiring human input. By leveraging the
power of large language models, Emojinize can choose appropriate emoji by
disambiguating based on context (eg, cricket-bat vs bat) and can express
complex concepts compositionally by combining multiple emoji (eq, "Emojinize"
is translated to input-latin-letters right-arrow grinning-face). In a cloze
test--based user study, we show that Emojinize's emoji translations increase
the human guessability of masked words by 55%, whereas human-picked emoji
translations do so by only 29%. These results suggest that emoji provide a
sufficiently rich vocabulary to accurately translate a wide variety of words.
Moreover, annotating words and phrases with Emojinize's emoji translations
opens the door to numerous downstream applications, including children learning
how to read, adults learning foreign languages, and text understanding for
people with learning disabilities.
| 2,024 | Computation and Language |
Designing Informative Metrics for Few-Shot Example Selection | Pretrained language models (PLMs) have shown remarkable few-shot learning
capabilities when provided with properly formatted examples. However, selecting
the "best" examples remains an open challenge. We propose a complexity-based
prompt selection approach for sequence tagging tasks. This approach avoids the
training of a dedicated model for selection of examples, and instead uses
certain metrics to align the syntactico-semantic complexity of test sentences
and examples. We use both sentence- and word-level metrics to match the
complexity of examples to the (test) sentence being considered. Our results
demonstrate that our approach extracts greater performance from PLMs: it
achieves state-of-the-art performance on few-shot NER, achieving a 5% absolute
improvement in F1 score on the CoNLL2003 dataset for GPT-4. We also see large
gains of upto 28.85 points (F1/Acc.) in smaller models like GPT-j-6B.
| 2,024 | Computation and Language |
X-Shot: A Unified System to Handle Frequent, Few-shot and Zero-shot
Learning Simultaneously in Classification | In recent years, few-shot and zero-shot learning, which learn to predict
labels with limited annotated instances, have garnered significant attention.
Traditional approaches often treat frequent-shot (freq-shot; labels with
abundant instances), few-shot, and zero-shot learning as distinct challenges,
optimizing systems for just one of these scenarios. Yet, in real-world
settings, label occurrences vary greatly. Some of them might appear thousands
of times, while others might only appear sporadically or not at all. For
practical deployment, it is crucial that a system can adapt to any label
occurrence. We introduce a novel classification challenge: X-shot, reflecting a
real-world context where freq-shot, few-shot, and zero-shot labels co-occur
without predefined limits. Here, X can span from 0 to positive infinity. The
crux of X-shot centers on open-domain generalization and devising a system
versatile enough to manage various label scenarios. To solve X-shot, we propose
BinBin (Binary INference Based on INstruction following) that leverages the
Indirect Supervision from a large collection of NLP tasks via instruction
following, bolstered by Weak Supervision provided by large language models.
BinBin surpasses previous state-of-the-art techniques on three benchmark
datasets across multiple domains. To our knowledge, this is the first work
addressing X-shot learning, where X remains variable.
| 2,024 | Computation and Language |
KIWI: A Dataset of Knowledge-Intensive Writing Instructions for
Answering Research Questions | Large language models (LLMs) adapted to follow user instructions are now
widely deployed as conversational agents. In this work, we examine one
increasingly common instruction-following task: providing writing assistance to
compose a long-form answer. To evaluate the capabilities of current LLMs on
this task, we construct KIWI, a dataset of knowledge-intensive writing
instructions in the scientific domain. Given a research question, an initial
model-generated answer and a set of relevant papers, an expert annotator
iteratively issues instructions for the model to revise and improve its answer.
We collect 1,260 interaction turns from 234 interaction sessions with three
state-of-the-art LLMs. Each turn includes a user instruction, a model response,
and a human evaluation of the model response. Through a detailed analysis of
the collected responses, we find that all models struggle to incorporate new
information into an existing answer, and to perform precise and unambiguous
edits. Further, we find that models struggle to judge whether their outputs
successfully followed user instructions, with accuracy at least 10 points short
of human agreement. Our findings indicate that KIWI will be a valuable resource
to measure progress and improve LLMs' instruction-following capabilities for
knowledge intensive writing tasks.
| 2,024 | Computation and Language |
On the Origins of Linear Representations in Large Language Models | Recent works have argued that high-level semantic concepts are encoded
"linearly" in the representation space of large language models. In this work,
we study the origins of such linear representations. To that end, we introduce
a simple latent variable model to abstract and formalize the concept dynamics
of the next token prediction. We use this formalism to show that the next token
prediction objective (softmax with cross-entropy) and the implicit bias of
gradient descent together promote the linear representation of concepts.
Experiments show that linear representations emerge when learning from data
matching the latent variable model, confirming that this simple structure
already suffices to yield linear representations. We additionally confirm some
predictions of the theory using the LLaMA-2 large language model, giving
evidence that the simplified model yields generalizable insights.
| 2,024 | Computation and Language |
Learning to Decode Collaboratively with Multiple Language Models | We propose a method to teach multiple large language models (LLM) to
collaborate by interleaving their generations at the token level. We model the
decision of which LLM generates the next token as a latent variable. By
optimizing the marginal likelihood of a training set under our latent variable
model, the base LLM automatically learns when to generate itself and when to
call on one of the ``assistant'' language models to generate, all without
direct supervision. Token-level collaboration during decoding allows for a
fusion of each model's expertise in a manner tailored to the specific task at
hand. Our collaborative decoding is especially useful in cross-domain settings
where a generalist base LLM learns to invoke domain expert models. On
instruction-following, domain-specific QA, and reasoning tasks, we show that
the performance of the joint system exceeds that of the individual models.
Through qualitative analysis of the learned latent decisions, we show models
trained with our method exhibit several interesting collaboration patterns,
e.g., template-filling. Our code is available at
https://github.com/clinicalml/co-llm.
| 2,024 | Computation and Language |
Impoverished Language Technology: The Lack of (Social) Class in NLP | Since Labov's (1964) foundational work on the social stratification of
language, linguistics has dedicated concerted efforts towards understanding the
relationships between socio-demographic factors and language production and
perception. Despite the large body of evidence identifying significant
relationships between socio-demographic factors and language production,
relatively few of these factors have been investigated in the context of NLP
technology. While age and gender are well covered, Labov's initial target,
socio-economic class, is largely absent. We survey the existing Natural
Language Processing (NLP) literature and find that only 20 papers even mention
socio-economic status. However, the majority of those papers do not engage with
class beyond collecting information of annotator-demographics. Given this
research lacuna, we provide a definition of class that can be operationalised
by NLP researchers, and argue for including socio-economic class in future
language technologies.
| 2,024 | Computation and Language |
SaulLM-7B: A pioneering Large Language Model for Law | In this paper, we introduce SaulLM-7B, a large language model (LLM) tailored
for the legal domain. With 7 billion parameters, SaulLM-7B is the first LLM
designed explicitly for legal text comprehension and generation. Leveraging the
Mistral 7B architecture as its foundation, SaulLM-7B is trained on an English
legal corpus of over 30 billion tokens. SaulLM-7B exhibits state-of-the-art
proficiency in understanding and processing legal documents. Additionally, we
present a novel instructional fine-tuning method that leverages legal datasets
to further enhance SaulLM-7B's performance in legal tasks. SaulLM-7B is
released under the MIT License.
| 2,024 | Computation and Language |
FaaF: Facts as a Function for the evaluation of RAG systems | Factual recall from a reference source is crucial for evaluating the
performance of Retrieval Augmented Generation (RAG) systems, as it directly
probes into the quality of both retrieval and generation. However, it still
remains a challenge to perform this evaluation reliably and efficiently. Recent
work has focused on fact verification via prompting language model (LM)
evaluators, however we demonstrate that these methods are unreliable in the
presence of incomplete or inaccurate information. We introduce Facts as a
Function (FaaF), a new approach to fact verification that utilizes the function
calling abilities of LMs and a framework for RAG factual recall evaluation.
FaaF substantially improves the ability of LMs to identify unsupported facts in
text with incomplete information whilst improving efficiency and lowering cost
by several times, compared to prompt-based approaches.
| 2,024 | Computation and Language |
From One to Many: Expanding the Scope of Toxicity Mitigation in Language
Models | To date, toxicity mitigation in language models has almost entirely been
focused on single-language settings. As language models embrace multilingual
capabilities, it's crucial our safety measures keep pace. Recognizing this
research gap, our approach expands the scope of conventional toxicity
mitigation to address the complexities presented by multiple languages. In the
absence of sufficient annotated datasets across languages, we employ translated
data to evaluate and enhance our mitigation techniques. We also compare
finetuning mitigation approaches against retrieval-augmented techniques under
both static and continual toxicity mitigation scenarios. This allows us to
examine the effects of translation quality and the cross-lingual transfer on
toxicity mitigation. We also explore how model size and data quantity affect
the success of these mitigation efforts. Covering nine languages, our study
represents a broad array of linguistic families and levels of resource
availability, ranging from high to mid-resource languages. Through
comprehensive experiments, we provide insights into the complexities of
multilingual toxicity mitigation, offering valuable insights and paving the way
for future research in this increasingly important field. Code and data are
available at https://github.com/for-ai/goodtriever.
| 2,024 | Computation and Language |
A Measure for Transparent Comparison of Linguistic Diversity in
Multilingual NLP Data Sets | Typologically diverse benchmarks are increasingly created to track the
progress achieved in multilingual NLP. Linguistic diversity of these data sets
is typically measured as the number of languages or language families included
in the sample, but such measures do not consider structural properties of the
included languages. In this paper, we propose assessing linguistic diversity of
a data set against a reference language sample as a means of maximising
linguistic diversity in the long run. We represent languages as sets of
features and apply a version of the Jaccard index suitable for comparing sets
of measures. In addition to the features extracted from typological data bases,
we propose an automatic text-based measure, which can be used as a means of
overcoming the well-known problem of data sparsity in manually collected
features. Our diversity score is interpretable in terms of linguistic features
and can identify the types of languages that are not represented in a data set.
Using our method, we analyse a range of popular multilingual data sets (UD,
Bible100, mBERT, XTREME, XGLUE, XNLI, XCOPA, TyDiQA, XQuAD). In addition to
ranking these data sets, we find, for example, that (poly)synthetic languages
are missing in almost all of them.
| 2,024 | Computation and Language |
Did Translation Models Get More Robust Without Anyone Even Noticing? | Neural machine translation (MT) models achieve strong results across a
variety of settings, but it is widely believed that they are highly sensitive
to "noisy" inputs, such as spelling errors, abbreviations, and other formatting
issues. In this paper, we revisit this insight in light of recent multilingual
MT models and large language models (LLMs) applied to machine translation.
Somewhat surprisingly, we show through controlled experiments that these models
are far more robust to many kinds of noise than previous models, even when they
perform similarly on clean data. This is notable because, even though LLMs have
more parameters and more complex training processes than past models, none of
the open ones we consider use any techniques specifically designed to encourage
robustness. Next, we show that similar trends hold for social media translation
experiments -- LLMs are more robust to social media text. We include an
analysis of the circumstances in which source correction techniques can be used
to mitigate the effects of noise. Altogether, we show that robustness to many
types of noise has increased.
| 2,024 | Computation and Language |
The Heuristic Core: Understanding Subnetwork Generalization in
Pretrained Language Models | Prior work has found that pretrained language models (LMs) fine-tuned with
different random seeds can achieve similar in-domain performance but generalize
differently on tests of syntactic generalization. In this work, we show that,
even within a single model, we can find multiple subnetworks that perform
similarly in-domain, but generalize vastly differently. To better understand
these phenomena, we investigate if they can be understood in terms of
"competing subnetworks": the model initially represents a variety of distinct
algorithms, corresponding to different subnetworks, and generalization occurs
when it ultimately converges to one. This explanation has been used to account
for generalization in simple algorithmic tasks. Instead of finding competing
subnetworks, we find that all subnetworks -- whether they generalize or not --
share a set of attention heads, which we refer to as the heuristic core.
Further analysis suggests that these attention heads emerge early in training
and compute shallow, non-generalizing features. The model learns to generalize
by incorporating additional attention heads, which depend on the outputs of the
"heuristic" heads to compute higher-level features. Overall, our results offer
a more detailed picture of the mechanisms for syntactic generalization in
pretrained LMs.
| 2,024 | Computation and Language |
Can Large Language Models do Analytical Reasoning? | This paper explores the cutting-edge Large Language Model with analytical
reasoning on sports. Our analytical reasoning embodies the tasks of letting
large language models count how many points each team scores in a quarter in
the NBA and NFL games. Our major discoveries are in two folds. Firstly, we find
among all the models we employed, GPT-4 stands out in effectiveness, followed
by Claude-2.1, with GPT-3.5, Gemini-Pro, and Llama-2-70b lagging behind.
Specifically, we compare three different prompting techniques and a
divide-and-conquer approach, we find that the latter was the most effective.
Our divide-and-conquer approach breaks down play-by-play data into smaller,
more manageable segments, solves each piece individually, and then aggregates
them together. Besides the divide-and-conquer approach, we also explore the
Chain of Thought (CoT) strategy, which markedly improves outcomes for certain
models, notably GPT-4 and Claude-2.1, with their accuracy rates increasing
significantly. However, the CoT strategy has negligible or even detrimental
effects on the performance of other models like GPT-3.5 and Gemini-Pro.
Secondly, to our surprise, we observe that most models, including GPT-4,
struggle to accurately count the total scores for NBA quarters despite showing
strong performance in counting NFL quarter scores. This leads us to further
investigate the factors that impact the complexity of analytical reasoning
tasks with extensive experiments, through which we conclude that task
complexity depends on the length of context, the information density, and the
presence of related information. Our research provides valuable insights into
the complexity of analytical reasoning tasks and potential directions for
developing future large language models.
| 2,024 | Computation and Language |
Semi-Supervised Dialogue Abstractive Summarization via High-Quality
Pseudolabel Selection | Semi-supervised dialogue summarization (SSDS) leverages model-generated
summaries to reduce reliance on human-labeled data and improve the performance
of summarization models. While addressing label noise, previous works on
semi-supervised learning primarily focus on natural language understanding
tasks, assuming each sample has a unique label. However, these methods are not
directly applicable to SSDS, as it is a generative task, and each dialogue can
be summarized in different ways. In this work, we propose a novel scoring
approach, SiCF, which encapsulates three primary dimensions of summarization
model quality: Semantic invariance (indicative of model confidence), Coverage
(factual recall), and Faithfulness (factual precision). Using the SiCF score,
we select unlabeled dialogues with high-quality generated summaries to train
summarization models. Comprehensive experiments on three public datasets
demonstrate the effectiveness of SiCF scores in uncertainty estimation and
semi-supervised learning for dialogue summarization tasks. Our code is
available at \url{https://github.com/amazon-science/summarization-sicf-score}.
| 2,024 | Computation and Language |
Transformers and Language Models in Form Understanding: A Comprehensive
Review of Scanned Document Analysis | This paper presents a comprehensive survey of research works on the topic of
form understanding in the context of scanned documents. We delve into recent
advancements and breakthroughs in the field, highlighting the significance of
language models and transformers in solving this challenging task. Our research
methodology involves an in-depth analysis of popular documents and forms of
understanding of trends over the last decade, enabling us to offer valuable
insights into the evolution of this domain. Focusing on cutting-edge models, we
showcase how transformers have propelled the field forward, revolutionizing
form-understanding techniques. Our exploration includes an extensive
examination of state-of-the-art language models designed to effectively tackle
the complexities of noisy scanned documents. Furthermore, we present an
overview of the latest and most relevant datasets, which serve as essential
benchmarks for evaluating the performance of selected models. By comparing and
contrasting the capabilities of these models, we aim to provide researchers and
practitioners with useful guidance in choosing the most suitable solutions for
their specific form understanding tasks.
| 2,024 | Computation and Language |
Don't Blame the Data, Blame the Model: Understanding Noise and Bias When
Learning from Subjective Annotations | Researchers have raised awareness about the harms of aggregating labels
especially in subjective tasks that naturally contain disagreements among human
annotators. In this work we show that models that are only provided aggregated
labels show low confidence on high-disagreement data instances. While previous
studies consider such instances as mislabeled, we argue that the reason the
high-disagreement text instances have been hard-to-learn is that the
conventional aggregated models underperform in extracting useful signals from
subjective tasks. Inspired by recent studies demonstrating the effectiveness of
learning from raw annotations, we investigate classifying using Multiple Ground
Truth (Multi-GT) approaches. Our experiments show an improvement of confidence
for the high-disagreement instances.
| 2,024 | Computation and Language |
DA-Net: A Disentangled and Adaptive Network for Multi-Source
Cross-Lingual Transfer Learning | Multi-Source cross-lingual transfer learning deals with the transfer of task
knowledge from multiple labelled source languages to an unlabeled target
language under the language shift. Existing methods typically focus on
weighting the predictions produced by language-specific classifiers of
different sources that follow a shared encoder. However, all source languages
share the same encoder, which is updated by all these languages. The extracted
representations inevitably contain different source languages' information,
which may disturb the learning of the language-specific classifiers.
Additionally, due to the language gap, language-specific classifiers trained
with source labels are unable to make accurate predictions for the target
language. Both facts impair the model's performance. To address these
challenges, we propose a Disentangled and Adaptive Network (DA-Net). Firstly,
we devise a feedback-guided collaborative disentanglement method that seeks to
purify input representations of classifiers, thereby mitigating mutual
interference from multiple sources. Secondly, we propose a class-aware parallel
adaptation method that aligns class-level distributions for each source-target
language pair, thereby alleviating the language pairs' language gap.
Experimental results on three different tasks involving 38 languages validate
the effectiveness of our approach.
| 2,024 | Computation and Language |
Attempt Towards Stress Transfer in Speech-to-Speech Machine Translation | The language diversity in India's education sector poses a significant
challenge, hindering inclusivity. Despite the democratization of knowledge
through online educational content, the dominance of English, as the internet's
lingua franca, limits accessibility, emphasizing the crucial need for
translation into Indian languages. Despite existing Speech-to-Speech Machine
Translation (SSMT) technologies, the lack of intonation in these systems gives
monotonous translations, leading to a loss of audience interest and
disengagement from the content. To address this, our paper introduces a dataset
with stress annotations in Indian English and also a Text-to-Speech (TTS)
architecture capable of incorporating stress into synthesized speech. This
dataset is used for training a stress detection model, which is then used in
the SSMT system for detecting stress in the source speech and transferring it
into the target language speech. The TTS architecture is based on FastPitch and
can modify the variances based on stressed words given. We present an Indian
English-to-Hindi SSMT system that can transfer stress and aim to enhance the
overall quality and engagement of educational content.
| 2,024 | Computation and Language |
Metric-aware LLM inference | Large language models (LLMs) have demonstrated strong results on a range of
NLP tasks. Typically, outputs are obtained via autoregressive sampling from the
LLM's underlying distribution. We show that this inference strategy can be
suboptimal for a range of tasks and associated evaluation metrics. As a remedy,
we propose metric aware LLM inference: a decision theoretic approach optimizing
for custom metrics at inference time. We report improvements over baselines on
academic benchmarks and publicly available models.
| 2,024 | Computation and Language |
Large Language Models are In-Context Molecule Learners | Large Language Models (LLMs) have demonstrated exceptional performance in
biochemical tasks, especially the molecule caption translation task, which aims
to bridge the gap between molecules and natural language texts. However,
previous methods in adapting LLMs to the molecule-caption translation task
required extra domain-specific pre-training stages, suffered weak alignment
between molecular and textual spaces, or imposed stringent demands on the scale
of LLMs. To resolve the challenges, we propose In-Context Molecule Adaptation
(ICMA), as a new paradigm allowing LLMs to learn the molecule-text alignment
from context examples via In-Context Molecule Tuning. Specifically, ICMA
incorporates the following three stages: Cross-modal Retrieval, Post-retrieval
Re-ranking, and In-context Molecule Tuning. Initially, Cross-modal Retrieval
utilizes BM25 Caption Retrieval and Molecule Graph Retrieval to retrieve
informative context examples. Additionally, we also propose Post-retrieval
Re-ranking with Sequence Reversal and Random Walk to further improve the
quality of retrieval results. Finally, In-Context Molecule Tuning unlocks the
in-context molecule learning capability of LLMs with retrieved examples and
adapts the parameters of LLMs for the molecule-caption translation task.
Experimental results demonstrate that ICMT can empower LLMs to achieve
state-of-the-art or comparable performance without extra training corpora and
intricate structures, showing that LLMs are inherently in-context molecule
learners.
| 2,024 | Computation and Language |
Persona Extraction Through Semantic Similarity for Emotional Support
Conversation Generation | Providing emotional support through dialogue systems is becoming increasingly
important in today's world, as it can support both mental health and social
interactions in many conversation scenarios. Previous works have shown that
using persona is effective for generating empathetic and supportive responses.
They have often relied on pre-provided persona rather than inferring them
during conversations. However, it is not always possible to obtain a user
persona before the conversation begins. To address this challenge, we propose
PESS (Persona Extraction through Semantic Similarity), a novel framework that
can automatically infer informative and consistent persona from dialogues. We
devise completeness loss and consistency loss based on semantic similarity
scores. The completeness loss encourages the model to generate missing persona
information, and the consistency loss guides the model to distinguish between
consistent and inconsistent persona. Our experimental results demonstrate that
high-quality persona information inferred by PESS is effective in generating
emotionally supportive responses.
| 2,024 | Computation and Language |
Self-Evaluation of Large Language Model based on Glass-box Features | The proliferation of open-source Large Language Models (LLMs) underscores the
pressing need for evaluation methods. Existing works primarily rely on external
evaluators, focusing on training and prompting strategies. However, a crucial
aspect - model-aware glass-box features - is overlooked. In this study, we
explore the utility of glass-box features under the scenario of
self-evaluation, namely applying an LLM to evaluate its own output. We
investigate various glass-box feature groups and discovered that the softmax
distribution serves as a reliable indicator for quality evaluation.
Furthermore, we propose two strategies to enhance the evaluation by
incorporating features derived from references. Experimental results on public
benchmarks validate the feasibility of self-evaluation of LLMs using glass-box
features.
| 2,024 | Computation and Language |
Aligners: Decoupling LLMs and Alignment | Large Language Models (LLMs) need to be aligned with human expectations to
ensure their safety and utility in most applications. Alignment is challenging,
costly, and needs to be repeated for every LLM and alignment criterion. We
propose to decouple LLMs and alignment by training aligner models that can be
used to align any LLM for a given criteria on an as-needed basis, thus also
reducing the potential negative impacts of alignment on performance. Our recipe
for training the aligner models solely relies on synthetic data generated with
a (prompted) LLM and can be easily adjusted for a variety of alignment
criteria. We illustrate our method by training an "ethical" aligner and verify
its efficacy empirically.
| 2,024 | Computation and Language |
DEEP-ICL: Definition-Enriched Experts for Language Model In-Context
Learning | It has long been assumed that the sheer number of parameters in large
language models (LLMs) drives in-context learning (ICL) capabilities, enabling
remarkable performance improvements by leveraging task-specific demonstrations.
Challenging this hypothesis, we introduce DEEP-ICL, a novel task Definition
Enriched ExPert Ensembling methodology for ICL. DEEP-ICL explicitly extracts
task definitions from given demonstrations and generates responses through
learning task-specific examples. We argue that improvement from ICL does not
directly rely on model size, but essentially stems from understanding task
definitions and task-guided learning. Inspired by this, DEEP-ICL combines two
3B models with distinct roles (one for concluding task definitions and the
other for learning task demonstrations) and achieves comparable performance to
LLaMA2-13B. Furthermore, our framework outperforms conventional ICL by
overcoming pretraining sequence length limitations, by supporting unlimited
demonstrations. We contend that DEEP-ICL presents a novel alternative for
achieving efficient few-shot learning, extending beyond the conventional ICL.
| 2,024 | Computation and Language |
UltraWiki: Ultra-fine-grained Entity Set Expansion with Negative Seed
Entities | Entity Set Expansion (ESE) aims to identify new entities belonging to the
same semantic class as a given set of seed entities. Traditional methods
primarily relied on positive seed entities to represent a target semantic
class, which poses challenge for the representation of ultra-fine-grained
semantic classes. Ultra-fine-grained semantic classes are defined based on
fine-grained semantic classes with more specific attribute constraints.
Describing it with positive seed entities alone cause two issues: (i) Ambiguity
among ultra-fine-grained semantic classes. (ii) Inability to define "unwanted"
semantic. Due to these inherent shortcomings, previous methods struggle to
address the ultra-fine-grained ESE (Ultra-ESE). To solve this issue, we first
introduce negative seed entities in the inputs, which belong to the same
fine-grained semantic class as the positive seed entities but differ in certain
attributes. Negative seed entities eliminate the semantic ambiguity by contrast
between positive and negative attributes. Meanwhile, it provide a
straightforward way to express "unwanted". To assess model performance in
Ultra-ESE, we constructed UltraWiki, the first large-scale dataset tailored for
Ultra-ESE. UltraWiki encompasses 236 ultra-fine-grained semantic classes, where
each query of them is represented with 3-5 positive and negative seed entities.
A retrieval-based framework RetExpan and a generation-based framework GenExpan
are proposed to comprehensively assess the efficacy of large language models
from two different paradigms in Ultra-ESE. Moreover, we devised three
strategies to enhance models' comprehension of ultra-fine-grained entities
semantics: contrastive learning, retrieval augmentation, and chain-of-thought
reasoning. Extensive experiments confirm the effectiveness of our proposed
strategies and also reveal that there remains a large space for improvement in
Ultra-ESE.
| 2,024 | Computation and Language |
Proxy-RLHF: Decoupling Generation and Alignment in Large Language Model
with Proxy | Reinforcement Learning from Human Feedback (RLHF) is the prevailing approach
to ensure Large Language Models (LLMs) align with human values. However,
existing RLHF methods require a high computational cost, one main reason being
that RLHF assigns both the generation and alignment tasks to the LLM
simultaneously. In this paper, we introduce Proxy-RLHF, which decouples the
generation and alignment processes of LLMs, achieving alignment with human
values at a much lower computational cost. We start with a novel Markov
Decision Process (MDP) designed for the alignment process and employ
Reinforcement Learning (RL) to train a streamlined proxy model that oversees
the token generation of the LLM, without altering the LLM itself. Experiments
show that our method achieves a comparable level of alignment with only 1\% of
the training parameters of other methods.
| 2,024 | Computation and Language |
HaluEval-Wild: Evaluating Hallucinations of Language Models in the Wild | Hallucinations pose a significant challenge to the reliability of large
language models (LLMs) in critical domains. Recent benchmarks designed to
assess LLM hallucinations within conventional NLP tasks, such as
knowledge-intensive question answering (QA) and summarization, are insufficient
for capturing the complexities of user-LLM interactions in dynamic, real-world
settings. To address this gap, we introduce HaluEval-Wild, the first benchmark
specifically designed to evaluate LLM hallucinations in the wild. We
meticulously collect challenging (adversarially filtered by Alpaca) user
queries from existing real-world user-LLM interaction datasets, including
ShareGPT, to evaluate the hallucination rates of various LLMs. Upon analyzing
the collected queries, we categorize them into five distinct types, which
enables a fine-grained analysis of the types of hallucinations LLMs exhibit,
and synthesize the reference answers with the powerful GPT-4 model and
retrieval-augmented generation (RAG). Our benchmark offers a novel approach
towards enhancing our comprehension and improvement of LLM reliability in
scenarios reflective of real-world interactions.
| 2,024 | Computation and Language |
Can Your Model Tell a Negation from an Implicature? Unravelling
Challenges With Intent Encoders | Conversational systems often rely on embedding models for intent
classification and intent clustering tasks. The advent of Large Language Models
(LLMs), which enable instructional embeddings allowing one to adjust semantics
over the embedding space using prompts, are being viewed as a panacea for these
downstream conversational tasks. However, traditional evaluation benchmarks
rely solely on task metrics that don't particularly measure gaps related to
semantic understanding. Thus, we propose an intent semantic toolkit that gives
a more holistic view of intent embedding models by considering three tasks --
(1) intent classification, (2) intent clustering, and (3) a novel triplet task.
The triplet task gauges the model's understanding of two semantic concepts
paramount in real-world conversational systems -- negation and implicature. We
observe that current embedding models fare poorly in semantic understanding of
these concepts. To address this, we propose a pre-training approach to improve
the embedding model by leveraging augmentation with data generated by an
auto-regressive model and a contrastive loss term. Our approach improves the
semantic understanding of the intent embedding model on the aforementioned
linguistic dimensions while slightly effecting their performance on downstream
task metrics.
| 2,024 | Computation and Language |
Measuring Meaning Composition in the Human Brain with Composition Scores
from Large Language Models | The process of meaning composition, wherein smaller units like morphemes or
words combine to form the meaning of phrases and sentences, is essential for
human sentence comprehension. Despite extensive neurolinguistic research into
the brain regions involved in meaning composition, a computational metric to
quantify the extent of composition is still lacking. Drawing on the key-value
memory interpretation of transformer feed-forward network blocks, we introduce
the Composition Score, a novel model-based metric designed to quantify the
degree of meaning composition during sentence comprehension. Experimental
findings show that this metric correlates with brain clusters associated with
word frequency, structural processing, and general sensitivity to words,
suggesting the multifaceted nature of meaning composition during human sentence
comprehension.
| 2,024 | Computation and Language |
Computational Modelling of Plurality and Definiteness in Chinese Noun
Phrases | Theoretical linguists have suggested that some languages (e.g., Chinese and
Japanese) are "cooler" than other languages based on the observation that the
intended meaning of phrases in these languages depends more on their contexts.
As a result, many expressions in these languages are shortened, and their
meaning is inferred from the context. In this paper, we focus on the omission
of the plurality and definiteness markers in Chinese noun phrases (NPs) to
investigate the predictability of their intended meaning given the contexts. To
this end, we built a corpus of Chinese NPs, each of which is accompanied by its
corresponding context, and by labels indicating its singularity/plurality and
definiteness/indefiniteness. We carried out corpus assessments and analyses.
The results suggest that Chinese speakers indeed drop plurality and
definiteness markers very frequently. Building on the corpus, we train a bank
of computational models using both classic machine learning models and
state-of-the-art pre-trained language models to predict the plurality and
definiteness of each NP. We report on the performance of these models and
analyse their behaviours.
| 2,024 | Computation and Language |
Acceleron: A Tool to Accelerate Research Ideation | Several tools have recently been proposed for assisting researchers during
various stages of the research life-cycle. However, these primarily concentrate
on tasks such as retrieving and recommending relevant literature, reviewing and
critiquing the draft, and writing of research manuscripts. Our investigation
reveals a significant gap in availability of tools specifically designed to
assist researchers during the challenging ideation phase of the research
life-cycle. To aid with research ideation, we propose `Acceleron', a research
accelerator for different phases of the research life cycle, and which is
specially designed to aid the ideation process. Acceleron guides researchers
through the formulation of a comprehensive research proposal, encompassing a
novel research problem. The proposals motivation is validated for novelty by
identifying gaps in the existing literature and suggesting a plausible list of
techniques to solve the proposed problem. We leverage the reasoning and
domain-specific skills of Large Language Models (LLMs) to create an agent-based
architecture incorporating colleague and mentor personas for LLMs. The LLM
agents emulate the ideation process undertaken by researchers, engaging
researchers in an interactive fashion to aid in the development of the research
proposal. Notably, our tool addresses challenges inherent in LLMs, such as
hallucinations, implements a two-stage aspect-based retrieval to manage
precision-recall trade-offs, and tackles issues of unanswerability. As
evaluation, we illustrate the execution of our motivation validation and method
synthesis workflows on proposals from the ML and NLP domain, given by 3
distinct researchers. Our observations and evaluations provided by the
researchers illustrate the efficacy of the tool in terms of assisting
researchers with appropriate inputs at distinct stages and thus leading to
improved time efficiency.
| 2,024 | Computation and Language |
Exploring Continual Learning of Compositional Generalization in NLI | Compositional Natural Language Inference has been explored to assess the true
abilities of neural models to perform NLI. Yet, current evaluations assume
models to have full access to all primitive inferences in advance, in contrast
to humans that continuously acquire inference knowledge. In this paper, we
introduce the Continual Compositional Generalization in Inference (C2Gen NLI)
challenge, where a model continuously acquires knowledge of constituting
primitive inference tasks as a basis for compositional inferences. We explore
how continual learning affects compositional generalization in NLI, by
designing a continual learning setup for compositional NLI inference tasks. Our
experiments demonstrate that models fail to compositionally generalize in a
continual scenario. To address this problem, we first benchmark various
continual learning algorithms and verify their efficacy. We then further
analyze C2Gen, focusing on how to order primitives and compositional inference
types and examining correlations between subtasks. Our analyses show that by
learning subtasks continuously while observing their dependencies and
increasing degrees of difficulty, continual learning can enhance composition
generalization ability.
| 2,024 | Computation and Language |
Promising and worth-to-try future directions for advancing
state-of-the-art surrogates methods of agent-based models in social and
health computational sciences | The execution and runtime performance of model-based analysis tools for
realistic large-scale ABMs (Agent-Based Models) can be excessively long. This
due to the computational demand exponentially proportional to the model size
(e.g. Population size) and the number of model parameters. Even the runtime of
a single simulation of a realistic ABM may demand huge computational resources
when attempting to employ realistic population size. The main aim of this
ad-hoc brief report is to highlight some of surrogate models that were adequate
and computationally less demanding for nonlinear dynamical models in various
modeling application areas.To the author knowledge, these methods have been
not, at least extensively, employed for ABMs within the field of (SHCS) Social
Health Computational Sciences, yet. Thus, they might be, but not necessarily,
useful in progressing state of the art for establishing surrogate models for
ABMs in the field of SHCS.
| 2,024 | Computation and Language |
Classist Tools: Social Class Correlates with Performance in NLP | Since the foundational work of William Labov on the social stratification of
language (Labov, 1964), linguistics has made concentrated efforts to explore
the links between sociodemographic characteristics and language production and
perception. But while there is strong evidence for socio-demographic
characteristics in language, they are infrequently used in Natural Language
Processing (NLP). Age and gender are somewhat well represented, but Labov's
original target, socioeconomic status, is noticeably absent. And yet it
matters. We show empirically that NLP disadvantages less-privileged
socioeconomic groups. We annotate a corpus of 95K utterances from movies with
social class, ethnicity and geographical language variety and measure the
performance of NLP systems on three tasks: language modelling, automatic speech
recognition, and grammar error correction. We find significant performance
disparities that can be attributed to socioeconomic status as well as ethnicity
and geographical differences. With NLP technologies becoming ever more
ubiquitous and quotidian, they must accommodate all language varieties to avoid
disadvantaging already marginalised groups. We argue for the inclusion of
socioeconomic class in future language technologies.
| 2,024 | Computation and Language |
Low-Resource Court Judgment Summarization for Common Law Systems | Common law courts need to refer to similar precedents' judgments to inform
their current decisions. Generating high-quality summaries of court judgment
documents can facilitate legal practitioners to efficiently review previous
cases and assist the general public in accessing how the courts operate and how
the law is applied. Previous court judgment summarization research focuses on
civil law or a particular jurisdiction's judgments. However, judges can refer
to the judgments from all common law jurisdictions. Current summarization
datasets are insufficient to satisfy the demands of summarizing precedents
across multiple jurisdictions, especially when labeled data are scarce for many
jurisdictions. To address the lack of datasets, we present CLSum, the first
dataset for summarizing multi-jurisdictional common law court judgment
documents. Besides, this is the first court judgment summarization work
adopting large language models (LLMs) in data augmentation, summary generation,
and evaluation. Specifically, we design an LLM-based data augmentation method
incorporating legal knowledge. We also propose a legal knowledge enhanced
evaluation metric based on LLM to assess the quality of generated judgment
summaries. Our experimental results verify that the LLM-based summarization
methods can perform well in the few-shot and zero-shot settings. Our LLM-based
data augmentation method can mitigate the impact of low data resources.
Furthermore, we carry out comprehensive comparative experiments to find
essential model components and settings that are capable of enhancing
summarization performance.
| 2,024 | Computation and Language |
Pearl: A Review-driven Persona-Knowledge Grounded Conversational
Recommendation Dataset | Conversational recommender system is an emerging area that has garnered an
increasing interest in the community, especially with the advancements in large
language models (LLMs) that enable diverse reasoning over conversational input.
Despite the progress, the field has many aspects left to explore. The currently
available public datasets for conversational recommendation lack specific user
preferences and explanations for recommendations, hindering high-quality
recommendations. To address such challenges, we present a novel conversational
recommendation dataset named PEARL, synthesized with persona- and
knowledge-augmented LLM simulators. We obtain detailed persona and knowledge
from real-world reviews and construct a large-scale dataset with over 57k
dialogues. Our experimental results demonstrate that utterances in PEARL
include more specific user preferences, show expertise in the target domain,
and provide recommendations more relevant to the dialogue context than those in
prior datasets.
| 2,024 | Computation and Language |
Do Large Language Model Understand Multi-Intent Spoken Language ? | This study marks a significant advancement by harnessing Large Language
Models (LLMs) for multi-intent spoken language understanding (SLU), proposing a
unique methodology that capitalizes on the generative power of LLMs within an
SLU context. Our innovative technique reconfigures entity slots specifically
for LLM application in multi-intent SLU environments and introduces the concept
of Sub-Intent Instruction (SII), enhancing the dissection and interpretation of
intricate, multi-intent communication within varied domains. The resultant
datasets, dubbed LM-MixATIS and LM-MixSNIPS, are crafted from pre-existing
benchmarks. Our research illustrates that LLMs can match and potentially excel
beyond the capabilities of current state-of-the-art multi-intent SLU models. It
further explores LLM efficacy across various intent configurations and dataset
proportions. Moreover, we introduce two pioneering metrics, Entity Slot
Accuracy (ESA) and Combined Semantic Accuracy (CSA), to provide an in-depth
analysis of LLM proficiency in this complex field.
| 2,024 | Computation and Language |
NLPre: a revised approach towards language-centric benchmarking of
Natural Language Preprocessing systems | With the advancements of transformer-based architectures, we observe the rise
of natural language preprocessing (NLPre) tools capable of solving preliminary
NLP tasks (e.g. tokenisation, part-of-speech tagging, dependency parsing, or
morphological analysis) without any external linguistic guidance. It is arduous
to compare novel solutions to well-entrenched preprocessing toolkits, relying
on rule-based morphological analysers or dictionaries. Aware of the
shortcomings of existing NLPre evaluation approaches, we investigate a novel
method of reliable and fair evaluation and performance reporting. Inspired by
the GLUE benchmark, the proposed language-centric benchmarking system enables
comprehensive ongoing evaluation of multiple NLPre tools, while credibly
tracking their performance. The prototype application is configured for Polish
and integrated with the thoroughly assembled NLPre-PL benchmark. Based on this
benchmark, we conduct an extensive evaluation of a variety of Polish NLPre
systems. To facilitate the construction of benchmarking environments for other
languages, e.g. NLPre-GA for Irish or NLPre-ZH for Chinese, we ensure full
customization of the publicly released source code of the benchmarking system.
The links to all the resources (deployed platforms, source code, trained
models, datasets etc.) can be found on the project website:
https://sites.google.com/view/nlpre-benchmark.
| 2,024 | Computation and Language |
Where does In-context Translation Happen in Large Language Models | Self-supervised large language models have demonstrated the ability to
perform Machine Translation (MT) via in-context learning, but little is known
about where the model performs the task with respect to prompt instructions and
demonstration examples. In this work, we attempt to characterize the region
where large language models transition from in-context learners to translation
models. Through a series of layer-wise context-masking experiments on
\textsc{GPTNeo2.7B}, \textsc{Bloom3B}, \textsc{Llama7b} and
\textsc{Llama7b-chat}, we demonstrate evidence of a "task recognition" point
where the translation task is encoded into the input representations and
attention to context is no longer necessary. We further observe correspondence
between the low performance when masking out entire layers, and the task
recognition layers. Taking advantage of this redundancy results in 45\%
computational savings when prompting with 5 examples, and task recognition
achieved at layer 14 / 32. Our layer-wise fine-tuning experiments indicate that
the most effective layers for MT fine-tuning are the layers critical to task
recognition.
| 2,024 | Computation and Language |
Uncertainty-Aware Relational Graph Neural Network for Few-Shot Knowledge
Graph Completion | Few-shot knowledge graph completion (FKGC) aims to query the unseen facts of
a relation given its few-shot reference entity pairs. The side effect of noises
due to the uncertainty of entities and triples may limit the few-shot learning,
but existing FKGC works neglect such uncertainty, which leads them more
susceptible to limited reference samples with noises. In this paper, we propose
a novel uncertainty-aware few-shot KG completion framework (UFKGC) to model
uncertainty for a better understanding of the limited data by learning
representations under Gaussian distribution. Uncertainty representation is
first designed for estimating the uncertainty scope of the entity pairs after
transferring feature representations into a Gaussian distribution. Further, to
better integrate the neighbors with uncertainty characteristics for entity
features, we design an uncertainty-aware relational graph neural network
(UR-GNN) to conduct convolution operations between the Gaussian distributions.
Then, multiple random samplings are conducted for reference triples within the
Gaussian distribution to generate smooth reference representations during the
optimization. The final completion score for each query instance is measured by
the designed uncertainty optimization to make our approach more robust to the
noises in few-shot scenarios. Experimental results show that our approach
achieves excellent performance on two benchmark datasets compared to its
competitors.
| 2,024 | Computation and Language |
MaCmS: Magahi Code-mixed Dataset for Sentiment Analysis | The present paper introduces new sentiment data, MaCMS, for
Magahi-Hindi-English (MHE) code-mixed language, where Magahi is a
less-resourced minority language. This dataset is the first
Magahi-Hindi-English code-mixed dataset for sentiment analysis tasks. Further,
we also provide a linguistics analysis of the dataset to understand the
structure of code-mixing and a statistical study to understand the language
preferences of speakers with different polarities. With these analyses, we also
train baseline models to evaluate the dataset's quality.
| 2,024 | Computation and Language |
QAQ: Quality Adaptive Quantization for LLM KV Cache | The emergence of LLMs has ignited a fresh surge of breakthroughs in NLP
applications, particularly in domains such as question-answering systems and
text generation. As the need for longer context grows, a significant bottleneck
in model deployment emerges due to the linear expansion of the Key-Value (KV)
cache with the context length. Existing methods primarily rely on various
hypotheses, such as sorting the KV cache based on attention scores for
replacement or eviction, to compress the KV cache and improve model throughput.
However, heuristics used by these strategies may wrongly evict essential KV
cache, which can significantly degrade model performance. In this paper, we
propose QAQ, a Quality Adaptive Quantization scheme for the KV cache. We
theoretically demonstrate that key cache and value cache exhibit distinct
sensitivities to quantization, leading to the formulation of separate
quantization strategies for their non-uniform quantization. Through the
integration of dedicated outlier handling, as well as an improved
attention-aware approach, QAQ achieves up to 10x the compression ratio of the
KV cache size with a neglectable impact on model performance. QAQ significantly
reduces the practical hurdles of deploying LLMs, opening up new possibilities
for longer-context applications. The code is available at
github.com/ClubieDong/KVCacheQuantization.
| 2,024 | Computation and Language |
Yi: Open Foundation Models by 01.AI | We introduce the Yi model family, a series of language and multimodal models
that demonstrate strong multi-dimensional capabilities. The Yi model family is
based on 6B and 34B pretrained language models, then we extend them to chat
models, 200K long context models, depth-upscaled models, and vision-language
models. Our base models achieve strong performance on a wide range of
benchmarks like MMLU, and our finetuned chat models deliver strong human
preference rate on major evaluation platforms like AlpacaEval and Chatbot
Arena. Building upon our scalable super-computing infrastructure and the
classical transformer architecture, we attribute the performance of Yi models
primarily to its data quality resulting from our data-engineering efforts. For
pretraining, we construct 3.1 trillion tokens of English and Chinese corpora
using a cascaded data deduplication and quality filtering pipeline. For
finetuning, we polish a small scale (less than 10K) instruction dataset over
multiple iterations such that every single instance has been verified directly
by our machine learning engineers. For vision-language, we combine the chat
language model with a vision transformer encoder and train the model to align
visual representations to the semantic space of the language model. We further
extend the context length to 200K through lightweight continual pretraining and
demonstrate strong needle-in-a-haystack retrieval performance. We show that
extending the depth of the pretrained checkpoint through continual pretraining
further improves performance. We believe that given our current results,
continuing to scale up model parameters using thoroughly optimized data will
lead to even stronger frontier models.
| 2,024 | Computation and Language |
Chain of Thought Explanation for Dialogue State Tracking | Dialogue state tracking (DST) aims to record user queries and goals during a
conversational interaction achieved by maintaining a predefined set of slots
and their corresponding values. Current approaches decide slot values opaquely,
while humans usually adopt a more deliberate approach by collecting information
from relevant dialogue turns and then reasoning the appropriate values. In this
work, we focus on the steps needed to figure out slot values by proposing a
model named Chain-of-Thought-Explanation (CoTE) for the DST task. CoTE, which
is built on the generative DST framework, is designed to create detailed
explanations step by step after determining the slot values. This process leads
to more accurate and reliable slot values. More-over, to improve the reasoning
ability of the CoTE, we further construct more fluent and high-quality
explanations with automatic paraphrasing, leading the method CoTE-refined.
Experimental results on three widely recognized DST benchmarks-MultiWOZ 2.2,
WoZ 2.0, and M2M-demonstrate the remarkable effectiveness of the CoTE.
Furthermore, through a meticulous fine-grained analysis, we observe significant
benefits of our CoTE on samples characterized by longer dialogue turns, user
responses, and reasoning steps.
| 2,024 | Computation and Language |
Telecom Language Models: Must They Be Large? | The increasing interest in Large Language Models (LLMs) within the
telecommunications sector underscores their potential to revolutionize
operational efficiency. However, the deployment of these sophisticated models
is often hampered by their substantial size and computational demands, raising
concerns about their viability in resource-constrained environments. Addressing
this challenge, recent advancements have seen the emergence of small language
models that surprisingly exhibit performance comparable to their larger
counterparts in many tasks, such as coding and common-sense reasoning. Phi-2, a
compact yet powerful model, exemplifies this new wave of efficient small
language models. This paper conducts a comprehensive evaluation of Phi-2's
intrinsic understanding of the telecommunications domain. Recognizing the
scale-related limitations, we enhance Phi-2's capabilities through a
Retrieval-Augmented Generation approach, meticulously integrating an extensive
knowledge base specifically curated with telecom standard specifications. The
enhanced Phi-2 model demonstrates a profound improvement in accuracy, answering
questions about telecom standards with a precision that closely rivals the more
resource-intensive GPT-3.5. The paper further explores the refined capabilities
of Phi-2 in addressing problem-solving scenarios within the telecom sector,
highlighting its potential and limitations.
| 2,024 | Computation and Language |
Greater than the sum of its parts: The role of minority and majority
status in collaborative problem-solving communication | Collaborative problem-solving (CPS) is a vital skill used both in the
workplace and in educational environments. CPS is useful in tackling
increasingly complex global, economic, and political issues and is considered a
central 21st century skill. The increasingly connected global community
presents a fruitful opportunity for creative and collaborative problem-solving
interactions and solutions that involve diverse perspectives. Unfortunately,
women and underrepresented minorities (URMs) often face obstacles during
collaborative interactions that hinder their key participation in these
problem-solving conversations. Here, we explored the communication patterns of
minority and non-minority individuals working together in a CPS task. Group
Communication Analysis (GCA), a temporally-sensitive computational linguistic
tool, was used to examine how URM status impacts individuals' sociocognitive
linguistic patterns. Results show differences across racial/ethnic groups in
key sociocognitive features that indicate fruitful collaborative interactions.
We also investigated how the groups' racial/ethnic composition impacts both
individual and group communication patterns. In general, individuals in more
demographically diverse groups displayed more productive communication
behaviors than individuals who were in majority-dominated groups. We discuss
the implications of individual and group diversity on communication patterns
that emerge during CPS and how these patterns can impact collaborative
outcomes.
| 2,024 | Computation and Language |
Fact-Checking the Output of Large Language Models via Token-Level
Uncertainty Quantification | Large language models (LLMs) are notorious for hallucinating, i.e., producing
erroneous claims in their output. Such hallucinations can be dangerous, as
occasional factual inaccuracies in the generated text might be obscured by the
rest of the output being generally factual, making it extremely hard for the
users to spot them. Current services that leverage LLMs usually do not provide
any means for detecting unreliable generations. Here, we aim to bridge this
gap. In particular, we propose a novel fact-checking and hallucination
detection pipeline based on token-level uncertainty quantification. Uncertainty
scores leverage information encapsulated in the output of a neural network or
its layers to detect unreliable predictions, and we show that they can be used
to fact-check the atomic claims in the LLM output. Moreover, we present a novel
token-level uncertainty quantification method that removes the impact of
uncertainty about what claim to generate on the current step and what surface
form to use. Our method Claim Conditioned Probability (CCP) measures only the
uncertainty of particular claim value expressed by the model. Experiments on
the task of biography generation demonstrate strong improvements for CCP
compared to the baselines for six different LLMs and three languages. Human
evaluation reveals that the fact-checking pipeline based on uncertainty
quantification is competitive with a fact-checking tool that leverages external
knowledge.
| 2,024 | Computation and Language |
Common 7B Language Models Already Possess Strong Math Capabilities | Mathematical capabilities were previously believed to emerge in common
language models only at a very large scale or require extensive math-related
pre-training. This paper shows that the LLaMA-2 7B model with common
pre-training already exhibits strong mathematical abilities, as evidenced by
its impressive accuracy of 97.7% and 72.0% on the GSM8K and MATH benchmarks,
respectively, when selecting the best response from 256 random generations. The
primary issue with the current base model is the difficulty in consistently
eliciting its inherent mathematical capabilities. Notably, the accuracy for the
first answer drops to 49.5% and 7.9% on the GSM8K and MATH benchmarks,
respectively. We find that simply scaling up the SFT data can significantly
enhance the reliability of generating correct answers. However, the potential
for extensive scaling is constrained by the scarcity of publicly available math
questions. To overcome this limitation, we employ synthetic data, which proves
to be nearly as effective as real data and shows no clear saturation when
scaled up to approximately one million samples. This straightforward approach
achieves an accuracy of 82.6% on GSM8K and 40.6% on MATH using LLaMA-2 7B
models, surpassing previous models by 14.2% and 20.8%, respectively. We also
provide insights into scaling behaviors across different reasoning complexities
and error types.
| 2,024 | Computation and Language |
LLMs in the Imaginarium: Tool Learning through Simulated Trial and Error | Tools are essential for large language models (LLMs) to acquire up-to-date
information and take consequential actions in external environments. Existing
work on tool-augmented LLMs primarily focuses on the broad coverage of tools
and the flexibility of adding new tools. However, a critical aspect that has
surprisingly been understudied is simply how accurately an LLM uses tools for
which it has been trained. We find that existing LLMs, including GPT-4 and
open-source LLMs specifically fine-tuned for tool use, only reach a correctness
rate in the range of 30% to 60%, far from reliable use in practice. We propose
a biologically inspired method for tool-augmented LLMs, simulated trial and
error (STE), that orchestrates three key mechanisms for successful tool use
behaviors in the biological system: trial and error, imagination, and memory.
Specifically, STE leverages an LLM's 'imagination' to simulate plausible
scenarios for using a tool, after which the LLM interacts with the tool to
learn from its execution feedback. Both short-term and long-term memory are
employed to improve the depth and breadth of the exploration, respectively.
Comprehensive experiments on ToolBench show that STE substantially improves
tool learning for LLMs under both in-context learning and fine-tuning settings,
bringing a boost of 46.7% to Mistral-Instruct-7B and enabling it to outperform
GPT-4. We also show effective continual learning of tools via a simple
experience replay strategy.
| 2,024 | Computation and Language |
Minimum Description Length and Compositionality | We present a non-vacuous definition of compositionality. It is based on the
idea of combining the minimum description length principle with the original
definition of compositionality (that is, that the meaning of the whole is a
function of the meaning of the parts).
The new definition is intuitive and allows us to distinguish between
compositional and non-compositional semantics, and between idiomatic and
non-idiomatic expressions. It is not ad hoc, since it does not make any
references to non-intrinsic properties of meaning functions (like being a
polynomial). Moreover, it allows us to compare different meaning functions with
respect to how compositional they are. It bridges linguistic and corpus-based,
statistical approaches to natural language understanding.
| 1,999 | Computation and Language |
Compositionality, Synonymy, and the Systematic Representation of Meaning | In a recent issue of Linguistics and Philosophy Kasmi and Pelletier (1998)
(K&P), and Westerstahl (1998) criticize Zadrozny's (1994) argument that any
semantics can be represented compositionally. The argument is based upon
Zadrozny's theorem that every meaning function m can be encoded by a function
\mu such that (i) for any expression E of a specified language L, m(E) can be
recovered from \mu(E), and (ii) \mu is a homomorphism from the syntactic
structures of L to interpretations of L.
In both cases, the primary motivation for the objections brought against
Zadrozny's argument is the view that his encoding of the original meaning
function does not properly reflect the synonymy relations posited for the
language.
In this paper, we argue that these technical criticisms do not go through. In
particular, we prove that \mu properly encodes synonymy relations, i.e. if two
expressions are synonymous, then their compositional meanings are identical.
This corrects some misconceptions about the function \mu, e.g. Janssen (1997).
We suggest that the reason that semanticists have been anxious to preserve
compositionality as a significant constraint on semantic theory is that it has
been mistakenly regarded as a condition that must be satisfied by any theory
that sustains a systematic connection between the meaning of an expression and
the meanings of its parts. Recent developments in formal and computational
semantics show that systematic theories of meanings need not be compositional.
| 2,007 | Computation and Language |
A Real World Implementation of Answer Extraction | In this paper we describe ExtrAns, an answer extraction system. Answer
extraction (AE) aims at retrieving those exact passages of a document that
directly answer a given user question. AE is more ambitious than information
retrieval and information extraction in that the retrieval results are phrases,
not entire documents, and in that the queries may be arbitrarily specific. It
is less ambitious than full-fledged question answering in that the answers are
not generated from a knowledge base but looked up in the text of documents. The
current version of ExtrAns is able to parse unedited Unix "man pages", and
derive the logical form of their sentences. User queries are also translated
into logical forms. A theorem prover then retrieves the relevant phrases, which
are presented through selective highlighting in their context.
| 1,998 | Computation and Language |