WiC: the Word-in-Context Dataset for Evaluating Context-Sensitive Meaning Representations
Abstract
A new benchmark named WiC is introduced for evaluating context-sensitive word representations, highlighting the limitations of existing datasets.
By design, word embeddings are unable to model the dynamic nature of words' semantics, i.e., the property of words to correspond to potentially different meanings. To address this limitation, dozens of specialized meaning representation techniques such as sense or contextualized embeddings have been proposed. However, despite the popularity of research on this topic, very few evaluation benchmarks exist that specifically focus on the dynamic semantics of words. In this paper we show that existing models have surpassed the performance ceiling of the standard evaluation dataset for the purpose, i.e., Stanford Contextual Word Similarity, and highlight its shortcomings. To address the lack of a suitable benchmark, we put forward a large-scale Word in Context dataset, called WiC, based on annotations curated by experts, for generic evaluation of context-sensitive representations. WiC is released in https://pilehvar.github.io/wic/.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- A Comparative Analysis of Static Word Embeddings for Hungarian (2025)
- Static Word Embeddings for Sentence Semantic Representation (2025)
- Multilinguality Does not Make Sense: Investigating Factors Behind Zero-Shot Transfer in Sense-Aware Tasks (2025)
- Fast, Not Fancy: Rethinking G2P with Rich Data and Rule-Based Models (2025)
- PolyBERT: Fine-Tuned Poly Encoder BERT-Based Model for Word Sense Disambiguation (2025)
- semantic-features: A User-Friendly Tool for Studying Contextual Word Embeddings in Interpretable Semantic Spaces (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 21
Browse 21 models citing this paperDatasets citing this paper 1
Spaces citing this paper 1,761
Collections including this paper 0
No Collection including this paper