[PDF] Integrating Weakly Supervised Word Sense Disambiguation into Neural Machine Translation | Semantic Scholar (2024)

Skip to search formSkip to main contentSkip to account menu

Semantic ScholarSemantic Scholar's Logo
@article{Pu2018IntegratingWS, title={Integrating Weakly Supervised Word Sense Disambiguation into Neural Machine Translation}, author={Xiao Pu and Nikolaos Pappas and James Henderson and Andrei Popescu-Belis}, journal={Transactions of the Association for Computational Linguistics}, year={2018}, volume={6}, pages={635-649}, url={https://api.semanticscholar.org/CorpusID:52933530}}
  • X. Pu, Nikolaos Pappas, Andrei Popescu-Belis
  • Published in Transactions of the… 5 October 2018
  • Computer Science, Linguistics

This paper demonstrates that word sense disambiguation (WSD) can improve neural machine translation (NMT) by widening the source context considered when modeling the senses of potentially ambiguous

38 Citations

Highly Influential Citations

4

Background Citations

25

Methods Citations

7

Topics

Word Sense Disambiguation (opens in a new tab)Neural MT (opens in a new tab)Sense Vectors (opens in a new tab)Neural Machine Translation (opens in a new tab)Random Walks (opens in a new tab)Chinese Restaurant Processes (opens in a new tab)Source Context (opens in a new tab)K-means (opens in a new tab)Word Vectors (opens in a new tab)BLEU Points (opens in a new tab)

Ask This Paper

BETA

AI-Powered

Our system tries to constrain to information found in this paper. Results quality may vary. Learn more about how we generate these answers.

Feedback?

38 Citations

Improving Machine Translation of Rare and Unseen Word Senses
    Viktor HangyaQianchu LiuDario StojanovskiAlexander M. FraserA. Korhonen

    Computer Science

    WMT

  • 2021

This work proposes CmBT (Contextually-mined Back-Translation), an approach for improving multi-sense word translation leveraging pre-trained cross-lingual contextual word representations (CCWRs), and shows that the system improves on the translation of difficult unseen and low frequency word senses.

  • 3
  • PDF
Multilingual Word Sense Disambiguation with Unified Sense Representation
    Ying SuHongming ZhangYangqiu SongTong Zhang

    Computer Science, Linguistics

    COLING

  • 2022

This paper proposes to build knowledge and supervised based Multilingual Word Sense Disambiguation systems and address the annotation scarcity problem for MWSD by transferring annotations from rich sourced languages.

CluBERT: A Cluster-Based Approach for Learning Sense Distributions in Multiple Languages
    Tommaso PasiniFederico ScozzafavaBianca Scarlini

    Computer Science, Linguistics

    ACL

  • 2020

This paper presents CluBERT, an automatic and multilingual approach for inducing the distributions of word senses from a corpus of raw sentences that attains state-of-the-art results on the English Word Sense Disambiguation tasks and helps to improve the disambiguated performance of two off- the-shelf WSD models.

  • 15
  • PDF
Zero-shot Word Sense Disambiguation using Sense Definition Embeddings
    Sawan KumarSharmistha JatKaran SaxenaP. Talukdar

    Computer Science, Linguistics

    ACL

  • 2019

This work proposes Extended WSD Incorporating Sense Embeddings (EWISE), a supervised model to perform WSD by predicting over a continuous sense embedding space as opposed to a discrete label space, which allows EWISE to generalize over both seen and unseen senses, thus achieving generalized zero-shot learning.

  • 91
  • PDF
Improving Word Sense Disambiguation in Neural Machine Translation with Salient Document Context
    Elijah Matthew RippethMarine CarpuatKevin DuhMatt Post

    Computer Science

    ArXiv

  • 2023

This work introduces a simple and scalable approach to resolve translation ambiguity by incorporating a small amount of extra-sentential context in neural neural models to translate ambiguous source words better than strong sentence- level baselines and comparable document-level baselines while reducing training costs.

Towards Effective Disambiguation for Machine Translation with Large Language Models
    Vivek IyerPinzhen ChenAlexandra Birch

    Computer Science, Linguistics

    WMT

  • 2023

The capabilities of LLMs to translate “ambiguous sentences” are studied, and two ways to improve their disambiguation capabilities are proposed, through a) in-context learning and b) fine-tuning on carefully curated ambiguous datasets.

An Evaluation Benchmark for Testing the Word Sense Disambiguation Capabilities of Machine Translation Systems
    Alessandro RaganatoYves ScherrerJ. Tiedemann

    Computer Science, Linguistics

    LREC

  • 2020

This paper presents an evaluation benchmark on WSD for machine translation for 10 language pairs, comprising training data with known sense distributions, and builds upon the wide-coverage multilingual sense inventory of BabelNet, the multilingual neural parsing pipeline TurkuNLP, and the OPUS collection of translated texts from the web.

  • 10
  • Highly Influenced
  • PDF
Word Sense Consistency in Statistical and Neural Machine Translation
    X. Pu

    Computer Science, Linguistics

  • 2018

This thesis proposes a method to decide whether two occurrences of the same noun in a source text should be translated consistently, and designs sense-aware MT systems that select the correct translations of ambiguous words by performing word sense disambiguation (WSD).

  • PDF
The Contribution of Selected Linguistic Markers for Unsupervised Arabic Verb Sense Disambiguation
    Asma DjaidriH. AlianeH. Azzoune

    Linguistics, Computer Science

    ACM Trans. Asian Low Resour. Lang. Inf. Process.

  • 2023

This work uses contextualized word embeddings for an unsupervised Arabic WSD that is based on linguistic markers and uses sentence-BERT Transformer pre-trained models, which yields encouraging results that outperform other existing un supervised neural AWSD approaches.

  • 1
Rare and Zero-shot Word Sense Disambiguation using Z-Reweighting
    Ying SuHongming ZhangYangqiu SongTong Zhang

    Computer Science

    ACL

  • 2022

This work investigates the statistical relation between word frequency rank and word sense number distribution and proposes a Z-reweighting method on the word level to adjust the training on the imbalanced dataset.

  • 6
  • PDF

...

...

48 References

Improving Statistical Machine Translation Using Word Sense Disambiguation
    Marine CarpuatDekai Wu

    Computer Science, Linguistics

    EMNLP

  • 2007

This paper investigates a new strategy for integrating WSD into an SMT system, that performs fully phrasal multi-word disambiguation, and provides the first known empirical evidence that lexical semantics are indeed useful for SMT, despite claims to the contrary.

  • 406
  • PDF
Improving Word Sense Disambiguation in Neural Machine Translation with Sense Embeddings
    Annette Rios GonzalesLaura MascarellRico Sennrich

    Computer Science, Linguistics

    WMT

  • 2017

While a baseline NMT system disambiguates frequent word senses quite reliably, the annotation with both sense labels and lexical chains improves the neural models’ performance on rare word senses.

  • 118
  • PDF
Word-Sense Disambiguation for Machine Translation
    David VickreyLuke BiewaldM. TeyssierD. Koller

    Computer Science, Linguistics

    HLT

  • 2005

It is shown that the word-translation system can be used to improve performance on a simplified machine-translation task and can effectively and accurately prune the set of candidate translations for a word.

  • 204
  • PDF
Sense-Aware Statistical Machine Translation using Adaptive Context-Dependent Clustering
    X. PuNikolaos PappasAndrei Popescu-Belis

    Computer Science

    WMT

  • 2017

This work demonstrates that WSD systems can be adapted to help SMT, thanks to three key achievements: it considers a larger context for WSD than SMT can afford to consider, and it adapts the number of senses per word to the ones observed in the training data.

  • 7
  • PDF
Multi-sense based neural machine translation
    Zhen YangWei ChenFeng WangBo Xu

    Computer Science

    2017 International Joint Conference on Neural…

  • 2017

This paper validates the hypothesis and proposes a simple and flexible framework, which enables the NMT model to only focus on the relevant sense type of the input word in current context and achieves substantial improvements on every test set over competitive baselines.

  • 6
A Sense-Based Translation Model for Statistical Machine Translation
    Deyi XiongMin Zhang

    Computer Science, Linguistics

    ACL

  • 2014

A sense-based translation model to integrate word senses into statistical machine translation based on a nonparametric Bayesian topic model that automatically learns sense clusters for words in the source language is proposed.

  • 34
  • PDF
Do Multi-Sense Embeddings Improve Natural Language Understanding?
    Jiwei LiDan Jurafsky

    Computer Science, Linguistics

    EMNLP

  • 2015

A multisense embedding model based on Chinese Restaurant Processes is introduced that achieves state of the art performance on matching human word similarity judgments, and a pipelined architecture for incorporating multi-sense embeddings into language understanding is proposed.

Word Sense-Aware Machine Translation: Including Senses as Contextual Features for Improved Translation Models
    Steven NealeLuís Manuel dos Santos GomesEneko AgirreOier Lopez de LacalleA. Branco

    Computer Science, Linguistics

    LREC

  • 2016

Training on a large, open-domain corpus (Europarl) and including word senses as contextual features in maxent-based translation models yields significant improvements in machine translation from English to Portuguese.

  • 28
  • PDF
Handling hom*ographs in Neural Machine Translation
    Frederick LiuHan LuGraham Neubig

    Computer Science, Linguistics

    NAACL

  • 2018

Empirical evidence is provided that existing NMT systems in fact still have significant problems in properly translating ambiguous words, and methods are described that model the context of the input word with context-aware word embeddings that help to differentiate the word sense before feeding it into the encoder.

Context-dependent word representation for neural machine translation
    Heeyoul ChoiKyunghyun ChoYoshua Bengio

    Computer Science

    Comput. Speech Lang.

  • 2017

...

...

Related Papers

Showing 1 through 3 of 0 Related Papers

    [PDF] Integrating Weakly Supervised Word Sense Disambiguation into Neural Machine Translation | Semantic Scholar (2024)
    Top Articles
    Latest Posts
    Article information

    Author: Tish Haag

    Last Updated:

    Views: 5789

    Rating: 4.7 / 5 (47 voted)

    Reviews: 86% of readers found this page helpful

    Author information

    Name: Tish Haag

    Birthday: 1999-11-18

    Address: 30256 Tara Expressway, Kutchburgh, VT 92892-0078

    Phone: +4215847628708

    Job: Internal Consulting Engineer

    Hobby: Roller skating, Roller skating, Kayaking, Flying, Graffiti, Ghost hunting, scrapbook

    Introduction: My name is Tish Haag, I am a excited, delightful, curious, beautiful, agreeable, enchanting, fancy person who loves writing and wants to share my knowledge and understanding with you.