TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Robust Multilingual Part-of-Speech Tagging via Adversarial...

Robust Multilingual Part-of-Speech Tagging via Adversarial Training

Michihiro Yasunaga, Jungo Kasai, Dragomir Radev

2017-11-14NAACL 2018 6POSPart-Of-Speech TaggingNamed Entity Recognition (NER)ChunkingPOS TaggingDependency Parsing
PaperPDFCode(official)

Abstract

Adversarial training (AT) is a powerful regularization method for neural networks, aiming to achieve robustness to input perturbations. Yet, the specific effects of the robustness obtained from AT are still unclear in the context of natural language processing. In this paper, we propose and analyze a neural POS tagging model that exploits AT. In our experiments on the Penn Treebank WSJ corpus and the Universal Dependencies (UD) dataset (27 languages), we find that AT not only improves the overall tagging accuracy, but also 1) prevents over-fitting well in low resource languages and 2) boosts tagging accuracy for rare / unseen words. We also demonstrate that 3) the improved tagging performance by AT contributes to the downstream task of dependency parsing, and that 4) AT helps the model to learn cleaner word representations. 5) The proposed AT model is generally effective in different sequence labeling tasks. These positive results motivate further use of AT for natural language tasks.

Results

TaskDatasetMetricValueModel
Part-Of-Speech TaggingPenn TreebankAccuracy97.59Adversarial Bi-LSTM
Part-Of-Speech TaggingUDAvg accuracy96.65Adversarial Bi-LSTM
Named Entity Recognition (NER)CoNLL 2003 (English)F191.56Adversarial Bi-LSTM
ChunkingCoNLL 2000Exact Span F195.25Adversarial Training
ChunkingCoNLL 2000Exact Span F195.18BiLSTM-CRF
Shallow SyntaxCoNLL 2000Exact Span F195.25Adversarial Training
Shallow SyntaxCoNLL 2000Exact Span F195.18BiLSTM-CRF

Related Papers

Dynamic Chunking for End-to-End Hierarchical Sequence Modeling2025-07-10CLI-RAG: A Retrieval-Augmented Framework for Clinically Structured and Context Aware Text Generation with LLMs2025-07-09Flippi: End To End GenAI Assistant for E-Commerce2025-07-08Selecting and Merging: Towards Adaptable and Scalable Named Entity Recognition with Large Language Models2025-06-28Can LLMs Replace Humans During Code Chunking?2025-06-24CronusVLA: Transferring Latent Motion Across Time for Multi-Frame Prediction in Manipulation2025-06-24cAST: Enhancing Code Retrieval-Augmented Generation with Structural Chunking via Abstract Syntax Tree2025-06-18LingoLoop Attack: Trapping MLLMs via Linguistic Context and State Entrapment into Endless Loops2025-06-17