TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Learning Better Internal Structure of Words for Sequence L...

Learning Better Internal Structure of Words for Sequence Labeling

Yingwei Xin, Ethan Hart, Vibhuti Mahajan, Jean-David Ruvini

2018-10-29EMNLP 2018 10Part-Of-Speech Taggingnamed-entity-recognitionNamed Entity RecognitionNamed Entity Recognition (NER)Chunking
PaperPDF

Abstract

Character-based neural models have recently proven very useful for many NLP tasks. However, there is a gap of sophistication between methods for learning representations of sentences and words. While most character models for learning representations of sentences are deep and complex, models for learning representations of words are shallow and simple. Also, in spite of considerable research on learning character embeddings, it is still not clear which kind of architecture is the best for capturing character-to-word representations. To address these questions, we first investigate the gaps between methods for learning word and sentence representations. We conduct detailed experiments and comparisons of different state-of-the-art convolutional models, and also investigate the advantages and disadvantages of their constituents. Furthermore, we propose IntNet, a funnel-shaped wide convolutional neural architecture with no down-sampling for learning representations of the internal structure of words by composing their characters from limited, supervised training corpora. We evaluate our proposed model on six sequence labeling datasets, including named entity recognition, part-of-speech tagging, and syntactic chunking. Our in-depth analysis shows that IntNet significantly outperforms other character embedding models and obtains new state-of-the-art performance without relying on any external knowledge or resources.

Results

TaskDatasetMetricValueModel
Part-Of-Speech TaggingPenn TreebankAccuracy97.58IntNet + BiLSTM-CRF
Named Entity Recognition (NER)CoNLL 2003 (English)F191.64IntNet + BiLSTM-CRF
ChunkingPenn TreebankF1 score95.29IntNet + BiLSTM-CRF
Shallow SyntaxPenn TreebankF1 score95.29IntNet + BiLSTM-CRF

Related Papers

Dynamic Chunking for End-to-End Hierarchical Sequence Modeling2025-07-10CLI-RAG: A Retrieval-Augmented Framework for Clinically Structured and Context Aware Text Generation with LLMs2025-07-09Flippi: End To End GenAI Assistant for E-Commerce2025-07-08Selecting and Merging: Towards Adaptable and Scalable Named Entity Recognition with Large Language Models2025-06-28Can LLMs Replace Humans During Code Chunking?2025-06-24CronusVLA: Transferring Latent Motion Across Time for Multi-Frame Prediction in Manipulation2025-06-24cAST: Enhancing Code Retrieval-Augmented Generation with Structural Chunking via Abstract Syntax Tree2025-06-18Chunk Twice, Embed Once: A Systematic Study of Segmentation and Representation Trade-offs in Chemistry-Aware Retrieval-Augmented Generation2025-06-13