TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/HiTIN: Hierarchy-aware Tree Isomorphism Network for Hierar...

HiTIN: Hierarchy-aware Tree Isomorphism Network for Hierarchical Text Classification

He Zhu, Chong Zhang, JunJie Huang, Junran Wu, Ke Xu

2023-05-24Text Classificationtext-classificationMulti-Label ClassificationMulti-Label Text ClassificationHierarchical Multi-label Classification
PaperPDFCode(official)

Abstract

Hierarchical text classification (HTC) is a challenging subtask of multi-label classification as the labels form a complex hierarchical structure. Existing dual-encoder methods in HTC achieve weak performance gains with huge memory overheads and their structure encoders heavily rely on domain knowledge. Under such observation, we tend to investigate the feasibility of a memory-friendly model with strong generalization capability that could boost the performance of HTC without prior statistics or label semantics. In this paper, we propose Hierarchy-aware Tree Isomorphism Network (HiTIN) to enhance the text representations with only syntactic information of the label hierarchy. Specifically, we convert the label hierarchy into an unweighted tree structure, termed coding tree, with the guidance of structural entropy. Then we design a structure encoder to incorporate hierarchy-aware information in the coding tree into text representations. Besides the text encoder, HiTIN only contains a few multi-layer perceptions and linear transformations, which greatly saves memory. We conduct experiments on three commonly used datasets and the results demonstrate that HiTIN could achieve better test performance and less memory consumption than state-of-the-art (SOTA) methods.

Results

TaskDatasetMetricValueModel
Multi-Label ClassificationRCV1-v2Macro F169.95HiTIN+BERT
Multi-Label ClassificationRCV1-v2Micro F186.71HiTIN+BERT
Multi-Label ClassificationRCV1-v2Macro F164.37HiTIN
Multi-Label ClassificationRCV1-v2Micro F184.81HiTIN
Multi-Label ClassificationNew York Times Annotated CorpusMacro F169.31HiTIN+BERT
Multi-Label ClassificationNew York Times Annotated CorpusMicro F179.65HiTIN+BERT
Multi-Label ClassificationNew York Times Annotated CorpusMacro F161.09HiTIN
Multi-Label ClassificationNew York Times Annotated CorpusMicro F175.13HiTIN
Multi-Label ClassificationWOSMacro F181.57HiTIN+BERT
Multi-Label ClassificationWOSMicro F187.19HiTIN+BERT
Multi-Label ClassificationWOSMacro F181.11HiTIN
Multi-Label ClassificationWOSMicro F186.66HiTIN

Related Papers

Making Language Model a Hierarchical Classifier and Generator2025-07-17GNN-CNN: An Efficient Hybrid Model of Convolutional and Graph Neural Networks for Text Representation2025-07-10The Trilemma of Truth in Large Language Models2025-06-30Robustness of Misinformation Classification Systems to Adversarial Examples Through BeamAttack2025-06-30Perspectives in Play: A Multi-Perspective Approach for More Inclusive NLP Systems2025-06-25Can Generated Images Serve as a Viable Modality for Text-Centric Multimodal Learning?2025-06-21SHREC and PHEONA: Using Large Language Models to Advance Next-Generation Computational Phenotyping2025-06-19Privacy-Preserving Chest X-ray Classification in Latent Space with Homomorphically Encrypted Neural Inference2025-06-18