TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Self-training from Self-memory in Data-to-text Generation

Self-training from Self-memory in Data-to-text Generation

Hoang-Thang Ta

2024-01-19Continual LearningData-to-Text GenerationText Generation
PaperPDFCode(official)

Abstract

This paper introduces a novel training model, self-training from self-memory (STSM) in data-to-text generation (DTG), allowing the model to self-train on subsets, including self-memory as outputs inferred directly from the trained models and/or the new data. The quality of self-memory is validated by two models, data-to-text (D2T) and text-to-data (T2D), by two pre-defined conditions: (1) the appearance of all source values in the outputs of the D2T model and (2) the ability to convert back to source data in the outputs in the T2D model. We utilize a greedy algorithm to generate shorter D2T outputs if they contain all source values. Subsequently, we use the T2D model to confirm that these outputs can capture input relationships by demonstrating their capacity to convert text back into data. With 30% of the dataset, we can train the D2T model with a competitive performance compared to full training in the same setup. We experiment with our model on two datasets, E2E NLG and DART. STSM offers the D2T model a generalization capability from its subset memory while reducing training data volume. Ultimately, we anticipate that this paper will contribute to continual learning solutions that adapt to new training data, incorporating it as a form of self-memory in DTG tasks. The curated dataset is publicly available at: https://github.com/hoangthangta/STSM.

Results

TaskDatasetMetricValueModel
Text GenerationDARTBLEU47.76self-mem + new data
Text GenerationE2EMETEOR46.11self-mem + new data (random)
Text GenerationE2EMETEOR46.07self-mem + new data (fixed)
Text GenerationE2E NLG ChallengeBLEU65.11Self-memory
Text GenerationE2E NLG ChallengeCIDEr2.16Self-memory
Text GenerationE2E NLG ChallengeMETEOR46.11Self-memory
Text GenerationE2E NLG ChallengeNIST8.35Self-memory
Text GenerationE2E NLG ChallengeROUGE-L68.41Self-memory
Data-to-Text GenerationDARTBLEU47.76self-mem + new data
Data-to-Text GenerationE2EMETEOR46.11self-mem + new data (random)
Data-to-Text GenerationE2EMETEOR46.07self-mem + new data (fixed)
Data-to-Text GenerationE2E NLG ChallengeBLEU65.11Self-memory
Data-to-Text GenerationE2E NLG ChallengeCIDEr2.16Self-memory
Data-to-Text GenerationE2E NLG ChallengeMETEOR46.11Self-memory
Data-to-Text GenerationE2E NLG ChallengeNIST8.35Self-memory
Data-to-Text GenerationE2E NLG ChallengeROUGE-L68.41Self-memory

Related Papers

Making Language Model a Hierarchical Classifier and Generator2025-07-17RegCL: Continual Adaptation of Segment Anything Model via Model Merging2025-07-16Information-Theoretic Generalization Bounds of Replay-based Continual Learning2025-07-16PROL : Rehearsal Free Continual Learning in Streaming Data via Prompt Online Learning2025-07-16Mitigating Object Hallucinations via Sentence-Level Early Intervention2025-07-16Fast Last-Iterate Convergence of SGD in the Smooth Interpolation Regime2025-07-15A Neural Network Model of Complementary Learning Systems: Pattern Separation and Completion for Continual Learning2025-07-15The Devil behind the mask: An emergent safety vulnerability of Diffusion LLMs2025-07-15