TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Multilingual training for Software Engineering

Multilingual training for Software Engineering

Toufique Ahmed, Premkumar Devanbu

2021-12-03Type predictionCode SummarizationRetrieval
PaperPDF

Abstract

Well-trained machine-learning models, which leverage large amounts of open-source software data, have now become an interesting approach to automating many software engineering tasks. Several SE tasks have all been subject to this approach, with performance gradually improving over the past several years with better models and training methods. More, and more diverse, clean, labeled data is better for training; but constructing good-quality datasets is time-consuming and challenging. Ways of augmenting the volume and diversity of clean, labeled data generally have wide applicability. For some languages (e.g., Ruby) labeled data is less abundant; in others (e.g., JavaScript) the available data maybe more focused on some application domains, and thus less diverse. As a way around such data bottlenecks, we present evidence suggesting that human-written code in different languages (which performs the same function), is rather similar, and particularly preserving of identifier naming patterns; we further present evidence suggesting that identifiers are a very important element of training data for software engineering tasks. We leverage this rather fortuitous phenomenon to find evidence that available multilingual training data (across different languages) can be used to amplify performance. We study this for 3 different tasks: code summarization, code retrieval, and function naming. We note that this data-augmenting approach is broadly compatible with different tasks, languages, and machine-learning models.

Results

TaskDatasetMetricValueModel
Program SynthesisManyTypes4TypeScriptAverage Accuracy61.29PolyGot
Program SynthesisManyTypes4TypeScriptAverage F158.86PolyGot
Program SynthesisManyTypes4TypeScriptAverage Precision58.81PolyGot
Program SynthesisManyTypes4TypeScriptAverage Recall58.91PolyGot
Program SynthesisManyTypes4TypeScriptAverage Accuracy61GraphPolyGot
Program SynthesisManyTypes4TypeScriptAverage F158.63GraphPolyGot
Program SynthesisManyTypes4TypeScriptAverage Precision58.36GraphPolyGot
Program SynthesisManyTypes4TypeScriptAverage Recall58.91GraphPolyGot
Type predictionManyTypes4TypeScriptAverage Accuracy61.29PolyGot
Type predictionManyTypes4TypeScriptAverage F158.86PolyGot
Type predictionManyTypes4TypeScriptAverage Precision58.81PolyGot
Type predictionManyTypes4TypeScriptAverage Recall58.91PolyGot
Type predictionManyTypes4TypeScriptAverage Accuracy61GraphPolyGot
Type predictionManyTypes4TypeScriptAverage F158.63GraphPolyGot
Type predictionManyTypes4TypeScriptAverage Precision58.36GraphPolyGot
Type predictionManyTypes4TypeScriptAverage Recall58.91GraphPolyGot

Related Papers

From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17HapticCap: A Multimodal Dataset and Task for Understanding User Experience of Vibration Haptic Signals2025-07-17A Survey of Context Engineering for Large Language Models2025-07-17MCoT-RE: Multi-Faceted Chain-of-Thought and Re-Ranking for Training-Free Zero-Shot Composed Image Retrieval2025-07-17Developing Visual Augmented Q&A System using Scalable Vision Embedding Retrieval & Late Interaction Re-ranker2025-07-16Language-Guided Contrastive Audio-Visual Masked Autoencoder with Automatically Generated Audio-Visual-Text Triplets from Videos2025-07-16Context-Aware Search and Retrieval Over Erasure Channels2025-07-16Seq vs Seq: An Open Suite of Paired Encoders and Decoders2025-07-15