TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/DPT

DPT

Dense Prediction Transformer

Computer VisionIntroduced 200025 papers
Source Paper

Description

Dense Prediction Transformers (DPT) are a type of vision transformer for dense prediction tasks.

The input image is transformed into tokens (orange) either by extracting non-overlapping patches followed by a linear projection of their flattened representation (DPT-Base and DPT-Large) or by applying a ResNet-50 feature extractor (DPT-Hybrid). The image embedding is augmented with a positional embedding and a patch-independent readout token (red) is added. The tokens are passed through multiple transformer stages. The tokens are reassembled from different stages into an image-like representation at multiple resolutions (green). Fusion modules (purple) progressively fuse and upsample the representations to generate a fine-grained prediction.

Papers Using This Method

Can In-Context Reinforcement Learning Recover From Reward Poisoning Attacks?2025-06-07Filtering Learning Histories Enhances In-Context Reinforcement Learning2025-05-21SAR Object Detection with Self-Supervised Pretraining and Curriculum-Aware Sampling2025-04-17Random Policy Enables In-Context Reinforcement Learning within Trust Horizons2024-10-25Theoretical limits of descending $\ell_0$ sparse-regression ML algorithms2024-10-10Endogenous Crashes as Phase Transitions2024-08-12Pretraining Decision Transformers with Reward Prediction for In-Context Multi-task Structured Bandit Learning2024-06-07Developmental Pretraining (DPT) for Image Classification Networks2023-12-01Enhancing Diffusion Models with 3D Perspective Geometry Constraints2023-12-01Depth-guided Free-space Segmentation for a Mobile Robot2023-11-03The serotonergic psychedelic N,N-dipropyltryptamine alters information-processing dynamics in cortical neural circuits2023-10-31Supervised Pretraining Can Learn In-Context Reinforcement Learning2023-06-26High-Resolution Synthetic RGB-D Datasets for Monocular Depth Estimation2023-05-02Diffusion Models and Semi-Supervised Learners Benefit Mutually with Few Labels2023-02-21Denoising and Prompt-Tuning for Multi-Behavior Recommendation2023-02-12DPTDR: Deep Prompt Tuning for Dense Passage Retrieval2022-08-24Dual Modality Prompt Tuning for Vision-Language Pre-Trained Model2022-08-17SSDPT: Self-Supervised Dual-Path Transformer for Anomalous Sound Detection in Machine Condition Monitoring2022-08-06Prompt Tuning for Discriminative Pre-trained Language Models2022-05-23Declaration-based Prompt Tuning for Visual Question Answering2022-05-05