{ "1512.03385": { "arxivId": "1512.03385", "title": "Deep Residual Learning for Image Recognition" }, "1706.03762": { "arxivId": "1706.03762", "title": "Attention is All you Need" }, "1810.04805": { "arxivId": "1810.04805", "title": "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding" }, "1505.04597": { "arxivId": "1505.04597", "title": "U-Net: Convolutional Networks for Biomedical Image Segmentation" }, "1506.01497": { "arxivId": "1506.01497", "title": "Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks" }, "1405.0312": { "arxivId": "1405.0312", "title": "Microsoft COCO: Common Objects in Context" }, "2005.14165": { "arxivId": "2005.14165", "title": "Language Models are Few-Shot Learners" }, "1506.02640": { "arxivId": "1506.02640", "title": "You Only Look Once: Unified, Real-Time Object Detection" }, "2010.11929": { "arxivId": "2010.11929", "title": "An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale" }, "1703.06870": { "arxivId": "1703.06870", "title": "Mask R-CNN" }, "1709.01507": { "arxivId": "1709.01507", "title": "Squeeze-and-Excitation Networks" }, "1907.11692": { "arxivId": "1907.11692", "title": "RoBERTa: A Robustly Optimized BERT Pretraining Approach" }, "2103.00020": { "arxivId": "2103.00020", "title": "Learning Transferable Visual Models From Natural Language Supervision" }, "1612.03144": { "arxivId": "1612.03144", "title": "Feature Pyramid Networks for Object Detection" }, "1409.3215": { "arxivId": "1409.3215", "title": "Sequence to Sequence Learning with Neural Networks" }, "1801.04381": { "arxivId": "1801.04381", "title": "MobileNetV2: Inverted Residuals and Linear Bottlenecks" }, "1606.00915": { "arxivId": "1606.00915", "title": "DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs" }, "2103.14030": { "arxivId": "2103.14030", "title": "Swin Transformer: Hierarchical Vision Transformer using Shifted Windows" }, "1905.11946": { "arxivId": "1905.11946", "title": "EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks" }, "1807.06521": { "arxivId": "1807.06521", "title": "CBAM: Convolutional Block Attention Module" }, "2005.12872": { "arxivId": "2005.12872", "title": "End-to-End Object Detection with Transformers" }, "1611.05431": { "arxivId": "1611.05431", "title": "Aggregated Residual Transformations for Deep Neural Networks" }, "1706.02413": { "arxivId": "1706.02413", "title": "PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space" }, "1607.06450": { "arxivId": "1607.06450", "title": "Layer Normalization" }, "1711.07971": { "arxivId": "1711.07971", "title": "Non-local Neural Networks" }, "1906.08237": { "arxivId": "1906.08237", "title": "XLNet: Generalized Autoregressive Pretraining for Language Understanding" }, "2111.06377": { "arxivId": "2111.06377", "title": "Masked Autoencoders Are Scalable Vision Learners" }, "1909.11942": { "arxivId": "1909.11942", "title": "ALBERT: A Lite BERT for Self-supervised Learning of Language Representations" }, "1411.4555": { "arxivId": "1411.4555", "title": "Show and tell: A neural image caption generator" }, "2012.12877": { "arxivId": "2012.12877", "title": "Training data-efficient image transformers & distillation through attention" }, "1602.07332": { "arxivId": "1602.07332", "title": "Visual Genome: Connecting Language and Vision Using Crowdsourced Dense Image Annotations" }, "1505.00468": { "arxivId": "1505.00468", "title": "VQA: Visual Question Answering" }, "1503.04069": { "arxivId": "1503.04069", "title": "LSTM: A Search Space Odyssey" }, "1703.06211": { "arxivId": "1703.06211", "title": "Deformable Convolutional Networks" }, "2104.14294": { "arxivId": "2104.14294", "title": "Emerging Properties in Self-Supervised Vision Transformers" }, "1904.01355": { "arxivId": "1904.01355", "title": "FCOS: Fully Convolutional One-Stage Object Detection" }, "1712.00726": { "arxivId": "1712.00726", "title": "Cascade R-CNN: Delving Into High Quality Object Detection" }, "1905.04899": { "arxivId": "1905.04899", "title": "CutMix: Regularization Strategy to Train Strong Classifiers With Localizable Features" }, "2010.04159": { "arxivId": "2010.04159", "title": "Deformable DETR: Deformable Transformers for End-to-End Object Detection" }, "2102.12092": { "arxivId": "2102.12092", "title": "Zero-Shot Text-to-Image Generation" }, "1707.07998": { "arxivId": "1707.07998", "title": "Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering" }, "2105.15203": { "arxivId": "2105.15203", "title": "SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers" }, "1902.09212": { "arxivId": "1902.09212", "title": "Deep High-Resolution Representation Learning for Human Pose Estimation" }, "2006.09882": { "arxivId": "2006.09882", "title": "Unsupervised Learning of Visual Features by Contrasting Cluster Assignments" }, "1807.05511": { "arxivId": "1807.05511", "title": "Object Detection With Deep Learning: A Review" }, "1908.02265": { "arxivId": "1908.02265", "title": "ViLBERT: Pretraining Task-Agnostic Visiolinguistic Representations for Vision-and-Language Tasks" }, "2102.12122": { "arxivId": "2102.12122", "title": "Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions" }, "1910.03151": { "arxivId": "1910.03151", "title": "ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks" }, "2102.05918": { "arxivId": "2102.05918", "title": "Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision" }, "1806.01261": { "arxivId": "1806.01261", "title": "Relational inductive biases, deep learning, and graph networks" }, "1612.00837": { "arxivId": "1612.00837", "title": "Making the V in VQA Matter: Elevating the Role of Image Understanding in Visual Question Answering" }, "2102.04306": { "arxivId": "2102.04306", "title": "TransUNet: Transformers Make Strong Encoders for Medical Image Segmentation" }, "2012.15840": { "arxivId": "2012.15840", "title": "Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective with Transformers" }, "2106.08254": { "arxivId": "2106.08254", "title": "BEiT: BERT Pre-Training of Image Transformers" }, "1811.11721": { "arxivId": "1811.11721", "title": "CCNet: Criss-Cross Attention for Semantic Segmentation" }, "1908.07490": { "arxivId": "1908.07490", "title": "LXMERT: Learning Cross-Modality Encoder Representations from Transformers" }, "1707.02968": { "arxivId": "1707.02968", "title": "Revisiting Unreasonable Effectiveness of Data in Deep Learning Era" }, "1803.02155": { "arxivId": "1803.02155", "title": "Self-Attention with Relative Position Representations" }, "2101.01169": { "arxivId": "2101.01169", "title": "Transformers in Vision: A Survey" }, "1908.03557": { "arxivId": "1908.03557", "title": "VisualBERT: A Simple and Performant Baseline for Vision and Language" }, "2004.06165": { "arxivId": "2004.06165", "title": "Oscar: Object-Semantics Aligned Pre-training for Vision-Language Tasks" }, "2101.11986": { "arxivId": "2101.11986", "title": "Tokens-to-Token ViT: Training Vision Transformers from Scratch on ImageNet" }, "2103.15808": { "arxivId": "2103.15808", "title": "CvT: Introducing Convolutions to Vision Transformers" }, "2011.00931": { "arxivId": "2011.00931", "title": "Point Transformer" }, "1807.10221": { "arxivId": "1807.10221", "title": "Unified Perceptual Parsing for Scene Understanding" }, "1908.08530": { "arxivId": "1908.08530", "title": "VL-BERT: Pre-training of Generic Visual-Linguistic Representations" }, "2104.02057": { "arxivId": "2104.02057", "title": "An Empirical Study of Training Self-Supervised Vision Transformers" }, "1802.05751": { "arxivId": "1802.05751", "title": "Image Transformer" }, "1811.08383": { "arxivId": "1811.08383", "title": "TSM: Temporal Shift Module for Efficient Video Understanding" }, "2012.12556": { "arxivId": "2012.12556", "title": "A Survey on Vision Transformer" }, "2012.00364": { "arxivId": "2012.00364", "title": "Pre-Trained Image Processing Transformer" }, "2102.03334": { "arxivId": "2102.03334", "title": "ViLT: Vision-and-Language Transformer Without Convolution or Region Supervision" }, "1904.11492": { "arxivId": "1904.11492", "title": "GCNet: Non-Local Networks Meet Squeeze-Excitation Networks and Beyond" }, "2004.08955": { "arxivId": "2004.08955", "title": "ResNeSt: Split-Attention Networks" }, "2103.00112": { "arxivId": "2103.00112", "title": "Transformer in Transformer" }, "2012.09688": { "arxivId": "2012.09688", "title": "PCT: Point cloud transformer" }, "1801.00868": { "arxivId": "1801.00868", "title": "Panoptic Segmentation" }, "1712.04851": { "arxivId": "1712.04851", "title": "Rethinking Spatiotemporal Feature Learning: Speed-Accuracy Trade-offs in Video Classification" }, "2107.06278": { "arxivId": "2107.06278", "title": "Per-Pixel Classification is Not All You Need for Semantic Segmentation" }, "2106.13797": { "arxivId": "2106.13797", "title": "PVT v2: Improved baselines with Pyramid Vision Transformer" }, "1811.00982": { "arxivId": "1811.00982", "title": "The Open Images Dataset V4" }, "2105.05633": { "arxivId": "2105.05633", "title": "Segmenter: Transformer for Semantic Segmentation" }, "1711.11575": { "arxivId": "1711.11575", "title": "Relation Networks for Object Detection" }, "2103.14899": { "arxivId": "2103.14899", "title": "CrossViT: Cross-Attention Multi-Scale Vision Transformer for Image Classification" }, "1904.01766": { "arxivId": "1904.01766", "title": "VideoBERT: A Joint Model for Video and Language Representation Learning" }, "1906.05909": { "arxivId": "1906.05909", "title": "Stand-Alone Self-Attention in Vision Models" }, "1912.11370": { "arxivId": "1912.11370", "title": "Big Transfer (BiT): General Visual Representation Learning" }, "1906.09756": { "arxivId": "1906.09756", "title": "Cascade R-CNN: High Quality Object Detection and Instance Segmentation" }, "2106.04803": { "arxivId": "2106.04803", "title": "CoAtNet: Marrying Convolution and Attention for All Data Sizes" }, "2009.06732": { "arxivId": "2009.06732", "title": "Efficient Transformers: A Survey" }, "2203.03605": { "arxivId": "2203.03605", "title": "DINO: DETR with Improved DeNoising Anchor Boxes for End-to-End Object Detection" }, "1904.09925": { "arxivId": "1904.09925", "title": "Attention Augmented Convolutional Networks" }, "2011.12450": { "arxivId": "2011.12450", "title": "Sparse R-CNN: End-to-End Object Detection with Learnable Proposals" }, "2106.04560": { "arxivId": "2106.04560", "title": "Scaling Vision Transformers" }, "2101.11605": { "arxivId": "2101.11605", "title": "Bottleneck Transformers for Visual Recognition" }, "2104.13840": { "arxivId": "2104.13840", "title": "Twins: Revisiting the Design of Spatial Attention in Vision Transformers" }, "2103.17239": { "arxivId": "2103.17239", "title": "Going deeper with Image Transformers" }, "1909.11059": { "arxivId": "1909.11059", "title": "Unified Vision-Language Pre-Training for Image Captioning and VQA" }, "2106.04554": { "arxivId": "2106.04554", "title": "A Survey of Transformers" }, "2103.15436": { "arxivId": "2103.15436", "title": "Transformer Tracking" }, "2104.12763": { "arxivId": "2104.12763", "title": "MDETR - Modulated Detection for End-to-End Multi-Modal Understanding" }, "2103.10697": { "arxivId": "2103.10697", "title": "ConViT: improving vision transformers with soft convolutional inductive biases" }, "2202.03555": { "arxivId": "2202.03555", "title": "data2vec: A General Framework for Self-supervised Learning in Speech, Vision and Language" }, "2004.13621": { "arxivId": "2004.13621", "title": "Exploring Self-Attention for Image Recognition" }, "2111.11418": { "arxivId": "2111.11418", "title": "MetaFormer is Actually What You Need for Vision" }, "2108.10904": { "arxivId": "2108.10904", "title": "SimVLM: Simple Visual Language Model Pretraining with Weak Supervision" }, "2106.14881": { "arxivId": "2106.14881", "title": "Early Convolutions Help Transformers See Better" }, "2011.14503": { "arxivId": "2011.14503", "title": "End-to-End Video Instance Segmentation with Transformers" }, "2201.12329": { "arxivId": "2201.12329", "title": "DAB-DETR: Dynamic Anchor Boxes are Better Queries for DETR" }, "2110.06922": { "arxivId": "2110.06922", "title": "DETR3D: 3D Object Detection from Multi-view Images via 3D-to-2D Queries" }, "2102.10882": { "arxivId": "2102.10882", "title": "Conditional Positional Encodings for Vision Transformers" }, "2111.14819": { "arxivId": "2111.14819", "title": "Point-BERT: Pre-training 3D Point Cloud Transformers with Masked Point Modeling" }, "1810.11579": { "arxivId": "1810.11579", "title": "A2-Nets: Double Attention Networks" }, "2103.16302": { "arxivId": "2103.16302", "title": "Rethinking Spatial Dimensions of Vision Transformers" }, "1911.03584": { "arxivId": "1911.03584", "title": "On the Relationship between Self-Attention and Convolutional Layers" }, "2108.06152": { "arxivId": "2108.06152", "title": "Conditional DETR for Fast Training Convergence" }, "2012.00759": { "arxivId": "2012.00759", "title": "MaX-DeepLab: End-to-End Panoptic Segmentation with Mask Transformers" }, "2203.01305": { "arxivId": "2203.01305", "title": "DN-DETR: Accelerate DETR Training by Introducing Query DeNoising" }, "1904.11491": { "arxivId": "1904.11491", "title": "Local Relation Networks for Image Recognition" }, "2006.03677": { "arxivId": "2006.03677", "title": "Visual Transformers: Token-based Image Representation and Processing for Computer Vision" }, "2103.11886": { "arxivId": "2103.11886", "title": "DeepViT: Towards Deeper Vision Transformer" }, "2203.11496": { "arxivId": "2203.11496", "title": "TransFusion: Robust LiDAR-Camera Fusion for 3D Object Detection with Transformers" }, "2106.05974": { "arxivId": "2106.05974", "title": "Scaling Vision with Sparse Mixture of Experts" }, "2001.00309": { "arxivId": "2001.00309", "title": "BlendMask: Top-Down Meets Bottom-Up for Instance Segmentation" }, "2104.09224": { "arxivId": "2104.09224", "title": "Multi-Modal Fusion Transformer for End-to-End Autonomous Driving" }, "2103.11816": { "arxivId": "2103.11816", "title": "Incorporating Convolution Designs into Visual Transformers" }, "2109.08141": { "arxivId": "2109.08141", "title": "An End-to-End Transformer Model for 3D Object Detection" }, "2104.05707": { "arxivId": "2104.05707", "title": "LocalViT: Bringing Locality to Vision Transformers" }, "1902.02181": { "arxivId": "1902.02181", "title": "Attention in Natural Language Processing" }, "2202.06709": { "arxivId": "2202.06709", "title": "How Do Vision Transformers Work?" }, "2107.00641": { "arxivId": "2107.00641", "title": "Focal Self-attention for Local-Global Interactions in Vision Transformers" }, "2108.08839": { "arxivId": "2108.08839", "title": "PoinTr: Diverse Point Cloud Completion with Geometry-Aware Transformers" }, "2103.12731": { "arxivId": "2103.12731", "title": "Scaling Local Self-Attention for Parameter Efficient Visual Backbones" }, "2203.06604": { "arxivId": "2203.06604", "title": "Masked Autoencoders for Point Cloud Self-supervised Learning" }, "2109.02497": { "arxivId": "2109.02497", "title": "Voxel Transformer for 3D Object Detection" }, "2102.07074": { "arxivId": "2102.07074", "title": "TransGAN: Two Pure Transformers Can Make One Strong GAN, and That Can Scale Up" }, "2001.08248": { "arxivId": "2001.08248", "title": "How Much Position Information Do Convolutional Neural Networks Encode?" }, "2012.11409": { "arxivId": "2012.11409", "title": "3D Object Detection with Pointformer" }, "2103.03404": { "arxivId": "2103.03404", "title": "Attention is Not All You Need: Pure Attention Loses Rank Doubly Exponentially with Depth" }, "2103.15358": { "arxivId": "2103.15358", "title": "Multi-Scale Vision Longformer: A New Vision Transformer for High-Resolution Image Encoding" }, "2109.10852": { "arxivId": "2109.10852", "title": "Pix2seq: A Language Modeling Framework for Object Detection" }, "2011.10881": { "arxivId": "2011.10881", "title": "Rethinking Transformer-based Set Prediction for Object Detection" }, "2104.08541": { "arxivId": "2104.08541", "title": "TransVG: End-to-End Visual Grounding with Transformers" }, "2107.14222": { "arxivId": "2107.14222", "title": "Rethinking and Improving Relative Position Encoding for Vision Transformer" }, "2101.07448": { "arxivId": "2101.07448", "title": "Fast Convergence of DETR with Spatially Modulated Co-Attention" }, "2106.13112": { "arxivId": "2106.13112", "title": "VOLO: Vision Outlooker for Visual Recognition" }, "2104.00678": { "arxivId": "2104.00678", "title": "Group-Free 3D Object Detection via Transformers" }, "1904.05373": { "arxivId": "1904.05373", "title": "Pixel-Adaptive Convolutional Neural Networks" }, "2106.00666": { "arxivId": "2106.00666", "title": "You Only Look at One Sequence: Rethinking Transformer in Vision through Object Detection" }, "2102.10772": { "arxivId": "2102.10772", "title": "UniT: Multimodal Multitask Learning with a Unified Transformer" }, "1907.12009": { "arxivId": "1907.12009", "title": "Representation Degeneration Problem in Training Natural Language Generation Models" }, "2103.14167": { "arxivId": "2103.14167", "title": "COTR: Correspondence Transformer for Matching Across Images" }, "2105.01928": { "arxivId": "2105.01928", "title": "Instances as Queries" }, "2007.09451": { "arxivId": "2007.09451", "title": "Feature Pyramid Transformer" }, "2112.06375": { "arxivId": "2112.06375", "title": "Embracing Single Stride 3D Object Detector with Sparse Transformer" }, "2105.13677": { "arxivId": "2105.13677", "title": "ResT: An Efficient Transformer for Visual Recognition" }, "2110.09408": { "arxivId": "2110.09408", "title": "HRFormer: High-Resolution Transformer for Dense Prediction" }, "2108.10723": { "arxivId": "2108.10723", "title": "Improving 3D Object Detection with Channel-wise Transformer" }, "2104.10858": { "arxivId": "2104.10858", "title": "All Tokens Matter: Token Labeling for Training Better Vision Transformers" }, "2108.04444": { "arxivId": "2108.04444", "title": "SnowflakeNet: Point Cloud Completion by Snowflake Point Deconvolution with Skip-Transformer" }, "2203.10642": { "arxivId": "2203.10642", "title": "FUTR3D: A Unified Sensor Fusion Framework for 3D Detection" }, "2011.09315": { "arxivId": "2011.09315", "title": "End-to-End Object Detection with Adaptive Clustering Transformer" }, "2104.01318": { "arxivId": "2104.01318", "title": "Efficient DETR: Improving End-to-End Object Detector with Dense Prior" }, "2105.04553": { "arxivId": "2105.04553", "title": "Self-Supervised Learning with Swin Transformers" }, "2106.03089": { "arxivId": "2106.03089", "title": "Referring Transformer: A One-step Approach to Multi-task Visual Grounding" }, "2105.15075": { "arxivId": "2105.15075", "title": "Not All Images are Worth 16x16 Words: Dynamic Transformers for Efficient Image Recognition" }, "1603.06759": { "arxivId": "1603.06759", "title": "Convolution in Convolution for Network in Network" }, "2106.05656": { "arxivId": "2106.05656", "title": "MST: Masked Self-Supervised Transformer for Visual Representation" }, "2203.10314": { "arxivId": "2203.10314", "title": "Voxel Set Transformer: A Set-to-Set Approach to 3D Object Detection from Point Clouds" }, "2203.11183": { "arxivId": "2203.11183", "title": "Masked Discrimination for Self-Supervised Learning on Point Clouds" }, "2112.04702": { "arxivId": "2112.04702", "title": "Fast Point Transformer" }, "2107.02191": { "arxivId": "2107.02191", "title": "TransformerFusion: Monocular RGB Scene Reconstruction using Transformers" }, "2203.10981": { "arxivId": "2203.10981", "title": "MonoDTR: Monocular 3D Object Detection with Depth-Aware Transformer" }, "2111.14330": { "arxivId": "2111.14330", "title": "Sparse DETR: Efficient End-to-End Object Detection with Learnable Sparsity" }, "2106.02351": { "arxivId": "2106.02351", "title": "SOLQ: Segmenting Objects by Learning Queries" }, "2105.00637": { "arxivId": "2105.00637", "title": "ISTR: End-to-End Instance Segmentation with Transformers" }, "2001.06891": { "arxivId": "2001.06891", "title": "Where Does It Exist: Spatio-Temporal Video Grounding for Multi-Form Sentences" }, "2206.02425": { "arxivId": "2206.02425", "title": "mmFormer: Multimodal Medical Transformer for Incomplete Multimodal Learning of Brain Tumor Segmentation" }, "2011.09763": { "arxivId": "2011.09763", "title": "Attention-Based Transformers for Instance Segmentation of Cells in Microstructures" }, "2204.02174": { "arxivId": "2204.02174", "title": "Multi-View Transformer for 3D Visual Grounding" }, "2108.02388": { "arxivId": "2108.02388", "title": "TransRefer3D: Entity-and-Relation Aware Transformer for Fine-Grained 3D Visual Grounding" }, "2107.03438": { "arxivId": "2107.03438", "title": "LanguageRefer: Spatial-Language Model for 3D Visual Grounding" }, "2012.05780": { "arxivId": "2012.05780", "title": "What Makes for End-to-End Object Detection?" }, "2203.16434": { "arxivId": "2203.16434", "title": "TubeDETR: Spatio-Temporal Video Grounding with Transformers" }, "2109.07036": { "arxivId": "2109.07036", "title": "PnP-DETR: Towards Efficient Visual Analysis with Transformers" }, "2103.12957": { "arxivId": "2103.12957", "title": "Multi-view 3D Reconstruction with Transformers" }, "2011.05049": { "arxivId": "2011.05049", "title": "Human-Centric Spatio-Temporal Video Grounding With Visual Transformers" }, "2011.13118": { "arxivId": "2011.13118", "title": "Multi-view Depth Estimation using Epipolar Spatio-Temporal Networks" }, "2105.04281": { "arxivId": "2105.04281", "title": "Visual Grounding with Transformers" }, "2104.12753": { "arxivId": "2104.12753", "title": "Vision Transformers with Patch Diversification" }, "2106.03714": { "arxivId": "2106.03714", "title": "Refiner: Refining Self-attention for Vision Transformers" }, "2203.00828": { "arxivId": "2203.00828", "title": "3DCTN: 3D Convolution-Transformer Network for Point Cloud Classification" }, "2203.13310": { "arxivId": "2203.13310", "title": "MonoDETR: Depth-aware Transformer for Monocular 3D Object Detection" }, "2103.11390": { "arxivId": "2103.11390", "title": "Multi-view analysis of unregistered medical images using cross-view transformers" }, "2006.00555": { "arxivId": "2006.00555", "title": "Transferring Inductive Biases through Knowledge Distillation" }, "2101.12322": { "arxivId": "2101.12322", "title": "Position, Padding and Predictions: A Deeper Look at Position Information in CNNs" }, "2111.14725": { "arxivId": "2111.14725", "title": "Searching the Search Space of Vision Transformer" }, "2203.08481": { "arxivId": "2203.08481", "title": "Pseudo-Q: Generating Pseudo Language Queries for Visual Grounding" }, "2108.05888": { "arxivId": "2108.05888", "title": "Multiview Detection with Shadow Transformer (and View-Coherent Data Augmentation)" }, "2110.13083": { "arxivId": "2110.13083", "title": "MVT: Multi-view Vision Transformer for 3D Object Recognition" }, "2111.11704": { "arxivId": "2111.11704", "title": "Deep Point Cloud Reconstruction" }, "2211.02006": { "arxivId": "2211.02006", "title": "SAP-DETR: Bridging the Gap Between Salient Points and Queries-Based Transformer Detector for Fast Model Convergency" } }