WDiscOOD: Out-of-Distribution Detection via Whitened Linear Discriminant Analysis |
|
|
➖ |
Pairwise Similarity Learning is SimPLE |
|
|
➖ |
No Fear of Classifier Biases: Neural Collapse Inspired Federated Learning with Synthetic and Fixed Classifier |
|
|
➖ |
Generalizable Neural Fields as Partially Observed Neural Processes |
|
|
➖ |
M2T: Masking Transformers Twice for Faster Decoding |
➖ |
|
➖ |
Keep it SimPool: Who Said Supervised Transformers Suffer from Attention Deficit? |
|
|
➖ |
Improving Pixel-based MIM by Reducing Wasted Modeling Capability |
|
|
➖ |
Learning Image-Adaptive Codebooks for Class-Agnostic Image Restoration |
|
|
|
Quality Diversity for Visual Pre-Training |
|
|
➖ |
Subclass-Balancing Contrastive Learning for Long-Tailed Recognition |
|
|
➖ |
Mastering Spatial Graph Prediction of Road Networks |
➖ |
|
➖ |
Poincaré ResNet |
|
|
➖ |
Exploring Model Transferability through the Lens of Potential Energy |
|
|
➖ |
Improving CLIP Fine-Tuning Performance |
➖ |
|
➖ |
Unsupervised Manifold Linearizing and Clustering |
➖ |
|
➖ |
Generalized Sum Pooling for Metric Learning |
|
|
➖ |
Partition Speeds Up Learning Implicit Neural Representations based on Exponential-Increase Hypothesis |
|
|
➖ |
The Effectiveness of MAE Pre-Pretraining for Billion-Scale Pretraining |
|
|
➖ |
Token-Label Alignment for Vision Transformers |
|
|
➖ |
Efficiently Robustify Pre-Trained Models |
➖ |
|
➖ |
OFVL-MS: Once for Visual Localization Across Multiple Indoor Scenes |
|
|
➖ |
Feature Prediction Diffusion Model for Video Anomaly Detection |
➖ |
|
➖ |
Joint Implicit Neural Representation for High-Fidelity and Compact Vector Fonts |
➖ |
|
➖ |
How Far Pre-Trained Models are from Neural Collapse on the Target Dataset Informs their Transferability |
|
|
➖ |
OPERA: Omni-Supervised Representation Learning with Hierarchical Supervisions |
|
|
➖ |
Perceptual Grouping in Contrastive Vision-Language Models |
|
|
➖ |
Fully Attentional Networks with Self-Emerging Token Labeling |
➖ |
|
➖ |
Instance and Category Supervision are Alternate Learners for Continual Learning |
➖ |
|
➖ |
SkeletonMAE: Graph-based Masked Autoencoder for Skeleton Sequence Pre-Training |
|
|
➖ |
Motion-Guided Masking for Spatiotemporal Representation Learning |
➖ |
|
➖ |
Data Augmented Flatness-Aware Gradient Projection for Continual Learning |
|
|
➖ |
Take-a-Photo: 3D-to-2D Generative Pre-Training of Point Cloud Models |
|
|
➖ |
BiViT: Extremely Compressed Binary Vision Transformers |
|
|
➖ |
Spatio-Temporal Crop Aggregation for Video Representation Learning |
|
|
➖ |
Hierarchical Visual Primitive Experts for Compositional Zero-Shot Learning |
|
|
➖ |
Semantic Information in Contrastive Learning |
|
|
➖ |
Cross-Domain Product Representation Learning for Rich-Content E-Commerce |
|
|
➖ |
Contrastive Continuity on Augmentation Stability Rehearsal for Continual Self-Supervised Learning |
➖ |
|
➖ |
HybridAugment++: Unified Frequency Spectra Perturbations for Model Robustness |
|
|
➖ |
Unleashing Text-to-Image Diffusion Models for Visual Perception |
|
|
➖ |