Learned Codecs, AI Compression, Rate-Distortion Theory, Entropy Models
Extracting Multimodal Learngene in CLIP: Unveiling the Multimodal Generalizable Knowledge
arxiv.org·2d
GLIMPSE: Gradient-Layer Importance Mapping for Prompted Visual Saliency Explanation for Generative LVLMs
arxiv.org·18h
AMF-MedIT: An Efficient Align-Modulation-Fusion Framework for Medical Image-Tabular Data
arxiv.org·18h
Episode-specific Fine-tuning for Metric-based Few-shot Learners with Optimization-based Training
arxiv.org·1d
ABC: Adaptive BayesNet Structure Learning for Computational Scalable Multi-task Image Compression
arxiv.org·6d
SLAP: Siamese Language-Audio Pretraining Without Negative Samples for Music Understanding
arxiv.org·1d
Loading...Loading more...