Results 51 to 60 of about 149,578 (172)
Multi- and Cross-Modal Semantics Beyond Vision: Grounding in Auditory Perception
Multi-modal semantics has relied on feature norms or raw image data for perceptual input. In this paper we examine grounding semantic representations in raw auditory data, using standard evaluations for multi-modal semantics, including measuring ...
Douwe Kiela, S. Clark
semanticscholar +1 more source
Vision-Language Models (VLMs) demonstrate significant potential for remote sensing interpretation through multimodal fusion and semantic representation of imagery.
Ziyan Wang +9 more
doaj +1 more source
Network-based mapping and neurotransmitter architecture of gray matter correlates of neuroticism
ObjectivesAlthough neuroticism is a major risk factor for adverse health outcomes, its neural basis is obscured by inconsistent findings from studies of regional gray matter volume (GMV) correlates.
Shu Wang +7 more
doaj +1 more source
"sounds like a Rainbow" - Sound-colour Mappings in Vowel Perception
The paper reports on an experiment conducted to investigate the nature of speech sound perception in terms of cross-modal mappings between vowel sound stimuli and colour spectrum associations.
M. Wrembel, K. Rataj
semanticscholar +1 more source
CM2-Net: Continual Cross-Modal Mapping Network For Driver Action Recognition
Driver action recognition has significantly advanced in enhancing driver-vehicle interactions and ensuring driving safety by integrating multiple modalities, such as infrared and depth. Nevertheless, compared to RGB modality only, it is always laborious and costly to collect extensive data for all types of non-RGB modalities in car cabin environments ...
Wang, Ruoyu +6 more
openaire +2 more sources
Cross-Modal Visual Relocalization in Prior LiDAR Maps Utilizing Intensity Textures
Cross-modal localization has drawn increasing attention in recent years, while the visual relocalization in prior LiDAR maps is less studied. Related methods usually suffer from inconsistency between the 2D texture and 3D geometry, neglecting the intensity features in the LiDAR point cloud.
Shen, Qiyuan +5 more
openaire +2 more sources
Disjoint Mapping Network for Cross-modal Matching of Voices and Faces
Tech ...
Wen, Yandong +4 more
openaire +2 more sources
Building-Guided Pseudo-Label Learning for Cross-Modal Building Damage Mapping
Accurate building damage assessment using bi-temporal multi-modal remote sensing images is essential for effective disaster response and recovery planning. This study proposes a novel Building-Guided Pseudo-Label Learning Framework to address the challenges of mapping building damage from pre-disaster optical and post-disaster SAR images.
Li, Jiepan +4 more
openaire +2 more sources
LHMap-loc: Cross-Modal Monocular Localization Using LiDAR Point Cloud Heat Map
Localization using a monocular camera in the pre-built LiDAR point cloud map has drawn increasing attention in the field of autonomous driving and mobile robotics. However, there are still many challenges (e.g. difficulties of map storage, poor localization robustness in large scenes) in accurately and efficiently implementing cross-modal localization.
Wu, Xinrui +4 more
openaire +2 more sources
Disparity Map Estimation from Cross-Modal Stereo
identifier:oai:t2r2.star.titech.ac.jp ...
openaire

