Curriculum Vitaes

Yamashita Haruka

  (山下 遥)

Profile Information

Affiliation
Associate Professor, Faculty of Science and Technology, Department of Information and Communication Sciences, Sophia University
Degree
学士(工学)(東京理科大学)
修士(工学)(慶應義塾大学)
博士(工学)(慶應義塾大学)

Researcher number
90754797
J-GLOBAL ID
201501092433192025
researchmap Member ID
7000011989

External link

Awards

 14

Papers

 56

Misc.

 102
  • Jiajie Lu, Haruka Yamashita
    The Proceedings of ACMSA2025, Dec, 2025  Peer-reviewedCorresponding author
  • Jiajie Lu, Yamashita Haruka
    Research square, Oct 27, 2025  Corresponding author
    Abstract <p>Multimodal recommendation systems have gained increasing attention for their ability to incorporate rich side information such as visual and textual features. However, a critical yet underexplored challenge is the insufficient preservation of modality-specific information during training, which can weaken the effectiveness of multimodal signals and limit recommendation accuracy. To address this limitation, we propose Contrastive Modality-Preserving Learning (CMPL), a novel framework that extends the state-of-the-art MONET architecture. CMPL introduces a before-and-after contrastive learning module that explicitly maximizes the mutual information between initial modality embeddings and their final representations, thereby ensuring stronger modality preservation. At the same time, a graph convolutional backbone captures high-order collaborative signals from the user–item interaction graph, while a target-aware attention mechanism adaptively emphasizes user preference patterns. This joint design allows CMPL to balance the preservation of modality cues with the exploitation of collaborative filtering signals. We conduct extensive experiments on two real-world Amazon datasets, Office and MenClothing, and results consistently show that CMPL outperforms competitive baselines, including MARIO and MONET, in terms of precision and recall. These findings validate both the effectiveness of our approach and further highlight the necessity of explicitly modeling modality preservation for robust multimodal recommendation.</p>
  • 本美桃佳, 山下遥
    人工知能学会全国大会論文集(Web), 39th, 2025  
  • 伊藤諒, 山下遥
    人工知能学会全国大会論文集(Web), 39th, 2025  
  • WANG Jiayi, SHAO Tengfei, YANG Tianxiang, 山下遥, 後藤正幸
    人工知能学会全国大会論文集(Web), 39th, 2025  

Presentations

 61

Research Projects

 9