no code implementations • 21 Apr 2024 • Xinxin Jiao, Liejun Wang, Yinfeng Yu
This paper introduces MFHCA, a novel method for Speech Emotion Recognition using Multi-Spatial Fusion and Hierarchical Cooperative Attention on spectrograms and raw audio.
no code implementations • 9 Oct 2023 • Yinfeng Yu, Changan Chen, Lele Cao, Fangkai Yang, Fuchun Sun
As humans, we hear sound every second of our life.
no code implementations • 4 Oct 2022 • Yinfeng Yu, Lele Cao, Fuchun Sun, Xiaohong Liu, Liejun Wang
Audio-visual embodied navigation, as a hot research topic, aims training a robot to reach an audio target using egocentric visual (from the sensors mounted on the robot) and audio (emitted from the target) input.
1 code implementation • ICLR 2022 • Yinfeng Yu, Wenbing Huang, Fuchun Sun, Changan Chen, Yikai Wang, Xiaohong Liu
In this work, we design an acoustically complex environment in which, besides the target sound, there exists a sound attacker playing a zero-sum game with the agent.