Seguir
Jeong Hun Yeo
Título
Citado por
Citado por
Año
Distinguishing homophenes using multi-head visual-audio memory for lip reading
M Kim, JH Yeo, YM Ro
Proceedings of the AAAI conference on artificial intelligence 36 (1), 1174-1182, 2022
422022
Multi-temporal lip-audio memory for visual speech recognition
JH Yeo, M Kim, YM Ro
ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and …, 2023
82023
Akvsr: Audio knowledge empowered visual speech recognition by compressing audio knowledge of a pretrained model
JH Yeo, M Kim, J Choi, DH Kim, YM Ro
IEEE Transactions on Multimedia, 2024
62024
Lip reading for low-resource languages by learning and combining general speech knowledge and language-specific knowledge
M Kim, JH Yeo, J Choi, YM Ro
Proceedings of the IEEE/CVF International Conference on Computer Vision …, 2023
62023
Towards practical and efficient image-to-speech captioning with vision-language pre-training and multi-modal tokens
M Kim, J Choi, S Maiti, JH Yeo, S Watanabe, YM Ro
ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and …, 2024
32024
Visual Speech Recognition for Languages with Limited Labeled Data Using Automatic Labels from Whisper
JH Yeo, M Kim, S Watanabe, YM Ro
ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and …, 2024
2*2024
Multilingual visual speech recognition with a single model by learning with discrete visual speech units
M Kim, JH Yeo, J Choi, SJ Park, YM Ro
arXiv preprint arXiv:2401.09802, 2024
12024
Where Visual Speech Meets Language: VSP-LLM Framework for Efficient and Context-Aware Visual Speech Processing
JH Yeo, S Han, M Kim, YM Ro
arXiv preprint arXiv:2402.15151, 2024
2024
El sistema no puede realizar la operación en estos momentos. Inténtalo de nuevo más tarde.
Artículos 1–8