| VLM — Sign Language and Figure-to-Caption |
| Room – Odessos Hall; Chair – Sandra Kübler |
| 11:35-12.00 | SiLVERScore: Semantically-Aware Embeddings for Sign Language Generation Evaluation (Saki Imai, Mert Inan, Anthony B. Sicilia and Malihe Alikhani) |
| 12.00-12.25 | Seeing, Signing, and Saying: A Vision-Language Model-Assisted Pipeline for Sign Language Data Acquisition and Curation from Social Media (Shakib Yazdani, Yasser Hamidullah, Cristina España-Bonet and Josef van Genabith) |
| 12.25-12.50 | FigCaps-HF: A Figure-to-Caption Generative Framework and Benchmark with Human Feedback (Ashish Singh, Ashutosh Singh, Prateek Agarwal, Zixuan Huang, Arpita Singh, Tong Yu, Sungchul Kim, Victor Soares Bursztyn, Nesreen K. Ahmed, Puneet Mathur, Erik Learned-Miller, Franck Dernoncourt and Ryan Rossi) |