- Exposé par Zied Bouyahia : "Towards an integrated framework for smart mobility: methods, applications and challenges"
- Exposé par Di Huang : "Multi-modal Visual Perception"
- Questions diverses
<!--
...
...
@@ -24,11 +23,16 @@
-->
<!--_Titre :_-->
<!--
_Résumé :_
A venir
-->
With the rapid development of the internet and sensing technology, different modalities of data are emerging at an unprecedented speed. Multimodal data convey more comprehensive clues, but due to significant differences in their underlying mechanisms and forms of expression, and the complexity of the coupling relation-ships between modalities, how to utilize and mine the complementary information of multi-modal data has become an important research direction in the field. This talk focuses on visual perception based on multimodal data and introduces the recent research work of the team of the presenter in this direction. The key discussion point is mainly on the fusion of texture and geometric information, involving data collected from various types of devices such as image sensors, depth sensors, LIDAR, for diverse tasks, including, facial analysis, robotic grasping, object detection, with the applications in industry, transportation, and other related areas.
_Biographie :_
Di Huang received the B.S. and M.S. degrees in computer science from Beihang University, Beijing, China, in 2005 and 2008, respectively, and the Ph.D. degree in computer science from the École Centrale de Lyon, Lyon, France, in 2011. He joined the Laboratory of Intelligent Recognition and Image Processing, School of Computer Science and Engineering, Beihang University, as a Faculty Member, where he is cur-rently a Professor. His research interests include computer vision, representation learning, robotics, etc. He has published 100+ papers in the major academic journals and conferences such as IEEE-TPAMI, IJCV, CVPR, ICCV, ECCV, with 12,000+ ci-tations and H-index 50. He has been among the top 2% world-wide AI scientist list since released in 2019 (by Stanford University and Elsevier). He was awarded the Best Paper at CCBR 2016 and AMFG 2017, the Best Student Paper at CCBR 2017, the Best Poster Paper at ICB 2016, and the Honorable Mention for Best Student Paper at FG 2024. He was the winner of the Affect Recognition Challenge of AVEC at MM 2013 and the OCRTOC track in Robotic Grasping and Manipulation Competitions at ICRA 2022.He served as a Guest Editor for ACM-TOMM, an Area Chair/Senior PC Member for CVPR 2022/2024-2025, ECCV 2022, MM 2019-2023, IJCAI 2021/2025, WACV 2024-2025, ICPR 2020/2024, IJCB 2021/2023-2024, ICMI 2021, ACII 2019-2023, and a Publicity Chair for FG 2019, IJCB 2020, and FG 2023.
<!-- A venir -->
<!--
Occlusions are very common in face images in the wild, leading to the degraded performance of face-related tasks. Although much effort has been devoted to removing occlusions from face images, the varying shapes and textures of occlusions still challenge the robustness of current methods. As a result, current methods either rely on manual occlusion masks or only apply to specific occlusions. This paper proposes a novel face de-occlusion model based on face segmentation and 3D face reconstruction, which automatically removes all kinds of face occlusions with even blurred boundaries,e.g., hairs. The proposed model consists of a 3D face reconstruction module, a face segmentation module, and an image generation module. With the face prior and the occlusion mask predicted by the first two, respectively, the image generation module can faithfully recover the missing facial textures. To supervise the training, we further build a large occlusion dataset, with both manually labeled and synthetic occlusions. Qualitative and quantitative results demonstrate the effectiveness and robustness of the proposed method.