The 3rd Vision-based Remote Physiological Signal Sensing (RePSS) Challenge & Workshop
To be held at IJCAI 2024, 5th August 2024, Jeju, South Korea
Invited Speakers
Prof. Pong Chi Yuen
Hong Kong Baptist University
Title: Towards Practical Remote Photoplethysmography DetectorÂ
Abstract: Remote photoplethysmography (rPPG) is a promising enabling technology for many non-contact and non-invasive computer vision and healthcare applications, such as face presentation attack detection, heart rate and respiration rate estimation, and sleep disorder detection. While rPPG research has been performed for almost 20 years, there has been great success in the past few years. To develop such applications, one of the key factors is to develop a robust and accurate rPPG detector that is insensitive to illumination and human motion. In this talk, I will first review the development of rPPG detection and its applications. Then, I will share two recent research works on rPPG detection from my group. Finally, some future directions will be discussed.Â
Bio: Pong C. Yuen received his B.Sc. in Electronic Engineering with first-class honours in 1989 from the City Polytechnic of Hong Kong and his Ph.D. in Electrical and Electronic Engineering in 1993 from The University of Hong Kong. He joined the Hong Kong Baptist University in 1993 and served as Head of the Department of Computer Science from 2011 to 2017. Currently, he is a Chair Professor in the Department of Computer Science and Associate Dean of the Science Faculty. Dr. Yuen has held key roles in international conferences and professional communities, including serving as Associate Editor of IEEE Transactions on Information Forensics and Security and Senior Editor of the SPIE Journal of Electronic Imaging. He is a Fellow of IAPR and has received prestigious awards such as the Natural Science Awards from Guangdong Province and the Ministry of Education, China. His research interests include biometric security and privacy, video surveillance, and medical informatics.
Prof. Limin Wang
Nanjing University
Title: InternVideo: A Multimodal Foundation Model for Video UnderstandingÂ
Abstract: How to build a foundation model for video understanding has become a very challenging task. This talk mainly introduces a multimodal foundation model InternVideo and the key technologies behind it, including the unimodal video self-supervised pre-training method of VideoMAE, the multimodal video weakly supervised pre-training method of UMT, and the video-centric chat model of VideoChat. At the same time, the multimodal video dataset InternVid and the multimodal video evaluation benchmark MVBench will also be introduced. Finally, we will discuss the future trend of multimodal video understanding foundation model.Â
Bio: Limin Wang received the B.Sc. degree from Nanjing University, Nanjing, China, in 2011, and the Ph.D. degree from The Chinese University of Hong Kong, Hong Kong, in 2015. From 2015 to 2018, he was a Postdoctoral Researcher with the Computer Vision Laboratory, ETH Zürich. He is currently a Professor with the Department of Computer Science and Technology, Nanjing University. His research interests include computer vision and deep learning. He was the first runner-up at the ImageNet Large Scale Visual Recognition Challenge 2015 in scene recognition and the winner at the ActivityNet Large Scale Activity Recognition Challenge 2016 in video classification. He served as the Area Chair for CVPR, ICCV, and NeurIPS. He is on the Editorial Board of IJCV.
Prof. Zhen Lei
Institute of Automation, Chinese Academy of Sciences
Title: Fine Grained 3D Face Reconstruction from a Single ImageÂ
Abstract: 3D information is important in many computer vision tasks. This talk introduces the progress of 3D face recovery from a single image. We present a solution to capture the personalized shape so that the reconstructed shape looks identical to the corresponding person, including the data augmentation method, the many-to-one network and visual effect loss function. Furthermore, we show a method that reconstructs faces with extreme expressions. The facial part segmentation information is incorporated and a Part Re-projection Distance Loss (PRDL) is proposed to improve the reconstruction results. Experiments demonstrate good quantitative and qualitative performance in several databases.Â
Bio: Zhen Lei received the B.S. degree in automation from the University of Science and Technology of China, in 2005, and the Ph.D. degree from the Institute of Automation, Chinese Academy of Sciences, in 2010, where he is currently a professor. He is IEEE/IAPR/AAIA Fellow. He has published over 200 papers in international journals and conferences with 31000+ citations in Google Scholar and h-index 82. He was the program co-chair of IJCB2023, was competition co-chair of IJCB2022 and has served as area chairs for several conferences and is associate editor for IEEE Transactions on Information Forensics and Security, IEEE Transactions on Biometrics, Behavior, and Identity Science, Pattern Recognition, Neurocomputing and IET Computer Vision journals. His research interests are in computer vision, pattern recognition, image processing, and face recognition in particular. He is the winner of 2019 IAPR Young Biometrics Investigator Award.
Call for papers
The workshop calls for high-quality and original research works. The topic includes but is not limited to:
New methodologies and data developed for remote measurement of all kinds of physiological responses of human bodies, e.g., heart rate, heart rate variability, respiration, blood oxygen saturation, blood pressure, and others.
Hardwires/apparatuses/imaging systems developed for the purpose of remote physiological measurement from the face and body.
Solutions for special challenges involved with remote physiological signal sensing, e.g., low-quality video, poor illumination conditions, motions, data with noisy labels or no labels, etc.
Applications of remote physiological sensing, e.g., for medical assessment in hospitals, for health surveillance at home or in other environments, for emotion assessment in various scenarios like for education, job interview, etc., for security and forensics like liveness detection, biometrics, and video editing/synthesis.
Methodologies and data for the applications of all kinds of physiological signals (e.g., ECG, PPG, EEG, EDA).
Self-supervised or unsupervised methods for model training on all kinds of physiological signals.
Paper Submission Instructions
Papers must comply with the CEURART paper style (1 column) and can fall in one of the following categories:
Full research papers(minimum 7 pages)
Short research papers(4-6 pages)
The CEURART template can be found on this Overleaf link.
The review process is double-blind. Please do not include your identity information in your submitted paper.
There are two tracks (challenge track and workshop track) for the paper submission.
For the challenge track, The top-3 teams are encouraged to submit papers to this track (Acceptance depends on the quality of the paper, so it is not guaranteed.). Other teams are also welcome to submit papers. The review criterion is based on both the team ranking and the paper quality.
For the workshop track, all work in the topic scope mentioned above is encouraged to submit. The review criterion is based on the paper quality and the relevance to RePSS workshop.
Accepted papers will be included in a volume of the CEUR Workshop Proceedings (EI-index, JUFO1).
Workshop submissions will be handled by the CMT submission system; the submission link is as follows: paper submission link. All questions about submissions should be emailed to xiaobai.li@zju.edu.cn.
At least one of the authors of accepted papers should register for the IJCAI 2024 workshop and be present onsite at the workshop.
Important Date
31 May (23:59 AoE) Paper submission deadline
4 June Notification to authors
8 June Camera-ready deadline
Note: Each paper must be presented on-site by an author/co-author at the conference.