| |||||||||||||||
AMFG 2018 : The 8th IEEE Workshop on Analysis and Modeling of Faces and Gestures (CVPR Workshop)Conference Series : Analysis and Modeling of Faces and Gestures | |||||||||||||||
Link: https://web.northeastern.edu/smilelab/AMFG2018/ | |||||||||||||||
| |||||||||||||||
Call For Papers | |||||||||||||||
Overview
Over the past five years, we have experienced rapid advances in facial recognition technologies. This due to many thanks to the deep learning (i.e., dating back to 2012, AlexNet) and large-scale, labeled facial image collections. The progress made in deep learning continues to push renown public facial recognition databases to near saturation which, thus, calls more evermore challenging image collections to be compiled as databases. To name a few: Labeled Faces in the Wild (LFW) database, YouTube Faces database, and, more recently, CASIA WebFace, MegaFace, MS-Celeb-1M. In practice, and even widely in applied research, using off-the-shelf deep learning models has become the norm, as numerous pre-trained networks are available for download and are readily deployed to new, unseen data (e.g., VGG-Face, ResNet, amongst other types). We have almost grown “spoiled” from such luxury, which, in all actuality, has enabled us to stay hidden from many truths. Theoretically, the truth behind what makes neural networks more discriminant than ever before is still, in all fairness, unclear—rather, they act as a sort of black box to most practitioners and even researchers, alike. More troublesome is the absence of tools to quantitatively and qualitatively characterize existing deep models, which, in itself, could yield greater insights about these all so familiar “black boxes”. With the frontier moving forward at rates incomparable to any spurt of the past, challenges such as high variations in illuminations, pose, age, etc., now confront us. However, state-of-the-art deep learning models often fail when faced with such challenges owed to the difficulties in modeling structured data and visual dynamics. Alongside the effort spent on conventional face recognition is the research done to automatically understand social media content. This line of work has attracted attention from industry and academic researchers from all sorts of domains. To understand social media the following capabilities must be satisfied: face and body tracking (e.g., facial expression analysis, face detection, gesture recognition), face and body characterization (e.g., behavioral understanding, emotion recognition), face, body and gesture characteristic analysis (e.g., gait, age, gender and ethnicity recognition), group understanding via social cues (e.g., kinship, non-blood relationships, personality), and visual sentiment analysis (e.g., temperament, arrangement). Thus, needing to be able to create effective models for visual certainty has significant value in both the scientific communities and the commercial market, with applications that span topics of human-computer interaction, social media analytics, video indexing, visual surveillance, and Internet vision. Currently, researchers have made significant progress addressing the many problems in the social domain, and especially when considering off-the-shelf and cost-efficient vision HW products available these days, e.g. Kinect, Leap, SHORE, and Affdex. Nonetheless, serious challenges still remain, which only amplifies when considering the unconstrained imaging conditions captured by different sources focused on non-cooperative subjects. It is these latter challenges that especially grabs our interest, as we sought out to bring together the cutting-edge techniques and recent advances in deep learning to solve the challenges above in social media. Previous AMFG Workshops The first workshop with this name was held in 2003, in conjunction with ICCV2003 in Nice, France. So far, it has been successfully held SEVEN times. The homepages of previous five AMFG are as follows: AMFG2003: http://brigade.umiacs.umd.edu/iccv2003/ AMFG2005: http://mmlab.ie.cuhk.edu.hk/iccv05/ AMFG2007: http://mmlab.ie.cuhk.edu.hk/iccv07/ AMFG2010: http://www.lv-nus.org/AMFG2010/cfp.html AMFG2013: http://www.northeastern.edu/smilelab/AMFG2013/home.html AMFG2015: http://www.northeastern.edu/smilelab/AMFG2015/home.html AMFG2017: https://web.northeastern.edu/smilelab/AMFG2017/index.html List of Topics -- Novel deep model, deep learning survey, or comparative study for face/gesture recognition; -- Deep learning methodology, theory, and its application to social media analytics; -- Deep learning for internet-scale soft biometrics and profiling: age, gender, ethnicity, personality, kinship, occupation, beauty ranking, and fashion classification by facial or body descriptor; -- Deep learning for detection and recognition of faces and bodies with large 3D rotation, illumination change, partial occlusion, unknown/changing background, and aging (i.e., in the wild); special attention will be given large 3D rotation robust face and gesture recognition; -- Motion analysis, tracking, and extraction of face and body models captured by mobile devices; -- Face, gait, and action recognition in low-quality (e.g., blurred), or low-resolution video from fixed or mobile device cameras; -- Novel mathematical models and algorithms, sensors and modalities for face & body gesture and action representation, analysis, and recognition for cross-domain social media; -- Social/psychological based studies that aids in understanding computational modeling and building better-automated face and gesture systems with interactive features; -- Novel social applications involving detection, tracking & recognition of face, body, and action; -- Face and gesture analysis for sentiment analysis in social media; -- Other applications involving face and gesture analysis in social media content. Contact Ming Shao (mshao@umassd.edu) Computer and Information Science, University of Massachusetts Dartmouth, Dartmouth, MA, USA Joseph Robinson (robinson.jo@husky.neu.edu) Department of Electrical and Computer Engineering, Northeastern University, Boston, MA, USA |
|