Program

CVM 2021 Programme

Download PDF conference programme

     
Day One   Wednesday, 21 April 2021
   
09:00 - 09:20 Opening Session
   
09:20 - 10:05 Keynote 1
 Lingqi Yan (UC Santa Barbara)
 Visual Appearance Synthesis: Challenges and Opportunities
   
   
 Industrial Session
10:05 - 10:20 Xi Wang (Booming Technology)
 From Game Engine to Matrix Architecture
10:20 - 10:35 Haozhi Huang (XVerse)
 Rethinking the Tech Roadmap to a Thriving Metaverse
10:35 - 10:50 Wenqiang Liu (Eazao)
 Mass Production for Personalized Ceramics Additive Manufacturing
   
10:50 - 11:05   Coffee Break
   
Session 1   Image Synthesis
11:05 - 11:25 Yuan Chang, Congyi Zhang, Yisong Chen, Guoping Wang
 Homography Guided Optimization for Motion Assisted Wide-Baseline Image Interpolation
11:25- 11:45 Yu Song, Fan Tang, Weiming Dong, Changsheng Xu
 Non-dominated Sorting based Multi-page Photo Collage
11:45 - 12:05 Zuyi Yang, Qinghui Dai, Junsong Zhang
 Visual Perception Driven Picture Collage Synthesis
   
12:15 - 13:40   Lunch
   
Session 2   Visual Localization and Navigation
13:40 - 14:00 Huixuan Wang, Changhe Tu, Xueying Qin
 ReLoc: Indoor visual localization with a hierarchical sitemap and view synthesis
14:00 - 14:20 Xinyu Zheng, Chen Zong, Jingliang Cheng, Shiqing Xin, Shuangmin Chen, Changhe Tu, Wenping Wang
 Visually Smooth Multi-UAV Formation Transformation
14:20 - 14:40 Yilin Liu, Di Lin, Ke Xie, Hui Huang
 VGF-Net: Visual-Geometric Fusion Learning for Simultaneous Drone Navigation and Height Mapping
   
Session 3   Geometric Algorithm
14:40 - 15:00 Qianwei Xia, Juyong Zhang, Zheng Fang, Jin Li, Bailin Deng, Ying He
 GeodesicEmbedding (GE): A High-Dimensional Embedding Approach for Fast Geodesic Distance Query
15:00 - 15:20 Bo-Yi Hu, Chunyang Ye, Jian-Ping Su, Ligang Liu
 Reference-constrained Geometric Optimization
15:20 - 15:40 Chuanfeng Hu, Hongwei Lin
 Heterogeneous Porous Scaffold Generation Using Trivariate B-spline Solids and Triply Periodic Minimal Surfaces
   
15:40 - 15:55   Coffee Break
   
Session 4   Visualization and Simulation
15:55 - 16:15 Ji Lan, Zheng Zhou, Jiachen Wang, Hui Zhang, Yingcai Wu
 Visual Analytics of Impacts of Behavior Adjustments on Scoring Rates in Table Tennis
16:15 - 16:35 Xiaopeng Sun, Jia Fu, Yu Dong, Teng Chen
 Wilting Deformation of Leaves Using Cell Dynamics and Time-Varying External Forces
16:35 - 16:55 Hongyang Zhou, Zhong Ren
 Adaptive Geometric Sound Propagation Based on A-weighting Variance Measure
   
16:55 - 17:55   Poster Session
 Zheng-Ning Liu, Yan-Pei Cao, Yu-Tao Yuan, Tai-Jiang Mu, Shi-Min Hu
  Semi-Supervised Learning of Disentangled Representations for Cross-Modal Translation
 Yu He, Guodong Zhao, Songhai Zhang, Shi-Min Hu
 Smoothness Preserving Layout for Dynamic Labels by Hybrid Optimization
 Yanlong Tang, Yun Zhang, Xiaoguang Han, Fang-Lue Zhang, Yu-Kun Lai
 3D Corrective Nose Reconstruction from a Single Image
 Xiaofeng Zhang, Xin Gao, Yan Zhang, Yujuan Sun, Caiming Zhang
 Improved Fuzzy Clustering Algorithm for Image Segmentation Based on Low-rank Prior
 Haoxuan Song, Jiahui Huang, Yanpei Cao, Shi-Min Hu
 HDR-Net-Fusion: Real-time 3D Dynamic Scene Reconstruction with a Hierarchical Deep Reinforcement Network
 Xiang Ma, Xuemei Li, Jiaye Wang, Yuanfeng Zhou, Caiming Zhang
 Image Smoothing Based on Global Sparsity Decomposition and Variable Parameter
 Dong Chen, Fan Tang, HanXing Yao, WeiMing Dong, Changsheng Xu
 SiamCPN: Visual Tracking with Siamese-based Center Prediction Network
 Renjie Chen, Craig Gotsman
 Efficient Fastest-Path Computations in Road Maps
 Yunai Yi, Diya Sun, Tae-Kyun Kim, Yuru Pei
 Unsupervised Random Forest for Affinity Estimation
   
19:00 - 21:30   CVM Workshop on Jittor
   
     
Day Two   Thursday, 22 April 2021
   
Session 5   Geometry and Texture
09:00 - 09:20 Shao-Kui Zhang, Wei-Yu Xie, Song-Hai Zhang
 Geometry-Based Layout Generation with Hyper-Relations AMONG Objects
09:20 - 09:40 Hongrui Cai, Yudong Guo, Zhuang Peng, Juyong Zhang
 Landmark Detection and 3D Face Reconstruction for Caricature using a Nonlinear Parametric Model
09:40 - 10:00 Sheldon Taylor, Owen Sharpe, Jiju Peethambaran
 Prime Gradient Noise
   
Session 6   Multiview Stereo
10:00 - 10:20 Qitong Zhang, Shan Luo, Lei Wang, Jieqing Feng
 CNLPA-MVS: Coarse-Hypotheses Guided Non-Local PAtchMatch Multi-View Stereo
10:20 - 10:40 Yakun Ju, Yuxin Peng, Muwei Jian, Feng Gao, Junyu Dong
 Learning Conditional Photometric Stereo with High-resolution Features
   
10:40 - 10:55   Coffee Break
   
Session 7   Motion Transfer and Prediction
10:55 - 11:15 Jinfeng Jiang, Guiqing Li, Shihao Wu, Huiqian Zhang
 BPA-GAN: Human Motion Transfer Using Body-Parts-Aware Generative Adversarial Networks
11:15 - 11:35 Qian Zheng, Weikai Wu, Hanting Pan, Niloy Mitra, Daniel Cohen-Or, Hui Huang
 Object Properties Inferring from and Transfer for Human Interaction Motions
11:35 - 11:55 Pei Lv, Hui Wei, Yuzhen Zhang, Xiaoheng Jiang, Tianxin Gu, Bing Zhou, Mingliang Xu
 Probability Trajectory: One New Movement Description for Trajectory Prediction
   
12:15 - 13:40   Lunch
   
Session 8   Detection and Recognition
13:40 - 14:00 Beiji Zou, Wenjun Yang, Kaiwen Li, Shu Liu
 A Character Flow Framework for Multi-oriented Scene Text Detection
14:00 - 14:20 Yang Liu, Ruili He, Wei Wang, Shengping Zhang
 Is it easy to recognize baby's age and gender?
14:20 - 14:40 Xinyu Li, Guangshun Wei, Jie Wang, Yuanfeng Zhou
 Multi-scale Feature Joint Network for Micro-expression Recognition
14:40 - 15:00 Zeyu Shen, Mingyang Zhao, Xiaohong Jia, Dong-Ming Yan
 Combining Convex Hull and Directed Graph for Fast and Accurate Ellipse Detection
   
Session 9   Image Understanding
15:00 - 15:20 Lixue Gong, Yiqun Zhang, Yunke Zhang, Ying Yang, Weiwei Xu
 Erroneous Pixel Prediction for Semantic Image Segmentation
15:20 - 15:40 Jin Zhou, Qing Zhang, Jian-Hao Fan, Wei Sun, Wei-Shi Zheng
 Joint Regression and Learning from Pairwise Rankings for Personalized Image Aesthetic Assessment
   
15:40 - 15:55   Coffee Break
   
Session 10   Face Art and Visual Exploration
15:55 - 16:15 Chenxu Zhang, Saifeng Ni, Zhipeng Fan, Hongbo Li, Ming Zeng, Madhukar Budagavi, Xiaohu Guo
 3D Talking Face with Personalized Pose Dynamics
16:15 - 16:35 Zhengyu Huang, Yichen Peng, Tomohiro Hibino, Chunqi Zhao, Haoran Xie, Tsukasa Fukusato, Kazunori Miyata
 dualFace: Two-Stage Drawing Guidance for Freehand Portrait Sketching
16:35 - 16:55 Ruoqi Sun, Chen Huang, Hengliang Zhu, Lizhuang Ma
 Mask-ware Photorealistic Face Attribute Manipulation
16:55 - 17:15 Yifang Wang, Hongye Liang, Xinhuan Shu, Jiachen Wang, Ke Xu, Zikun Deng, Cameron Campbell, Bijia Chen, Yingcai Wu, Huamin Qu
 Interactive Visual Exploration of Longitudinal Historical Career Mobility Data
   
17:15 - 18:00 Keynote 2
  Dani Lischinski (The Hebrew University of Jerusalem)
 Deep Learning for Graphics - A Personal Perspective
   
18:30 - 21:30   Banquet Dinner
     
Day Three   Friday, 23 April 2021
   
09:00 - 09:45 Keynote 3
  Lin Gao (Chinese Academy of Sciences)
 Deep Structural Generative Model for Visual Media Analysis and Synthesis: From Deep Face to Deformable Model
   
Session 11   3D Feature and Correspondence
09:45 - 10:05 Wenkai Han, Hai Wu, Luwen Cheng, Cheng Wang, Xin Li
 BLNet: Bidirectional Learning Network for Point Clouds
10:05 - 10:25 Jiachen Li, Fan Zhong, Songhua Xu, Xueying Qin
 3D Object Tracking with Adaptively Weighted Local Bundles
10:25 - 10:45 Xiao-Ming Fu, Jian-Ping Su, Zheng-Yu Zhao, Qing Fang, Chunyang Ye, Ligang Liu
 Inversion-free Geometric Mapping Construction: A Survey
   
10:45 - 11:00   Coffee Break
   
Session 12   Styles
11:00 - 11:20 Lan Chen, Juntao Ye, Xiaopeng Zhang
 Multi-Feature Super-Resolution Network for Cloth Wrinkle Synthesis
11:20 - 11:40 Songhua Xu, Yechuan Tian
 A Novel Three-Staged Generative Model for Skeletonizing Chinese Characters in Versatile Styles
11:40 - 12:00 Min Shi, Yukun Wei, Lan Chen, Dengming Zhu, Tianlu Mao, Zhaoqi Wang
 Learning a Shared Deformation Space for Efficient Design-Preserving Garment Transfer
   
12:15 - 13:40   Lunch
   
Session 13   Image Processing
13:40 - 14:00 Xueting Liu, Wenliang Wu, Chengze Li, Yifan Li, Huisi Wu
 Reference-Guided Structure-Aware Deep Sketch Colorization for Cartoons
14:00 - 14:20 Tianlin Zhang, Jinjiang Li, Hui Fan
 Progressive edge-sensing dynamic scene deblurring
14:20 - 14:40 Qi Mu, Xinyue Wang, Yanyan Wei, Zhanli Li
 Color Image Enhancement Algorithm Based on Weighted Guided Image Filtering Under Low and Non-uniform Illumination
   
Session 14   Video Warping and Interpolation
14:40 - 15:00 Jinbo Xing, Wenbo Hu, Yuechen Zhang, Tien-Tsin Wong
 Flow-aware Synthesis: A Generic Motion Model for Video Frame Interpolation
15:00 - 15:20 Jinliang Wu, Junjie Shi, Lei Zhang
 Rectangling Irregular Videos by Optimizing Spatio-Temporal Warping
15:20 - 15:40 Yudong Guo, Juyong Zhang, Yihua Chen, Hongrui Cai, Zhangjin Huang, Bailin Deng
 Real-Time Face View Correction for Front-Facing Cameras
   
15:40 - 15:55   Coffee Break
   
15:55 - 16:55 Forum: AI Enabled Computational Visual Media
 Shi-Min Hu (Tsinghua University), Baoquan Chen (Peking University), Ligang Liu (University of Science and Technology of China), Yebin Liu (Tsinghua University), Zhouhui Lian (Peking University), Yunhai Wang (Shandong University)
   
16:55 - 17:25  Closing Session
     
     

Keynote Speakers

Lingqi Yan, University of California, Santa Barbara

Title:

Visual Appearance Synthesis: Challenges and Opportunities

Abstract:

Visual appearance rendering is self-contradictory. Traditional rendering overly simplifies the appearance, leading to loss of details. Complex appearance rendering explicitly models details ranging from micron-level scratches and bumps to real-world-level hair strands and foliage. But they are often too costly to store, and too slow for real-time applications, both due to the huge amount of appearance data needed to be generated, recorded and processed. We want to eliminate this trade-off, generating and rendering visual appearance as accurate as "photorealistic", and fast enough even for real-time applications. To realize this goal, in this talk, I will introduce three key properties to consider: dynamic (on-the-fly synthesis as being queried), scalable (level-of-detail aggregation at zero additional storage cost) and fast (simultaneous syntheses in a massive amount). I will provide a brief view on the synthesis of visual appearance, analyzing its unique challenges and the failure of image synthesis approaches, and providing tentative procedural, example-based solutions, so we can create visual appearance of arbitrary complexity and realism, given just a series of random numbers, a measured block of fabrics, or a strand of hair.

Speaker's Biography:

Lingqi Yan is an Assistant Professor of Computer Science at UC Santa Barbara, co-director of the MIRAGE Lab, and affiliated faculty in the Four Eyes Lab. He received his Ph.D. degree from the Department of Electrical Engineering and Computer Sciences at UC Berkeley in 2018, advised remotely by Prof. Ravi Ramamoorthi at UC San Diego. His research is in Computer Graphics, mainly including appearance modeling, real-time ray tracing, sampling and reconstruction theory, volumetric scattering and light transport algorithms. During his Ph.D. study, he has published a record-breaking number of 7 first authored ACM SIGGRAPH/ACM TOG papers. And he won the SIGGRAPH Outstanding Doctoral Dissertation Award in 2019.


Dani Lischinski, The Hebrew University of Jerusalem

Title:

Deep Learning for Graphics - A Personal Perspective

Abstract:

Starting from the success of AlexNet in 2012, deep neural networks have taken over the field of computer vision by storm. In contrast, the adoption of deep learning machinery by the Computer Graphics community has been much slower. Nevertheless, today deep networks can be found in all the major areas of Computer Graphics, including first and foremost image and video processing, but also animation, rendering, and 3D modeling. In this talk, I will give a brief overview of my personal perspective of this trend, focusing on several exciting advances ranging from texture and image synthesis to motion style transfer for articulated characters. All of these results are enabled by the use of deep networks, demonstrating their potential for graphics, which the community has only begun to explore.

Speaker's Biography:

Dani Lischinski is a Professor at the School of Computer Science and Engineering at the Hebrew University of Jerusalem, Israel. He received his PhD from Cornell University in 1994, and was a post-doctoral research associate at the University of Washington until 1996. In 2002/3, he spent a sabbatical year at Pixar Animation Studios. In 2012 he received the Eurographics Outstanding Technical Contributions Award. In 2017, he served as the Technical Papers Chair for SIGGRAPH Asia 2017. His areas of interest span a wide variety of topics in the fields of computer graphics, image and video processing, and computer vision. Most of his recent work involves deep neural networks and their applications in graphics and vision.


Lin Gao, Chinese Academy of Sciences

Title:

Deep Structural Generative Model for Visual Media Analysis and Synthesis: From Deep Face to Deformable Model

Abstract:

Deep geometry learning has become a hot topic in recent years and is widely used to analyze and generate various kinds of visual media. At present, its main challenge is how to improve the generation quality of visual media and reduce the dependence on 3D annotation data. To deal with these challenges, we utilize the structural information of visual media, design the deep generative model according to the internal structure of visual media, and propose unsupervised 3D model analysis methods. Specifically, for facial image generation, we developed a deep generation network, which encodes facial images using the structure information, and use this network to translate the sketches to high-quality face images. And then we developed an intelligent face drawing system, named DeepFaceDrawing, based on the deep learning framework Jittor. For the CAD model, according to the robust supporting structure of the artificial model, we proposed a new deep generative model named SDM-NET, which is able to encode the supporting structure and rich details of 3D model. Furthermore, we proposed a deep generation network named DSG-NET, which encodes the geometric details and complex structures jointly and hierarchically. In addition, we also study how to use the global symmetry information to analyze symmetry of 3D model in the unsupervised manner.

Speaker's Biography:

Lin Gao is an Associate Professor (PhD supervisor) with Institute of Computing Technology, Chinese Academy of Sciences, also the director of the Intelligent Graphics Laboratory (IGL), IDEI. He received his Ph.D. from Tsinghua University in 2014 advised by Prof. Shi-Min Hu. He was a visiting professor in Prof. Leif Kobbelt group in RWTH Aachen University in 2016. His research primarily centers on algorithmic issues in geometry learning, mesh processing, 3d computer vision. He was selected into Royal Society-Newton Advanced Fellowship, Youth Innovation Promotion Association of Chinese Academy of Sciences, Young Elite Scientists Sponsorship Program of China Association for Science and Technology and so on. He won the Asia Graphics Young Researcher Award in 2020.