default search action
SIGGRAPH Asia 2022 Conference Papers: Daegu, Korea
- Soon Ki Jung, Jehee Lee, Adam W. Bargteil:
SIGGRAPH Asia 2022 Conference Papers, SA 2022, Daegu, Republic of Korea, December 6-9, 2022. ACM 2022, ISBN 978-1-4503-9470-3
Character Animation
- Zhan Xu, Yang Zhou, Li Yi, Evangelos Kalogerakis:
Morig: Motion-Aware Rigging of Character Meshes from Point Clouds. 1:1-1:9 - Alexander W. Winkler, Jungdam Won, Yuting Ye:
QuestSim: Human Motion Tracking from Sparse Sensors with Simulated Avatars. 2:1-2:8 - Yifeng Jiang, Yuting Ye, Deepak Gopinath, Jungdam Won, Alexander W. Winkler, C. Karen Liu:
Transformer Inertial Poser: Real-time Human Motion Reconstruction from Sparse IMUs with Simultaneous Terrain Generation. 3:1-3:9
Distances and Matching
- Craig Gotsman, Kai Hormann:
Compressing Geodesic Information for Fast Point-to-Point Geodesic Distance Queries. 4:1-4:9
Differentiable Rendering
- Sai Praveen Bangaru, Michaël Gharbi, Fujun Luan, Tzu-Mao Li, Kalyan Sunkavalli, Milos Hasan, Sai Bi, Zexiang Xu, Gilbert Bernstein, Frédo Durand:
Differentiable Rendering of Neural SDFs through Reparameterization. 5:1-5:9 - Jingsen Zhu, Fujun Luan, Yuchi Huo, Zihao Lin, Zhihua Zhong, Dianbing Xi, Rui Wang, Hujun Bao, Jiaxiang Zheng, Rui Tang:
Learning-based Inverse Rendering of Complex Indoor Scenes with Differentiable Monte Carlo Raytracing. 6:1-6:8 - Qiang Zhang, Seung-Hwan Baek, Szymon Rusinkiewicz, Felix Heide:
Differentiable Point-Based Radiance Fields for Efficient View Synthesis. 7:1-7:12
Image Generation
- Seongtae Kim, Kyoungkook Kang, Geonung Kim, Seung-Hwan Baek, Sunghyun Cho:
DynaGAN: Dynamic Few-shot Adaptation of GANs to Multiple Domains. 8:1-8:8 - Wonjoon Jin, Nuri Ryu, Geonung Kim, Seung-Hwan Baek, Sunghyun Cho:
Dr.3D: Adapting 3D GANs to Artistic Drawings. 9:1-9:8
Acquisition
- Zhiyi Kuang, Yiyang Chen, Hongbo Fu, Kun Zhou, Youyi Zheng:
DeepMVSHair: Deep Hair Modeling from Sparse Views. 10:1-10:8
Radiance Fields, Bases, and Probes
- Jiemin Fang, Taoran Yi, Xinggang Wang, Lingxi Xie, Xiaopeng Zhang, Wenyu Liu, Matthias Nießner, Qi Tian:
Fast Dynamic Radiance Fields with Time-Aware Neural Voxels. 11:1-11:9 - Jingbo Zhang, Xiaoyu Li, Ziyu Wan, Can Wang, Jing Liao:
FDNeRF: Few-shot Dynamic Neural Radiance Fields for Face Reconstruction and Expression Editing. 12:1-12:9 - Quewei Li, Jie Guo, Yang Fei, Feichao Li, Yanwen Guo:
NeuLighting: Neural Lighting for Free Viewpoint Outdoor Scene Relighting with Unconstrained Photo Collections. 13:1-13:9 - Zilin Xu, Zheng Zeng, Lifan Wu, Lu Wang, Ling-Qi Yan:
Lightweight Neural Basis Functions for All-Frequency Shading. 14:1-14:9
Stylization and Colorization
- Michal Kucera, David Mould, Daniel Sýkora:
StyleBin: Stylizing Video by Example in Stereo. 15:1-15:8
Face, Speech, and Gesture
- Byungkuk Choi, Haekwang Eom, Benjamin Mouscadet, Stephen Cullingford, Wan-Duo Kurt Ma, Stefanie Gassel, Suzi Kim, Andrew Moffat, Millicent Maier, Marco Revelant, Joe Letteri, Karan Singh:
Animatomy: an Animator-centric, Anatomically Inspired System for 3D Facial Modeling, Animation and Transfer. 16:1-16:9 - Yasheng Sun, Hang Zhou, Kaisiyuan Wang, Qianyi Wu, Zhibin Hong, Jingtuo Liu, Errui Ding, Jingdong Wang, Ziwei Liu, Hideki Koike:
Masked Lip-Sync Prediction by Audio-Visual Contextual Exploitation in Transformers. 17:1-17:9 - Yifang Pan, Chris Landreth, Eugene Fiume, Karan Singh:
VOCAL: Vowel and Consonant Layering for Expressive Animator-Centric Singing Animation. 18:1-18:9 - Jordan Juravsky, Yunrong Guo, Sanja Fidler, Xue Bin Peng:
PADL: Language-Directed Physics-Based Character Control. 19:1-19:9
Perception in VR and AR
- Joohwan Kim, Arjun Madhusudan, Benjamin Watson, Ben Boudaoud, Roland Tarrazo, Josef B. Spjut:
Display Size and Targeting Performance: Small Hurts, Large May Help. 20:1-20:8 - Nathan Matsuda, Alexandre Chapiro, Yang Zhao, Clinton Smith, Romain Bachy, Douglas Lanman:
Realistic Luminance in VR. 21:1-21:8 - Joseph March, Anantha Krishnan, Simon J. Watt, Marek Wernikowski, Hongyun Gao, Ali Özgür Yöntem, Rafal Mantiuk:
Impact of correct and simulated focus cues on perceived realism. 22:1-22:9
Faces and Avatars
- Shen Sang, Tiancheng Zhi, Guoxian Song, Minghao Liu, Chun-Pong Lai, Jing Liu, Xiang Wen, James Davis, Linjie Luo:
AgileAvatar: Stylized 3D Avatar Creation via Cascaded Domain Bridging. 23:1-23:8
Shape Generation
- Ka-Hei Hui, Ruihui Li, Jingyu Hu, Chi-Wing Fu:
Neural Wavelet-domain Diffusion for 3D Shape Generation. 24:1-24:9 - Nasir Mohammad Khalid, Tianhao Xie, Eugene Belilovsky, Tiberiu Popa:
CLIP-Mesh: Generating textured meshes from text using pretrained image-text models. 25:1-25:8 - Sifan Ye, Yixing Wang, Jiaman Li, Dennis Park, C. Karen Liu, Huazhe Xu, Jiajun Wu:
Scene Synthesis from Human Motion. 26:1-26:9 - Kurt Leimer, Paul Guerrero, Tomer Weiss, Przemyslaw Musialski:
LayoutEnhancer: Generating Good Indoor Layouts from Imperfect Data. 27:1-27:8
Reconstruction and Repair
- Bowen Zhang, Xi Zhao, He Wang, Ruizhen Hu:
Shape Completion with Points in the Shadow. 28:1-28:9
Image Editing and Manipulation
- Rotem Tzaban, Ron Mokady, Rinon Gal, Amit Bermano, Daniel Cohen-Or:
Stitch it in Time: GAN-Based Facial Editing of Real Videos. 29:1-29:9 - Kun Cheng, Xiaodong Cun, Yong Zhang, Menghan Xia, Fei Yin, Mingrui Zhu, Xuan Wang, Jue Wang, Nannan Wang:
VideoReTalking: Audio-based Lip Synchronization for Talking Head Video Editing In the Wild. 30:1-30:9 - Kaiwen Jiang, Shu-Yu Chen, Feng-Lin Liu, Hongbo Fu, Lin Gao:
NeRFFaceEditing: Disentangled Face Editing in Neural Radiance Fields. 31:1-31:9 - Ryusuke Sugimoto, Mingming He, Jing Liao, Pedro V. Sander:
Water Simulation and Rendering from a Still Photograph. 32:1-32:9
Appearance Modeling and Capture
- Wenhua Jin, Beibei Wang, Milos Hasan, Yu Guo, Steve Marschner, Ling-Qi Yan:
Woven Fabric Capture from a Single Photo. 33:1-33:8 - Xilong Zhou, Milos Hasan, Valentin Deschaintre, Paul Guerrero, Kalyan Sunkavalli, Nima Khademi Kalantari:
TileGen: Tileable, Controllable Material Generation and Capture. 34:1-34:9 - Bin Chen, Michal Piovarci, Chao Wang, Hans-Peter Seidel, Piotr Didyk, Karol Myszkowski, Ana Serrano:
Gloss management for consistent reproduction of real and virtual objects. 35:1-35:9
Maps and Fields
- Xingyi Du, Danny M. Kaufman, Qingnan Zhou, Shahar Z. Kovalsky, Yajie Yan, Noam Aigerman, Tao Ju:
Isometric Energies for Recovering Injectivity in Constrained Mapping. 36:1-36:9 - Yuta Noma, Nobuyuki Umetani, Yoshihiro Kawahara:
Fast Editing of Singularities in Field-Aligned Stripe Patterns. 37:1-37:8 - Jean-Marc Thiery, Tamy Boubekeur:
Green Coordinates for Triquad Cages in 3D. 38:1-38:8 - Haotong Lin, Sida Peng, Zhen Xu, Yunzhi Yan, Qing Shuai, Hujun Bao, Xiaowei Zhou:
Efficient Neural Radiance Fields for Interactive Free-viewpoint Video. 39:1-39:9
Solids and Fluids
- Ty Trusty, Danny M. Kaufman, David I. W. Levin:
Mixed Variational Finite Elements for Implicit Simulation of Deformables. 40:1-40:8
Sampling and Reconstruction
- He Li, Beibei Wang, Changhe Tu, Kun Xu, Nicolas Holzschuch, Ling-Qi Yan:
Unbiased Caustics Rendering Guided by Representative Specular Paths. 41:1-41:8 - Rex West, Iliyan Georgiev, Toshiya Hachisuka:
Marginal Multiple Importance Sampling. 42:1-42:8
Everything Interactive and Dynamic
- Haoyu Hu, Xinyu Yi, Hao Zhang, Jun-Hai Yong, Feng Xu:
Physical Interaction: Reconstructing Hand-object Interactions with Physics. 43:1-43:9 - Elias Jadon, Bernhard Thomaszewski, Aleksandra Anna Apolinarska, Roi Poranne:
Continuous deformation based panelization for design rationalization. 44:1-44:8 - Yao Feng, Jinlong Yang, Marc Pollefeys, Michael J. Black, Timo Bolkart:
Capturing and Animation of Body and Clothing from Monocular Video. 45:1-45:9 - Di Huang, Xiaopeng Ji, Xingyi He, Jiaming Sun, Tong He, Qing Shuai, Wanli Ouyang, Xiaowei Zhou:
Reconstructing Hand-Held Objects from Monocular Video. 46:1-46:9
Material and Rendering
- Marina Alterman, Evgeniia Saiko, Anat Levin:
Direct acquisition of volumetric scattering phase function using speckle correlations. 47:1-47:9 - Ruizhi Shao, Liliang Chen, Zerong Zheng, Hongwen Zhang, Yuxiang Zhang, Han Huang, Yandong Guo, Yebin Liu:
FloRen: Real-time High-quality Human Performance Rendering via Appearance Flow Using Sparse RGB Cameras. 48:1-48:10 - Brandon Yushan Feng, Susmija Jabbireddy, Amitabh Varshney:
VIINTER: View Interpolation with Implicit Neural Representations of Images. 49:1-49:9
VR and Interaction
- Shangchen Han, Po-Chen Wu, Yubo Zhang, Beibei Liu, Linguang Zhang, Zheng Wang, Weiguang Si, Peizhao Zhang, Yujun Cai, Tomas Hodan, Randi Cabezas, Luan Tran, Muzaffer Akbay, Tsz-Ho Yu, Cem Keskin, Robert Wang:
UmeTrack: Unified multi-view end-to-end hand tracking for VR. 50:1-50:9 - Yong Li, Shoaib Kamil, Alec Jacobson, Yotam I. Gingold:
H rtDown: Document Processor for Executable Linear Algebra Papers. 51:1-51:8
Simulation of Everything
- Seung-wook Kim, JungHyun Han:
Fast Stabilization of Inducible Magnet Simulation. 52:1-52:8
CAD
- Joseph George Lambourne, Karl D. D. Willis, Pradeep Kumar Jayaraman, Longfei Zhang, Aditya Sanghi, Kamal Rahimi Malekshan:
Reconstructing editable prismatic CAD from rounded voxel models. 53:1-53:9
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.