Most Influential SIGGRAPH Papers (2023-04)
To search or review papers within SIGGRAPH related to a specific topic, please use the search by venue (SIGGRAPH) and review by venue (SIGGRAPH) services. To browse the most productive SIGGRAPH authors by year ranked by #papers accepted, here is a list of most productive SIGGRAPH authors.
Based in New York, Paper Digest is dedicated to producing high-quality text analysis results that people can acturally use on a daily basis. Since 2018, we have been serving users across the world with a number of exclusive services to track, search, review and rewrite scientific literature.
You are welcome to follow us on Twitter and Linkedin to get updated with new conference digests.
Paper Digest Team
New York City, New York, 10017
team@paperdigest.org
TABLE 1: Most Influential SIGGRAPH Papers (2023-04)
Year | Rank | Paper | Author(s) |
---|---|---|---|
2022 | 1 | Palette: Image-to-Image Diffusion Models IF:6 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper develops a unified framework for image-to-image translation based on conditional diffusion models and evaluates this framework on four challenging image-to-image translation tasks, namely colorization, inpainting, uncropping, and JPEG restoration. |
CHITWAN SAHARIA et. al. |
2022 | 2 | StyleGAN-NADA: CLIP-guided Domain Adaptation of Image Generators IF:5 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Leveraging the semantic power of large scale Contrastive-Language-Image-Pre-training (CLIP) models, we present a text-driven method that allows shifting a generative model to new domains, without having to collect even a single image. |
RINON GAL et. al. |
2022 | 3 | StyleGAN-XL: Scaling StyleGAN to Large Diverse Datasets IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Our final model, StyleGAN-XL, sets a new state-of-the-art on large-scale image synthesis and is the first to generate images at a resolution of 10242 at such a dataset scale. |
Axel Sauer; Katja Schwarz; Andreas Geiger; |
2022 | 4 | AvatarCLIP: Zero-shot Text-driven Generation and Animation of 3D Avatars IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, the whole production process is prohibitively time-consuming and labor-intensive. To democratize this technology to a larger audience, we propose AvatarCLIP, a zero-shot text-driven framework for 3D avatar generation and animation. |
FANGZHOU HONG et. al. |
2022 | 5 | CLIP2StyleGAN: Unsupervised Extraction of StyleGAN Edit Directions IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we investigate how to effectively link the pretrained latent spaces of StyleGAN and CLIP, which in turn allows us to automatically extract semantically-labeled edit directions from StyleGAN, finding and naming meaningful edit operations, in a fully unsupervised setup, without additional human guidance. |
Rameen Abdal; Peihao Zhu; John Femiani; Niloy Mitra; Peter Wonka; |
2022 | 6 | CLIPasso: Semantically-aware Object Sketching IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present CLIPasso, an object sketching method that can achieve different levels of abstraction, guided by geometric and semantic simplifications. |
YAEL VINKER et. al. |
2022 | 7 | ASE: Large-scale Reusable Adversarial Skill Embeddings for Physically Simulated Characters IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we present a large-scale data-driven framework for learning versatile and reusable skill embeddings for physically simulated characters. |
Xue Bin Peng; Yunrong Guo; Lina Halper; Sergey Levine; Sanja Fidler; |
2022 | 8 | Text2Human: Text-driven Controllable Human Image Generation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we present a text-driven controllable framework, Text2Human, for a high-quality and diverse human generation. |
YUMING JIANG et. al. |
2022 | 9 | Authentic Volumetric Avatars from A Phone Scan IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Creating photorealistic avatars of existing people currently requires extensive person-specific data capture, which is usually only accessible to the VFX industry and not the general public. Our work aims to address this drawback by relying only on a short mobile phone capture to obtain a drivable 3D head avatar that matches a person’s likeness faithfully. |
CHEN CAO et. al. |
2022 | 10 | EAMM: One-Shot Emotional Talking Face Via Audio-Based Emotion-Aware Motion Model IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose the Emotion-Aware Motion Model (EAMM) to generate one-shot emotional talking faces by involving an emotion source video. |
XINYA JI et. al. |
2022 | 11 | Variable Bitrate Neural Fields IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Unfortunately, these feature grids usually come at the cost of significantly increased memory consumption compared to stand-alone neural network models. We present a dictionary method for compressing such feature grids, reducing their memory consumption by up to 100 × and permitting a multiresolution representation which can be useful for out-of-core streaming. |
TOWAKI TAKIKAWA et. al. |
2022 | 12 | Domain Enhanced Arbitrary Image Style Transfer Via Contrastive Learning IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we tackle the challenging problem of arbitrary image style transfer using a novel style feature representation learning method. |
YUXIN ZHANG et. al. |
2022 | 13 | ReLU Fields: The Little Non-linearity That Could IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Hence, in this work, we investigate what is the smallest change to grid-based representations that allows for retaining the high fidelity result of MLPs while enabling fast reconstruction and rendering times. |
Animesh Karnewar; Tobias Ritschel; Oliver Wang; Niloy Mitra; |
2022 | 14 | Differentiable Signed Distance Function Rendering IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this article, we show how to extend the commonly used sphere tracing algorithm so that it additionally outputs a reparameterization that provides the means to compute accurate shape parameter derivatives. |
Delio Vicini; Sébastien Speierer; Wenzel Jakob; |
2022 | 15 | Learning Smooth Neural Functions Via Lipschitz Regularization IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we introduce a novel regularization designed to encourage smooth latent spaces in neural fields by penalizing the upper bound on the field’s Lipschitz constant. |
Hsueh-Ti Derek Liu; Francis Williams; Alec Jacobson; Sanja Fidler; Or Litany; |
2021 | 1 | Designing An Encoder for StyleGAN Image Manipulation IF:6 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we carefully study the latent space of StyleGAN, the state-of-the-art unconditional generator. |
Omer Tov; Yuval Alaluf; Yotam Nitzan; Or Patashnik; Daniel Cohen-Or; |
2021 | 2 | Learning An Animatable Detailed 3D Face Model from In-the-wild Images IF:5 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present the first approach that regresses 3D face shape and animatable details that are specific to an individual but change with expression. |
Yao Feng; Haiwen Feng; Michael J. Black; Timo Bolkart; |
2021 | 3 | AMP: Adversarial Motion Priors for Stylized Physics-based Character Control IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose to obviate the need to manually design imitation objectives and mechanisms for motion selection by utilizing a fully automated approach based on adversarial imitation learning. |
Xue Bin Peng; Ze Ma; Pieter Abbeel; Sergey Levine; Angjoo Kanazawa; |
2021 | 4 | Acorn: Adaptive Coordinate Networks for Neural Scene Representation IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Here, we introduce a new hybrid implicit-explicit network architecture and training strategy that adaptively allocates resources during training and inference based on the local complexity of a signal of interest. |
JULIEN N. P. MARTEL et. al. |
2021 | 5 | Mixture of Volumetric Primitives for Efficient Neural Rendering IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present Mixture of Volumetric Primitives (MVP), a representation for rendering dynamic 3D content that combines the completeness of volumetric representations with the efficiency of primitive-based rendering, e.g., point-based or mesh-based methods. |
STEPHEN LOMBARDI et. al. |
2021 | 6 | Editable Free-viewpoint Video Using A Layered Neural Representation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To fill this gap, in this paper, we propose the first approach for editable free-viewpoint video generation for large-scale view-dependent dynamic scenes using only 16 cameras. |
JIAKAI ZHANG et. al. |
2021 | 7 | Only A Matter of Style: Age Transformation Using A Style-based Regression Model IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we present an image-to-image translation method that learns to directly encode real facial images into the latent space of a pre-trained unconditional GAN (e.g., StyleGAN) subject to a given aging shift. |
Yuval Alaluf; Or Patashnik; Daniel Cohen-Or; |
2021 | 8 | Total Relighting: Learning to Relight Portraits for Background Replacement IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a novel system for portrait relighting and background replacement, which maintains high-frequency boundary details and accurately synthesizes the subject’s appearance as lit by novel illumination, thereby producing realistic composite images for any desired scene. |
ROHIT PANDEY et. al. |
2021 | 9 | SP-GAN: Sphere-guided 3D Shape Generation and Manipulation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present SP-GAN, a new unsupervised sphere-guided generative model for direct synthesis of 3D shapes in the form of point clouds. |
Ruihui Li; Xianzhi Li; Ka-Hei Hui; Chi-Wing Fu; |
2021 | 10 | Real-time Deep Dynamic Characters IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a deep videorealistic 3D human character model displaying highly realistic shape, motion, and dynamic appearance learned in a new weakly supervised way from multi-view imagery. |
MARC HABERMANN et. al. |
2021 | 11 | TryOnGAN: Body-aware Try-on Via Layered Interpolation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Given a pair of images—target person and garment on another person—we automatically generate the target person in the given garment. |
Kathleen M Lewis; Srivatsan Varadharajan; Ira Kemelmacher-Shlizerman; |
2021 | 12 | FovVideoVDP: A Visible Difference Predictor for Wide Field-of-view Video IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: FovVideoVDP is a video difference metric that models the spatial, temporal, and peripheral aspects of perception. While many other metrics are available, our work provides the first practical treatment of these three central aspects of vision simultaneously. |
RAFA? K. MANTIUK et. al. |
2021 | 13 | TransPose: Real-time 3D Human Translation and Pose Estimation with Six Inertial Sensors IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present TransPose, a DNN-based approach to perform full motion capture (with both global translations and body poses) from only 6 Inertial Measurement Units (IMUs) at over 90 fps. |
Xinyu Yi; Yuxiao Zhou; Feng Xu; |
2021 | 14 | AgileGAN: Stylizing Portraits By Inversion-consistent Transfer Learning IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Hence we propose AgileGAN, a framework that can generate high quality stylistic portraits via inversion-consistent transfer learning. |
GUOXIAN SONG et. al. |
2021 | 15 | SWAGAN: A Style-based Wavelet-driven Generative Model IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To address this issue, we present a novel general-purpose Style and WAvelet based GAN (SWAGAN) that implements progressive generation in the frequency domain. |
Rinon Gal; Dana Cohen Hochberg; Amit Bermano; Daniel Cohen-Or; |
2020 | 1 | Consistent Video Depth Estimation IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present an algorithm for reconstructing dense, geometrically consistent depth for all pixels in a monocular video. |
Xuan Luo; Jia-Bin Huang; Richard Szeliski; Kevin Matzen; Johannes Kopf; |
2020 | 2 | Immersive Light Field Video With A Layered Mesh Representation IF:5 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a system for capturing, reconstructing, compressing, and rendering high quality immersive light field video. |
MICHAEL BROXTON et. al. |
2020 | 3 | XNect: Real-time Multi-person 3D Motion Capture With A Single RGB Camera IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a real-time approach for multi-person 3D motion capture at over 30 fps using a single RGB camera. |
DUSHYANT MEHTA et. al. |
2020 | 4 | Character Controllers Using Motion VAEs IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We learn data-driven generative models of human movement using autoregressive conditional variational autoencoders, or Motion VAEs. |
Hung Yu Ling; Fabio Zinno; George Cheng; Michiel Van De Panne; |
2020 | 5 | Learning Temporal Coherence Via Self-supervision For GAN-based Video Generation IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In contrast, we focus on improving learning objectives and propose a temporally self-supervised algorithm. |
Mengyu Chu; You Xie; Jonas Mayer; Laura Leal-Taixé; Nils Thuerey; |
2020 | 6 | Robust Motion In-betweening IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work we present a novel, robust transition generation technique that can serve as a new tool for 3D animators, based on adversarial recurrent neural networks. |
Félix G. Harvey; Mike Yurick; Derek Nowrouzezahrai; Christopher Pal; |
2020 | 7 | Point2Mesh: A Self-prior For Deformable Meshes IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we introduce Point2Mesh, a technique for reconstructing a surface mesh from an input point cloud. |
Rana Hanocka; Gal Metzer; Raja Giryes; Daniel Cohen-Or; |
2020 | 8 | DeepFaceDrawing: Deep Generation Of Face Images From Sketches IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To address this issue, our key idea is to implicitly model the shape space of plausible face images and synthesize a face image in this space to approximate an input sketch. |
Shu-Yu Chen; Wanchao Su; Lin Gao; Shihong Xia; Hongbo Fu; |
2020 | 9 | Local Motion Phases For Learning Multi-contact Character Movements IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel framework to learn fast and dynamic character interactions that involve multiple contacts between the body and an object, another character and the environment, from a rich, unstructured motion capture database. |
Sebastian Starke; Yiwei Zhao; Taku Komura; Kazi Zaman; |
2020 | 10 | A Scalable Approach To Control Diverse Behaviors For Physically Simulated Characters IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we develop a technique for learning controllers for a large set of heterogeneous behaviors. |
Jungdam Won; Deepak Gopinath; Jessica Hodgins; |
2020 | 11 | MEgATrack: Monochrome Egocentric Articulated Hand-tracking For Virtual Reality IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a system for real-time hand-tracking to drive virtual and augmented reality (VR/AR) experiences. |
SHANGCHEN HAN et. al. |
2020 | 12 | Single Image HDR Reconstruction Using A CNN With Masked Features And Perceptual Loss IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present a novel learning-based approach to reconstruct an HDR image by recovering the saturated pixels of an input LDR image in a visually pleasing way. |
Marcel Santana Santos; Tsang Ing Ren; Nima Khademi Kalantari; |
2020 | 13 | Fast Tetrahedral Meshing In The Wild IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a new tetrahedral meshing method, fTetWild, to convert triangle soups into high-quality tetrahedral meshes. |
Yixin Hu; Teseo Schneider; Bolun Wang; Denis Zorin; Daniele Panozzo; |
2020 | 14 | Skeleton-aware Networks For Deep Motion Retargeting IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce a novel deep learning framework for data-driven motion retargeting between skeletons, which may have different structure, yet corresponding to homeomorphic graphs. |
KFIR ABERMAN et. al. |
2020 | 15 | Path-space Differentiable Rendering IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we show how path integrals can be differentiated with respect to arbitrary differentiable changes of a scene. |
Cheng Zhang; Bailey Miller; Kai Yan; Ioannis Gkioulekas; Shuang Zhao; |
2019 | 1 | Deferred Neural Rendering: Image Synthesis Using Neural Textures IF:7 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we explore the use of imperfect 3D content, for instance, obtained from photo-metric reconstructions with noisy and incomplete surface geometry, while still aiming to produce photo-realistic (re-)renderings. |
Justus Thies; Michael Zollh�fer; Matthias Nie�ner; |
2019 | 2 | Local Light Field Fusion: Practical View Synthesis With Prescriptive Sampling Guidelines IF:6 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a practical and robust deep learning solution for capturing and rendering novel views of complex real world scenes for virtual exploration. |
BEN MILDENHALL et. al. |
2019 | 3 | MeshCNN: A Network With An Edge IF:6 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we utilize the unique properties of the mesh for a direct analysis of 3D shapes using MeshCNN, a convolutional neural network designed specifically for triangular meshes. |
RANA HANOCKA et. al. |
2019 | 4 | Semantic Photo Manipulation With A Generative Image Prior IF:5 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we address these issues by adapting the image prior learned by GANs to image statistics of an individual image. |
DAVID BAU et. al. |
2019 | 5 | Text-based Editing Of Talking-head Video IF:5 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a novel method to edit talking-head video based on its transcript to produce a realistic output video in which the dialogue of the speaker has been modified, while maintaining a seamless audio-visual flow (i.e. no jump cuts). |
OHAD FRIED et. al. |
2019 | 6 | Single Image Portrait Relighting IF:5 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To this end, we present a system for portrait relighting: a neural network that takes as input a single RGB image of a portrait taken with a standard cellphone camera in an unconstrained environment, and from that image produces a relit image of that subject as though it were illuminated according to any provided environment map. |
TIANCHENG SUN et. al. |
2019 | 7 | Learning To Optimize Halide With Tree Search And Random Programs IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a new algorithm to automatically schedule Halide programs for high-performance image processing and deep learning. |
ANDREW ADAMS et. al. |
2019 | 8 | Scalable Muscle-actuated Human Simulation And Control IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work aims to build a comprehensive musculoskeletal model and its control system that reproduces realistic human movements driven by muscle contraction dynamics. |
Seunghwan Lee; Moonseok Park; Kyoungmin Lee; Jehee Lee; |
2019 | 9 | Handheld Multi-frame Super-resolution IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we supplant the use of traditional demosaicing in single-frame and burst photography pipelines with a multiframe super-resolution algorithm that creates a complete RGB image directly from a burst of CFA raw images. |
BARTLOMIEJ WRONSKI et. al. |
2019 | 10 | PlanIT: Planning And Instantiating Indoor Scenes With Relation Graph And Spatial Prior Networks IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a new framework for interior scene synthesis that combines a high-level relation graph representation with spatial prior neural networks. |
KAI WANG et. al. |
2019 | 11 | Real-time Pose And Shape Reconstruction Of Two Interacting Hands With A Single Depth Camera IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a novel method for real-time pose and shape reconstruction of two strongly interacting hands. |
FRANZISKA MUELLER et. al. |
2019 | 12 | Deep Inverse Rendering For High-resolution SVBRDF Estimation From An Arbitrary Number Of Images IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper we present a unified deep inverse rendering framework for estimating the spatially-varying appearance properties of a planar exemplar from an arbitrary number of input photographs, ranging from just a single photograph to many photographs. |
DUAN GAO et. al. |
2019 | 13 | Interactive Hand Pose Estimation Using A Stretch-sensing Soft Glove IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a stretch-sensing soft glove to interactively capture hand poses with high accuracy and without requiring an external optical setup. |
Oliver Glauser; Shihao Wu; Daniele Panozzo; Otmar Hilliges; Olga Sorkine-Hornung; |
2019 | 14 | Deep View Synthesis From Sparse Photometric Images IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we synthesize novel viewpoints across a wide range of viewing directions (covering a 60? cone) from a sparse set of just six viewing directions. |
ZEXIANG XU et. al. |
2019 | 15 | VR Facial Animation Via Multiview Image Translation IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we present a bidirectional system that can animate avatar heads of both users’ full likeness using consumer-friendly headset mounted cameras (HMC). |
SHIH-EN WEI et. al. |