Data Efficient Neural Scaling Law via Model Reusing
Peihao Wang,
Rameswar Panda,
Atlas Wang
International Conference on Machine Learning (ICML), 2023
We empirically reveal that the existing power laws under-estimate the data inefficiency of large transformers, and leverage model reusing to reproduce the power law under the data scarcity regime.
[Paper]
[Code]
NeuralLift-360: Lifting An In-the-wild 2D Photo to A 3D Object with 360 Views
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2023, Highlight
We study how to lift a single image to a 3D object and generate its 360° views that correspond well with the given reference image.
[Project Page]
[Paper]
[Code]
Equivariant Hypergraph Diffusion Neural Operators
International Conference on Learning Representations (ICLR), 2023
This work proposes a new hypergraph neural network architecture, which provably represents any continuous equivariant hypergraph diffusion operators.
[Paper]
[Code]
Is Attention All That NeRF Needs?
International Conference on Learning Representations (ICLR), 2023
We present Generalizable NeRF Transformer (GNT), a pure, unified transformer-based architecture that efficiently reconstructs Neural Radiance Fields (NeRFs) on the fly.
[Project Page]
[Paper]
[Code]
NeRF-SOS: Any-View Self-supervised Object Segmentation on Complex Scenes
International Conference on Learning Representations (ICLR), 2023
We propose a novel collaborative contrastive loss for NeRF to segment objects in complex real-world scenes, without any annotation.
[Project Page]
[Paper]
[Code]
Learning to Grow Pretrained Models for Efficient Transformer Training
International Conference on Learning Representations (ICLR), 2023, Spotlight
This paper proposes to accelerate transformer training by re-using pretrained models via a learnable, linear and sparse model growth operator.
[MIT News]
[Paper]
[Code]
Signal Processing for Implicit Neural Representations
Advances in Neural Information Processing Systems (NeurIPS), 2022
We propose a theoretically grounded signal processing framework for Implicit Neural Representations (INR), which analytically manipulates INRs on the weight space through differential operators.
[Project Page]
[Paper]
[Code]
Old can be Gold: Better Gradient Flow can Make Vanilla-GCNs Great Again
Advances in Neural Information Processing Systems (NeurIPS), 2022
We derive a topology-aware isometric initialization and a Dirichlet Energy guided achitectural rewiring technique that boost vanilla-GCNs to be the state-of-the-art.
[Paper]
[Code]
Unified Implicit Neural Stylization
European Conference on Computer Vision (ECCV), 2022
This work explores stylizing an implicit neural representation, using a generalized approach that can apply to various 2D and 3D representations.
[Project Page]
[Paper]
[Code]
SinNeRF: Training Neural Radiance Fields on Complex Scenes from a Single Image
European Conference on Computer Vision (ECCV), 2022
We present Single View NeRF (SinNeRF) consisting of thoughtfully designed semantic and geometry regularizations to train neural radiance field using only a single view.
[Project Page]
[Paper]
[Code]
Neural Implicit Dictionary Learning via Mixture-of-Expert Training
International Conference on Machine Learning (ICML), 2022
We present Neural Implicit Dictionary (NID) that learns and represents implicit neural representation as a sparse mixture of expert networks.
[Paper]
[Code]
Bag of Tricks for Training Deeper Graph Neural Networks: A Comprehensive Benchmark Study
IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2022
We present the first fair and reproducible benchmark dedicated to assessing the "tricks" of training deep GNNs.
[Paper]
[Code]
Aug-NeRF: Training Stronger Neural Radiance Fields with Triple-Level Augmentations
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2022
We propose Aug-NeRF which augments NeRF with worst-case perturbations in three distinct levels with physical grounds.
[Paper]
[Code]
CADTransformer: Panoptic Symbol Spotting Transformer for CAD Drawings
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2022, Oral
We present CADTransformer, a transformer based framework, to tackle the panoptic symbol spotting task for computer-aided design drawings.
[Paper]
[Code]
Anti-Oversmoothing in Deep Vision Transformers via the Fourier Domain Analysis: From Theory to Practice
International Conference on Learning Representations (ICLR), 2022
We prove that self-attention is no more than low-pass filter, and propose two simple yet effective methods to counteract excessive smoothening.
[Paper]
[Code]
Delayed Propagation Transformer: A Universal Computation Engine towards Practical Control in Cyber-Physical Systems
Advances in Neural Information Processing Systems (NeurIPS), 2021
We presents the Delayed Propagation Transformer (DePT) that specializes in the global modeling of CPS while taking into account the immutable constraints from the physical world.
[Paper]
[Code]
SoGCN: Second-Order Graph Convolutional Networks
arXiv preprint, 2021
We prove that second-order graph convolution is the maximally localized kernel with full representation power.
[Paper]
[Code]
TightCap: 3D Human Shape Capture with Clothing Tightness Field
ACM Transactions on Graphics (TOG), 2021
We propose a data-driven approach to capture both the human shape and dressed garments with only a single 3D human scan, by predicting clothing tightness.
[Project Page]
[Paper]
[Code]