Export Citations
Save this search
Please login to be able to save your searches and receive alerts for new content matching your search criteria.
- research-articleDecember 2021
Early convolutions help transformers see better
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 2325, Pages 30392–30400Vision transformer (ViT) models exhibit substandard optimizability. In particular, they are sensitive to the choice of optimizer (AdamW vs. SGD), optimizer hyperpa-rameters, and training schedule length. In comparison, modern convolutional neural ...
- research-articleDecember 2021
Post-training quantization for vision transformer
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 2152, Pages 28092–28103Recently, transformer has achieved remarkable performance on a variety of computer vision applications. Compared with mainstream convolutional neural networks, vision transformers are often of sophisticated architectures for extracting powerful feature ...
- research-articleJune 2024
Moiré Attack (MA): a new potential risk of screen photos
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 2000, Pages 26117–26129Images, captured by a camera, play a critical role in training Deep Neural Networks (DNNs). Usually, we assume the images acquired by cameras are consistent with the ones perceived by human eyes. However, due to the different physical mechanisms between ...
- research-articleDecember 2021
Look at the variance! efficient black-box explanations with sobol-based sensitivity analysis
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 1991, Pages 26005–26014We describe a novel attribution method which is grounded in Sensitivity Analysis and uses Sobol indices. Beyond modeling the individual contributions of image regions, Sobol indices provide an efficient way to capture higher-order interactions between ...
- research-articleDecember 2021
IA-RED2: interpretability-aware redundancy reduction for vision transformers
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 1907, Pages 24898–24911The self-attention-based model, transformer, is recently becoming the leading backbone in the feld of computer vision. In spite of the impressive success made by transformers in a variety of vision tasks, it still suffers from heavy computation and ...
-
- research-articleJune 2024
Trash or treasure? an interactive dual-stream strategy for single image reflection separation
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 1890, Pages 24683–24694Single image reflection separation (SIRS), as a representative blind source separation task, aims to recover two layers, i.e., transmission and reflection, from one mixed observation, which is challenging due to the highly ill-posed nature. Existing deep ...
- research-articleDecember 2021
Memory efficient meta-learning with large images
- John Bronskill,
- Daniela Massiceti,
- Massimiliano Patacchiola,
- Katja Hofmann,
- Sebastian Nowozin,
- Richard E. Turner
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 1862, Pages 24327–24339Meta learning approaches to few-shot classification are computationally efficient at test time, requiring just a few optimization steps or single forward pass to learn a new task, but they remain highly memory-intensive to train. This limitation arises ...
- research-articleDecember 2021
Adaptive denoising via GainTuning
- Sreyas Mohan,
- Joshua L. Vincent,
- Ramon Manzorro,
- Peter A. Crozier,
- Carlos Fernandez-Granda,
- Eero P. Simoncelli
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 1817, Pages 23727–23740Deep convolutional neural networks (CNNs) for image denoising are typically trained on large datasets. These models achieve the current state of the art, but they do not generalize well to data that deviate from the training distribution. Recent work has ...
- research-articleJune 2024
Spectrum-to-kernel translation for accurate blind image super-resolution
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 1734, Pages 22643–22654Deep-learning based Super-Resolution (SR) methods have exhibited promising performance under non-blind setting where blur kernel is known. However, blur kernels of Low-Resolution (LR) images in different practical applications are usually unknown. It may ...
- research-articleDecember 2021
On the frequency bias of generative models
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 1387, Pages 18126–18136The key objective of Generative Adversarial Networks (GANs) is to generate new data with the same statistics as the provided training data. However, multiple recent works show that state-of-the-art architectures yet struggle to achieve this goal. In ...
- research-articleJune 2024
ResT: an efficient transformer for visual recognition
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 1185, Pages 15475–15485This paper presents an efficient multi-scale vision Transformer, called ResT, that capably served as a general-purpose backbone for image recognition. Unlike existing Transformer methods, which employ standard Transformer blocks to tackle raw images with ...
- research-articleDecember 2021
Augmented shortcuts for vision transformers
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 1173, Pages 15316–15327Transformer models have achieved great progress on computer vision tasks recently. The rapid development of vision transformers is mainly contributed by their high representation ability for extracting informative features from input images. However, the ...
- research-articleDecember 2021
DynamicViT: efficient vision transformers with dynamic token sparsification
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 1068, Pages 13937–13949Attention is sparse in vision transformers. We observe the final prediction in vision transformers is only based on a subset of most informative tokens, which is sufficient for accurate image recognition. Based on this observation, we propose a dynamic ...
- research-articleDecember 2021
Implicit transformer network for screen content image continuous super-resolution
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 1019, Pages 13304–13315Nowadays, there is an explosive growth of screen contents due to the wide application of screen sharing, remote cooperation, and online education. To match the limited terminal bandwidth, high-resolution (HR) screen contents may be downsampled and ...
- research-articleJune 2024
Glance-and-gaze vision transformer
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 995, Pages 12992–13003Recently, there emerges a series of vision Transformers, which show superior performance with a more compact model size than conventional convolutional neural networks, thanks to the strong ability of Transformers to model long-range dependencies. ...
- research-articleJune 2024
Diffusion models beat GANs on image synthesis
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 672, Pages 8780–8794We show that diffusion models can achieve image sample quality superior to the current state-of-the-art generative models. We achieve this on unconditional image synthesis by finding a better architecture through a series of ablations. For conditional ...
- research-articleJune 2024
Functional neural networks for parametric image restoration problems
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 518, Pages 6762–6775Almost every single image restoration problem has a closely related parameter, such as the scale factor in super-resolution, the noise level in image denoising, and the quality factor in JPEG deblocking. Although recent studies on image restoration ...
- research-articleJune 2024
A trainable spectral-spatial sparse coding model for hyperspectral image restoration
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 415, Pages 5430–5442Hyperspectral imaging offers new perspectives for diverse applications, ranging from the monitoring of the environment using airborne or satellite remote sensing, precision farming, food safety, planetary exploration, or astrophysics. Unfortunately, the ...
- research-articleDecember 2021
Blending anti-aliasing into vision transformer
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 414, Pages 5416–5429The transformer architectures, based on self-attention mechanism and convolution-free design, recently found superior performance and booming applications in computer vision. However, the discontinuous patch-wise tokenization process implicitly ...
- research-articleDecember 2021
Aligned structured sparsity learning for efficient image super-resolution
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 206, Pages 2695–2706Lightweight image super-resolution (SR) networks have obtained promising results with moderate model size. Many SR methods have focused on designing lightweight architectures, which neglect to further reduce the redundancy of network parameters. On the ...