SL$^{2}$A-INR: Single-Layer Learnable Activation for Implicit Neural Representation

📅 2024-09-17
📈 Citations: 2
Influential: 0
📄 PDF
🤖 AI Summary
Implicit Neural Representations (INRs) are constrained by fixed activation functions, limiting their efficiency in modeling high-frequency and heterogeneous signals. To address this, we propose a lightweight fusion architecture that replaces the conventional ReLU with a single-layer learnable piecewise-linear activation module, integrated synergistically with coordinate encoding. The model is optimized via multi-task joint training on image reconstruction, 3D shape reconstruction, and novel-view synthesis. Crucially, this work introduces the first parameterized activation function embedded directly into the INR backbone—enhancing high-frequency detail representation without increasing network depth. Our approach achieves state-of-the-art performance across multiple INR benchmarks, demonstrating superior visual quality and generalization. The implementation is publicly available.

Technology Category

Application Category

📝 Abstract
Implicit Neural Representation (INR), leveraging a neural network to transform coordinate input into corresponding attributes, has recently driven significant advances in several vision-related domains. However, the performance of INR is heavily influenced by the choice of the nonlinear activation function used in its multilayer perceptron (MLP) architecture. To date, multiple nonlinearities have been investigated, but current INRs still face limitations in capturing high-frequency components and diverse signal types. We show that these challenges can be alleviated by introducing a novel approach in INR architecture. Specifically, we propose SL$^{2}$A-INR, a hybrid network that combines a single-layer learnable activation function with an MLP that uses traditional ReLU activations. Our method performs superior across diverse tasks, including image representation, 3D shape reconstruction, and novel view synthesis. Through comprehensive experiments, SL$^{2}$A-INR sets new benchmarks in accuracy, quality, and robustness for INR. Our Code is publicly available on~href{https://github.com/Iceage7/SL2A-INR}{ extcolor{magenta}{GitHub}}.
Problem

Research questions and friction points this paper is trying to address.

Improving INR performance with learnable activation functions
Enhancing high-frequency component capture in neural representations
Addressing limitations in diverse signal type representation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Single-layer learnable activation function
Hybrid network combining learnable and ReLU activations
Superior performance in diverse vision tasks
🔎 Similar Papers
No similar papers found.
Moein Heidari
Moein Heidari
PhD Student, University of British Columbia
Computer VisionDeep LearningMachine LearningImage Processing
R
Reza Rezaeian
School of Electrical and Computer Engineering, College of Engineering, The University of Tehran, Iran
R
Reza Azad
Faculty of Electrical Engineering and Information Technology, RWTH Aachen University Aachen, Germany
D
D. Merhof
Faculty of Informatics and Data Science, University of Regensburg Regensburg, Germany
Hamid Soltanian-Zadeh
Hamid Soltanian-Zadeh
Senior Scientist, Henry Ford Hospital, Professor, University of Tehran,
Image ProcessingMedical ImagingPattern RecognitionNeural Networks
I
I. Hacihaliloglu
Department of Radiology, Department of Medicine, University of British Columbia, Vancouver, BC, Canada