RIGA-Fold: A General Framework for Protein Inverse Folding via Recurrent Interaction and Geometric Awareness

📅 2026-02-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the limitations of existing protein inverse folding methods, which often suffer from restricted receptive fields that neglect long-range dependencies and error accumulation during single-pass inference. To overcome these challenges, the authors propose RIGA-Fold, a novel framework featuring a geometric attention update module for SE(3)-invariant local encoding and a global context bridging mechanism that dynamically injects topological information. The architecture integrates a dual-stream design to fuse trainable geometric features with frozen evolutionary priors from ESM-2/ESM-IF, and employs an iterative denoising strategy based on prediction, recycling, and refinement. Evaluated on CATH 4.2, TS50, and TS500 benchmarks, RIGA-Fold significantly outperforms current state-of-the-art models, achieving notable advances in both sequence recovery accuracy and structural consistency.

Technology Category

Application Category

📝 Abstract
Protein inverse folding, the task of predicting amino acid sequences for desired structures, is pivotal for de novo protein design. However, existing GNN-based methods typically suffer from restricted receptive fields that miss long-range dependencies and a"single-pass"inference paradigm that leads to error accumulation. To address these bottlenecks, we propose RIGA-Fold, a framework that synergizes Recurrent Interaction with Geometric Awareness. At the micro-level, we introduce a Geometric Attention Update (GAU) module where edge features explicitly serve as attention keys, ensuring strictly SE(3)-invariant local encoding. At the macro-level, we design an attention-based Global Context Bridge that acts as a soft gating mechanism to dynamically inject global topological information. Furthermore, to bridge the gap between structural and sequence modalities, we introduce an enhanced variant, RIGA-Fold*, which integrates trainable geometric features with frozen evolutionary priors from ESM-2 and ESM-IF via a dual-stream architecture. Finally, a biologically inspired ``predict-recycle-refine''strategy is implemented to iteratively denoise sequence distributions. Extensive experiments on CATH 4.2, TS50, and TS500 benchmarks demonstrate that our geometric framework is highly competitive, while RIGA-Fold* significantly outperforms state-of-the-art baselines in both sequence recovery and structural consistency.
Problem

Research questions and friction points this paper is trying to address.

protein inverse folding
long-range dependencies
error accumulation
de novo protein design
sequence-structure mapping
Innovation

Methods, ideas, or system contributions that make the work stand out.

Protein Inverse Folding
Geometric Attention
Recurrent Interaction
SE(3) Invariance
Dual-stream Architecture
🔎 Similar Papers
No similar papers found.
S
Sisi Yuan
National Engineering Laboratory for Big Data System Computing Technology, Shenzhen University, Shenzhen, China
J
Jiehuang Chen
College of Computer Science and Software Engineering, Shenzhen University, Shenzhen, China
J
Junchuang Cai
National Engineering Laboratory for Big Data System Computing Technology, Shenzhen University, Shenzhen, China
Dong Xu
Dong Xu
Shenzhen University
Artificial intelligenceDrug Design
Xueliang Li
Xueliang Li
Shenzhen University
Artificial IntelligenceArtificial EmpathySustainable ComputingNearly Everything
Zexuan Zhu
Zexuan Zhu
Shenzhen University
Evolutionary ComputationMemetic ComputingBioinformaticsMachine Learning
J
Junkai Ji
National Engineering Laboratory for Big Data System Computing Technology, Shenzhen University, Shenzhen, China