Papers
Topics
Authors
Recent
Search
2000 character limit reached

NeRFlex: Real-Time Mobile Neural Rendering

Updated 9 February 2026
  • NeRFlex is a resource-aware real-time rendering framework that decomposes complex 3D scenes and optimally balances memory usage with visual quality.
  • It integrates multi-NeRF scene decomposition with a domain-informed profiler to predict storage and SSIM tradeoffs for enhanced visual fidelity.
  • Leveraging a pseudo-polynomial dynamic programming solution to the NP-hard MCK problem, NeRFlex ensures interactive performance and strict memory compliance on mobile platforms.

NeRFlex is a resource-aware real-time rendering framework that targets interactive, high-fidelity synthesis of complex 3D scenes on mobile devices, fundamentally re-architecting Neural Radiance Fields (@@@@4@@@@) methods to address memory and computational constraints. NeRFlex integrates a multi-NeRF scene decomposition, a domain-informed profiler for memory/quality tradeoffs, and a dynamic programming optimization over the NP-hard Multiple-Choice Knapsack (MCK) configuration problem. Its design enables real-time rendering at quality levels previously unattainable on commercial mobile platforms, robustly adhering to strict device storage budgets while leveraging a principled treatment of visual frequency and resource consumption (Wang et al., 4 Apr 2025).

1. Multi-NeRF Scene Decomposition

NeRFlex decomposes complex scenes into multiple sub-scenes, assigning each a dedicated NeRF network based on localized visual detail. The segmentation module operates by:

  • Performing object detection across all input images, yielding per-object binary masks.
  • Computing a 2D detail frequency ff for each object in every training view (using, e.g., Laplacian responses), and recording the maximum frequency fmaxf_{\max} per object.
  • Objects with fmaxαf_{\max} \geq \alpha (for threshold α\alpha) are classified as high-detail and assigned individual NeRFs; all others are grouped together with a shared NeRF.

To facilitate efficient representation learning for high-detail objects, NeRFlex applies interpolation scaling: each selected object is cropped from training images and rescaled so that it fills the full input resolution, alleviating the NeRF’s need to capture high-frequency details at small scales.

Table: Multi-NeRF Scene Decomposition Workflow

Step Operation Outcome
Object Detection Binary masks generation per candidate object Isolated object masks in all images
Detail Frequency Compute ff per view, record fmaxf_{\max} Quantitative detail measure for segmentation
Thresholding Classify by fmaxαf_{\max} \geq \alpha High/low-detail object distinction
Interpolation Scaling Crop and rescale selected objects per image High-detail object NeRFs optimized

2. Lightweight Profiler and Modeling Memory-Quality Tradeoffs

NeRFlex incorporates a lightweight profiler for each NeRF representation to estimate the tradeoff between memory usage and visual quality. It exposes two primary configuration parameters per NeRF:

  • Geometry grid resolution gg (voxel grid size, g3g^3 total voxels)
  • Texture patch size pp (patch per mesh face, p2p^2 texels per face)

The profiler fits white-box polynomial models for predicted data storage S(g,p)S(g, p) and quality Q(g,p)Q(g, p) (measured via SSIM):

S(g,p)=fs(g,p)=k(g+a)3(p+b)2+m Q(g,p)=fq(g,p)=k(g+a)3(p+b)2\begin{aligned} S(g,p) = f_s(g,p) &= \frac{-k}{(g + a)^3 (p + b)^2} + m\ Q(g,p) = f_q(g,p) &= k' (g + a')^3 (p + b')^2 \end{aligned}

where k,a,b,m,k,a,bk, a, b, m, k', a', b' are constants fitted empirically from a small grid of (g,p)(g, p) samples. The profiler achieves mean prediction errors of approximately 0.0065 (SSIM, σ=0.0088\sigma = 0.0088) and $3.34$ MB in storage (σ=2.73\sigma = 2.73 MB), ensuring reliable input to subsequent optimization.

3. Formal Resource-Aware Configuration as MCK

For each segmented object ii (i=1,,ni=1, \dots, n), NeRFlex selects a configuration θi=(gi,pi)\theta_i = (g_i, p_i) from a discrete candidate set CiC_i. Formally, the allocation of NeRF parameters across all sub-scenes is cast as a Multiple-Choice Knapsack (MCK) problem:

  • Objective: Maximize aggregate visual quality over all objects:

maxxiji=1nθjCifqi(θj)xij\max_{x_{ij} } \sum_{i=1}^n \sum_{\theta_j \in C_i} f_{q_i}(\theta_j) x_{ij}

where xij{0,1}x_{ij} \in \{0,1\} indicates object ii uses configuration jj.

  • Memory Constraint: Aggregate storage must not exceed device budget HH:

i=1nθjCifsi(θj)xijH\sum_{i=1}^n \sum_{\theta_j \in C_i} f_{s_i}(\theta_j) x_{ij} \leq H

  • Uniqueness Constraint: Each object receives exactly one configuration:

iθjCixij=1\forall i \quad \sum_{\theta_j \in C_i} x_{ij} = 1

Solving this NP-hard problem robustly is central to NeRFlex's guarantees on memory-efficiency and visual fidelity.

4. Dynamic Programming Solution to Configuration Selection

To practically solve the NP-hard MCK, NeRFlex employs a pseudo-polynomial time dynamic programming (DP) algorithm. For nn objects and device memory HH:

  • DP state: DP[i,c]DP[i, c] = maximum sum-SSIM achievable with the first ii objects not exceeding total size cc.
  • Recurrence: For each i=1ni=1 \dots n, c=0Hc=0 \dots H:

DP[i,c]=maxθCi,fsi(θ)c{DP[i1,cfsi(θ)]+fqi(θ)}DP[i, c] = \max_{\theta \in C_i, f_{s_i}(\theta) \leq c} \left\{ DP[i-1, c - f_{s_i}(\theta)] + f_{q_i}(\theta) \right\}

  • Initialization: DP[0,c]=0DP[0, c] = 0 for all cc.

Additional pruning is performed by precomputing ri=HhiminθChfsh(θ)r_i = H - \sum_{h \neq i} \min_{\theta \in C_h} f_{s_h}(\theta), excluding any θ\theta that would preclude a full solution. The backtracking phase retrieves the optimal configuration per object.

The algorithm exhibits complexity O(nHCmax)\mathcal{O}(n H |C_{\max}|), tractable for realistic settings (tens to hundreds of MB memory budgets).

5. Experimental Evaluation on Mobile Platforms

NeRFlex demonstrates strong empirical performance on commercial hardware, evaluated on both real and synthetic datasets.

  • Device Budgets: iPhone 13 (H=240H=240 MB), Google Pixel 4 (H=150H=150 MB)
  • Quality Metrics: PSNR, SSIM, LPIPS (higher is better for PSNR, SSIM; lower is better for LPIPS)
Method PSNR↑ SSIM↑ LPIPS↓
MipNeRF 360 26.55 0.815 0.183
NGP (Instant NGP) 27.21 0.851 0.136
MobileNeRF 26.03 0.785 0.207
NeRFlex 27.65 0.886 0.114
  • Data-Size vs Quality: Block-NeRF (≥400 MB, infeasible); Single NeRF (~250 MB, SSIM ≈0.84–0.88, may fail); NeRFlex consistently meets its budgets (150 or 240 MB), attaining SSIM ≈0.90+, matching Block-NeRF quality.
  • Frame Rate: Scene 3, 360° pan at 7.5 s/turn: iPhone 13 ≈ 35 FPS; Pixel 4 ≈ 25 FPS (2× Single NeRF); Block-NeRF unable to load; Single NeRF sometimes stalls at 0 FPS when exceeding budget.
  • Cloud Processing Overhead (per 20 images): ~3.8 s for segmentation/interpolation, ~0.28 s for profiler, ~1.9 s for DP selection; total ~5.9 s (one-time).

6. Synthesis and Formal Innovation

NeRFlex unifies domain-driven multi-NeRF decomposition for high-frequency detail management, high-accuracy polynomial profiling of resource/quality tradeoffs, an explicit combinatorial optimization framework grounded in MCK, and a practical DP solver tailored for mobile constraints. This synthesis achieves interactive rates and stringent memory compliance, substantially advancing the deployability and quality of real-time neural rendering on commercial mobile devices (Wang et al., 4 Apr 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to NaFlex.