PercHead: 3D Head Reconstruction & Editing
- PercHead is a method for converting a single RGB face image into a complete 3D head model with semantic editing capabilities, integrating perceptual supervision and a dual-branch transformer architecture.
- It achieves robust novel view synthesis and precise identity preservation by leveraging cross-attention between 2D encoder features and 3D patch embeddings, enhanced through differentiable Gaussian splatting.
- Its semantic editing feature enables independent manipulation of head geometry and style using segmentation maps and CLIP-inspired text prompts, offering high flexibility and control.
PercHead is a method for single-image 3D head reconstruction and semantic 3D editing that integrates perceptual supervision, dual-branch transformer architectures, and differentiable Gaussian splatting for high-fidelity and view-consistent output. It enables transforming a single RGB face image into a full 3D head representation and further supports semantic editing—where head geometry and appearance can be independently manipulated through segmentation maps and prompts—surpassing previous models in novel view synthesis, robustness to occlusions, and editing flexibility (Oroz et al., 4 Nov 2025).
1. Model Architecture
PercHead leverages a dual-branch encoder and a transformer-based 3D decoder to lift 2D appearance signals into a unified 3D head representation.
- Input Preprocessing: Each RGB face image is cropped and background-masked with a tracker adapted from GAGAvatar.
- Dual-Branch Encoder:
- Branch 1: A frozen DINOv2 backbone extracts dense patch features from intermediate layers (), capturing multi-scale semantic and low-level structure.
- Branch 2: A lightweight ViT (inspired by MAE), trained from scratch, captures fine-grained, context-specific features not available in DINOv2.
- Background patches (approx. 30%) are masked out to reduce computation. Surviving patches' feature vectors from both branches are concatenated and projected into the decoder embedding space via an MLP:
where is the decoder dimension (typically 512).
ViT-Based 3D Decoder with Iterative Cross-Attention:
- The head is initialized from a fixed FLAME mesh (k vertices), partitioned into 4096 3D patches, each with 16 vertices and a learnable embedding.
- Each decoder layer () applies cross-attention between current 3D patch embeddings () and 2D encoder features (), then refines with a per-patch MLP and residual connections:
- Cross-attention equations (fully stated):
where , , and is the key dimension.
3D-to-3D self-attention is deliberately omitted to limit computational burden; coherence arises from global cross-attention to the shared 2D context.
2. Perceptual Supervision and Losses
PercHead eschews pixel-wise losses in favor of multi-layer perceptual signals:
Supervision Sources:
- DINOv2 (layers 8, 11): Provides high-frequency and semantic correspondence signals, promoting accurate facial identity and detail.
- SAM 2.1 image encoder: Yields segmentation-oriented features emphasizing correct geometry and part delineation, e.g., eyes, hairline, mouth region.
- Loss Formulation:
- Training is driven by purely perceptual loss:
where is the -th feature extractor and is the loss weight. - After freezing the 3D decoder, a 2D CNN "refinement head" is trained with a combination of , LPIPS, and the above perceptual losses:
3. Differentiable Rendering via Gaussian Splatting
For efficient 3D-to-2D synthesis, PercHead utilizes differentiable Gaussian splatting:
Each of the 4096 decoded 3D patches is upsampled to 16 Gaussians via PixelShuffle, yielding Gaussians.
Each Gaussian is parameterized by:
- Mean
- Covariance (diagonal scale and rotation)
- Color
- Opacity
- The contribution of Gaussian to pixel :
where is the projection of the 3D center.
- Final color is computed as . This result is sharpened with a lightweight 2D CNN.
4. Semantic 3D Editing: Disentangling Geometry and Style
PercHead supports semantic editing by encoder swapping and careful input modality disentanglement:
- Editing Encoder:
- Geometry is prescribed by a 19-channel FARL segmentation map (semantic part labels).
- Appearance ("style") is specified either by a CLIP-encoded text prompt or a reference image; the global 512D CLIP style token is appended to each patch embedding.
- Editing Decoder: Remains unchanged from the base reconstruction model.
- Disentanglement Mechanism:
- Geometry is determined solely by the segmentation map, with no influence from CLIP style.
- Style is governed exclusively by the CLIP embedding, having no spatial effect on geometry.
- This approach achieves clean separation, with the model attending to the correct modality for each subtask.
- Editing-Specific Loss: The base perceptual loss is sufficient; an optional CLIP-guided loss may be used, but is not required.
5. Quantitative and Qualitative Performance
5.1. Novel View Synthesis
PercHead demonstrates strong results on Ava-256 (novel-view benchmarks, 11 heads × 5 views) and NeRSemble (5 heads × 16 views):
| Method | PSNR (Ava-256) | LPIPS (Ava-256) | ArcFace (Ava-256) | PSNR (NeRSemble) | LPIPS (NeRSemble) | ArcFace (NeRSemble) |
|---|---|---|---|---|---|---|
| PercHead | 16.08 | 0.2666 | 0.2935 | 18.04 | 0.1854 | 0.2559 |
| GAGAvatar | 15.87 | 0.2739 | 0.3481 | 16.88 | 0.2169 | 0.2883 |
Lower LPIPS and ArcFace values reflect greater perceptual and identity fidelity.
5.2. Extreme View Robustness
On challenging extreme left/right and vertical angles:
| Method | PSNR (Ava-256) | LPIPS (Ava-256) | ArcFace (Ava-256) |
|---|---|---|---|
| PercHead | 15.58 | 0.2866 | 0.2812 |
| GAGAvatar | 13.54 | 0.3643 | 0.5228 |
Minimal degradation is observed for PercHead relative to established baselines.
5.3. Qualitative and Video Consistency
Frame-by-frame analysis and visualizations demonstrate identity-preserving, detail-faithful renderings—maintained across time and pose transitions in both image and video domains.
6. Implementation Considerations and Limitations
- Data: PercHead is trained on Nersemble (real multi-view), Cafca (synthetic multi-view), and FFHQ (single-view) datasets.
- Computation:
- Base lifting model: 70 h on one RTX 3090 (perceptual losses only).
- CNN refinement: additional 24 h.
- Editing fine-tune: 30 h.
- Optimizer: AdamW, standard hyperparameters.
- Runtime: Inference requires several seconds per frame, dominated by Gaussian splatting and CNN stages.
- Limitations:
- No explicit dynamic expression transfer (e.g., reenactment is unsupported).
- Absence of real-time performance.
- Inability to relight: lighting is entangled in the input image; isolated editing is not possible.
- Failure cases can arise from extreme occlusions or nonhuman accessories.
7. Context and Significance
PercHead embodies a convergence of self-supervised vision, modern transformer design, and differentiable rendering. Its dual-branch encoder decouples global, semantic, and local, fine-grained cues; iterative cross-attention bridges 2D-to-3D relationships without computationally expensive 3D-3D self-attention. Perceptual supervision—eschewing MSE, , or SSIM losses except for final CNN refinement—directly optimizes fidelity in the embedding space of robust vision backbones (DINOv2, SAM2.1).
The semantic editing variant operationalizes the disentanglement of head geometry and style—enabling interactive tools where geometry and appearance can be sculpted and styled independently by end-users, with low parameter overhead.
Compared to classical HRTF-individualization pipelines essential for personalized binaural audio, PercHead’s pipeline—rapid smartphone-based scans, fast mesh-based simulation, learned latent compression and perceptual tuning—suggests practical deployment routes for audio-visual personalization systems (Guezenoc et al., 2020). While PercHead itself focuses on visual 3D head synthesis, its architecture and findings intersect with the future of perceptual user-specific modeling in audio, graphics, and multimodal systems.