Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 148 tok/s
Gemini 2.5 Pro 44 tok/s Pro
GPT-5 Medium 27 tok/s Pro
GPT-5 High 38 tok/s Pro
GPT-4o 85 tok/s Pro
Kimi K2 210 tok/s Pro
GPT OSS 120B 442 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

BrushNet: Interactive Neural Editing

Updated 21 October 2025
  • BrushNet is a framework for brush-based interactive editing that directly manipulates neural representations like signed distance functions and diffusion model latents.
  • It employs a brush template function and localized weight updates to achieve precise deformations while preserving global structure.
  • The method integrates with diffusion models and segmentation techniques to enable both 3D sculpting and consistent image editing.

BrushNet is a technical framework for brush-based, interactive editing in both neural 3D geometry and image synthesis applications. The underlying methodology enables localized, intuitive deformations or manipulations of implicit neural representations—most notably signed distance functions (SDFs) and diffusion model latents—via brush functions that mimic traditional art workflows but operate directly in high-dimensional, learned spaces.

1. Definition and Core Components

BrushNet generally refers to a family of approaches in which brush-like interactions—parameterized by location, radius, intensity, and profile—are applied to neural representations for direct, localized modification. These interactions occur on neural fields rather than explicit surfaces or images. The essential components include:

  • Brush template function: A compact, radially symmetric function (e.g., quintic polynomial) defines the “brush” profile.
  • Region of influence: The brush operates on a tangent space around a selected interaction point; this region is parameterized by radius and intensity.
  • Sampling and projection: Points in the region are sampled and mapped to local neural features, either on the SDF field for geometric editing or in latent/image space for visual editing.
  • Weight update mechanism: Local edits are enforced through selective network re-sampling and adaptation; weights are updated primarily to reflect the brush’s imposed deformation or appearance change.

2. Brush-Based Editing of Neural Signed Distance Functions

The methodology in "3D Neural Sculpting (3DNS): Editing Neural Signed Distance Functions" (Tzathas et al., 2022) establishes the foundational algorithm for BrushNet in neural geometry. Key technical details:

  • Brush function formulation: The brush template bT(x)b_T(\mathbf{x}) is specified as bT(x)={P(1x),x<1;0,otherwise}b_T(\mathbf{x}) = \{ P(1 - \|\mathbf{x}\|), \|\mathbf{x}\| < 1; 0, \text{otherwise} \}, with P(x)=6x515x4+10x3P(x) = 6x^5 - 15x^4 + 10x^3.
  • Parameterization and control: The full family becomes Br,s(x)=sbT(x/r)B_{r,s}(\mathbf{x}) = s \cdot b_T(\mathbf{x}/r), providing precise control over radius rr and intensity ss.
  • Local surface parameterization: The SDF zero-level set is locally characterized via the implicit function theorem; brush edits are added in 2D over the surface’s tangent plane, facilitating both displacement and normal adjustment via \nabla operations.
  • Local weight adaptation: Edits are localized by restricting training to samples in the interaction region (high density near the brush) while maintaining many “model samples” elsewhere to preserve global geometry. Samples near the brush edge on the unedited surface are filtered to avoid undesired spillover.

This approach is distinct from mesh-based sculpting—where vertices are moved directly—in that the neural network itself internalizes the deformation via re-training or fine-tuning in response to brush-induced local changes.

3. BrushNet in Diffusion Models and Cross-Modal Editing

BrushNet interfaces with diffusion-based editing frameworks for image synthesis and manipulation, as exemplified by integration with Edicho (Bai et al., 30 Dec 2024) and In-Context Brush (Xu et al., 26 May 2025):

  • Spatially coherent image editing: Edicho equips BrushNet with explicit image correspondence by integrating external extractors (e.g., DIFT, Dust3R). Attention queries in the network are “warped” via precomputed correspondences, so brush edits propagate consistently across images or views.
  • Classifier-Free Guidance (CFG): Edicho introduces Corr-CFG, wherein the unconditional branch of the denoising network is modified by injecting noise from correspondingly mapped latents in source images, balancing injection strength λ\lambda and region γ\gamma.
  • In-context subject insertion: In-Context Brush reformulates customization as in-context learning, blending subject and prompt semantics into the target image at inference time. Dual-level latent manipulation includes intra-head “feature shifting” (modifying attention outputs per region) and inter-head “attention reweighting” (emphasizing relevant semantic heads).

These integrations position BrushNet as both a direct manipulation tool and a backbone for plug-and-play artistic and semantic editing within pre-trained generative pipelines.

4. Algorithmic Structures and Artistic Rendering

In artistic rendering, BrushNet-style systems leverage region-based and hierarchical strategies, utilizing segmentation, SVG vectorization, and sequential stroke processing as described in (Prudviraj et al., 11 Jun 2025):

  • Semantic segmentation (SAM): The input image is decomposed into non-overlapping, semantically meaningful regions, refined by IOU metrics and area sorting.
  • SVG vectorization and stroke parameterization: Each region is vectorized (curves, angles) and parameterized for brush strokes as {x,y,w,h,θ,r,g,b}\{x, y, w, h, \theta, r, g, b\}; large regions are subdivided, and strokes are defined by minimum rotated rectangles and clustering techniques.
  • Stroke sequence establishment: The placement order is determined by proximity-based clustering (Gestalt principles) and TSP solutions, ensuring an evolution that mirrors artistic creation (background to detail).

This process is hierarchical: scene \to segment \to vectorized region \to stroke, supporting natural, high-fidelity painting evolutions.

5. Locality, Preservation, and Interactivity

A defining attribute of BrushNet methodologies is robust locality:

  • Local deformation: Through careful sample selection and brush parameterization, edits remain sharply localized. This minimizes artifacts in surrounding geometry or image regions and supports high-precision workflow.
  • Global structure preservation: By maintaining dense “model samples” outside the interaction area and discarding samples near the edge of the brush’s influence, the global shape or image remains stable.
  • Interactive sculpting: The frameworks provide real-time feedback, simulating tactile sculpting workflows, and enable fine-grained iterative adjustments (brush radius, intensity, direction).

High-accuracy is validated quantitatively—Chamfer distances for geometry, CLIP/DINO/FID metrics for images—where BrushNet approaches consistently outperform naive or grid-based counterparts.

6. Applications and Future Directions

BrushNet is applicable in:

  • 3D digital sculpting: Artists directly manipulate neural SDFs, achieving clay-like deformations but with the resolution independence and smoothness afforded by neural priors (Tzathas et al., 2022).
  • Consistent image editing: Edicho and similar modules enable cross-image consistency in both local (inpainting) and global (semantic translation) edits across variable poses, lighting, and context (Bai et al., 30 Dec 2024).
  • Zero-shot subject insertion: In-Context Brush allows test-time customizable insertion and compositional generation in high-quality visual synthesis tasks (Xu et al., 26 May 2025).
  • Artistic rendering: Region-driven stroke evolution supports education, collaborative art, and virtual exhibitions, emphasizing not only final output fidelity but also the process itself (Prudviraj et al., 11 Jun 2025).

Expansion may include more sophisticated edge-handling, adaptive brush templates, and support for broader neural scene representations (e.g., Neural Radiance Fields), with ongoing work addressing hard edges and richer semantic control.

7. Technical Significance and Challenges

BrushNet highlights several impactful technical innovations:

  • Direct neural field manipulation: Edits happen at the neural function level, not via mesh or raster post-processing, allowing for unprecedented expressiveness.
  • Localized learning: By using carefully weighted sample sets, the influence of parameter updates is sharply confined.
  • Plug-and-play compatibility: BrushNet and related modules (e.g., Edicho) operate entirely at inference, without the need for retraining, enabling broad adoption within existing pipelines.

Limitations include handling sharp geometries, edge effects, and dependency on quality of segmentation/correspondence extraction. Addressing these issues is central to broadening application scope and further improving user control.


BrushNet thus embodies a unified interface for art-directable, interactive neural editing, enabling significant advances in both geometric deformation and image-level manipulation. Its mathematically rigorous design, combined with robust empirical validation, positions it as a versatile tool across creative, engineering, and scientific domains.

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to BrushNet.