Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 124 tok/s
Gemini 2.5 Pro 52 tok/s Pro
GPT-5 Medium 25 tok/s Pro
GPT-5 High 31 tok/s Pro
GPT-4o 79 tok/s Pro
Kimi K2 206 tok/s Pro
GPT OSS 120B 435 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

Next Visual Granularity Generation (2508.12811v1)

Published 18 Aug 2025 in cs.CV, cs.AI, and cs.LG

Abstract: We propose a novel approach to image generation by decomposing an image into a structured sequence, where each element in the sequence shares the same spatial resolution but differs in the number of unique tokens used, capturing different level of visual granularity. Image generation is carried out through our newly introduced Next Visual Granularity (NVG) generation framework, which generates a visual granularity sequence beginning from an empty image and progressively refines it, from global layout to fine details, in a structured manner. This iterative process encodes a hierarchical, layered representation that offers fine-grained control over the generation process across multiple granularity levels. We train a series of NVG models for class-conditional image generation on the ImageNet dataset and observe clear scaling behavior. Compared to the VAR series, NVG consistently outperforms it in terms of FID scores (3.30 -> 3.03, 2.57 ->2.44, 2.09 -> 2.06). We also conduct extensive analysis to showcase the capability and potential of the NVG framework. Our code and models will be released.

Summary

  • The paper presents a structured, hierarchical model that decomposes images into stages for fine-to-coarse synthesis.
  • It employs a dual-generation pipeline with structure and content generators guided by RoPE and refined embeddings.
  • Empirical results on ImageNet show improved FID, Inception Score, and recall with fewer parameters.

Next Visual Granularity Generation: Structured, Hierarchical Image Synthesis

The "Next Visual Granularity Generation" (NVG) framework introduces a novel paradigm for image generation by explicitly modeling and controlling hierarchical visual structure. Unlike prior approaches that treat images as flat sequences or holistic distributions, NVG decomposes images into a sequence of stages, each corresponding to a distinct level of visual granularity. This enables structured, coarse-to-fine synthesis and fine-grained control over the generative process.

Hierarchical Visual Granularity: Representation and Construction

NVG represents an image as a sequence of content and structure pairs, each at a fixed spatial resolution but with a varying number of unique tokens. At each stage, a structure map defines the spatial arrangement of tokens, and a content vector specifies the token values. The construction of this sequence is fully data-driven: starting from the finest granularity (each spatial location is a unique token), tokens are iteratively clustered based on feature similarity, halving the number of unique tokens at each stage. This bottom-up clustering yields a hierarchy of structure maps, naturally capturing the progression from fine details to global layout. Figure 1

Figure 1

Figure 1

Figure 1

Figure 1: Construction of the visual granularity sequence on a 2562256^2 image and the next visual granularity generation in the 16216^2 latent space. Top-to-bottom: Number of unique tokens, structure map, generated image.

The content tokens for each stage are constructed in a residual manner, quantizing the difference between the current canvas and the ground truth latent. This approach, inspired by VAR but guided by the induced structure map rather than spatial downsampling, results in a more semantically meaningful decomposition and improved codebook utilization.

Structure Embedding and Hierarchical Encoding

To encode the hierarchical relationships across stages, NVG introduces a compact structure embedding scheme. Each spatial location is assigned a KK-dimensional bit-style vector, where each bit encodes the parent-child relationship at a given stage. This embedding is RoPE-compatible and enables efficient, stage-aware positional encoding for both content and structure generators. Figure 2

Figure 2: KK-dimensional structure embedding encodes parent-child relations across all stages, with padding and bit patterns distinguishing clusters and stages.

Generation Pipeline: Stage-wise Structure and Content Synthesis

The NVG generation pipeline alternates between structure and content generation at each stage. At each step, the structure generator predicts the structure map for the current granularity, conditioned on the input text, the current canvas, and the hierarchical structure embedding. The content generator then predicts the content tokens for the current structure, refining the canvas towards the final image. Figure 3

Figure 3: At each stage, structure is generated first, followed by content, both conditioned on text, current canvas, and hierarchical structure.

The structure generator employs a lightweight rectified flow model, treating structure prediction as a structure inpainting task. Known parts of the structure embedding are fixed, and the model predicts the remaining unknowns. Gumbel-top-kk sampling is used to increase diversity in structure assignment. The content generator uses a transformer-based architecture with structure-aware RoPE, predicting the final canvas and the content tokens for the current stage. Training is supervised by both MSE loss on the canvas and cross-entropy loss on the content tokens.

Empirical Results: Quantitative and Qualitative Analysis

NVG models are evaluated on class-conditional ImageNet generation, demonstrating strong scaling behavior and outperforming the VAR baseline in FID, Inception Score, and recall across all model sizes. Notably, NVG achieves these results with fewer training steps and parameters compared to many state-of-the-art models. Figure 4

Figure 4: Iterative generation process (top), alignment of structure maps with final images (middle), and diverse, high-quality samples from NVG-d24 (bottom).

Ablation studies confirm the importance of autoregressive content modeling, partial noise in structure inpainting, and structure-aware RoPE. Direct prediction of the next content token (as opposed to the final canvas) leads to overfitting and degraded performance, highlighting the benefit of richer supervision.

Structure-Guided and Stage-wise Controlled Generation

A key advantage of NVG is explicit structure control. By providing custom structure maps—such as geometric shapes or reference image structures—users can guide the generation process, enabling flexible and interpretable manipulation of image layout and semantics. Figure 5

Figure 5: Structure-guided generation using geometric binary structure maps and reference image structures.

Stage-wise control experiments demonstrate that fixing structure and/or content at different stages allows for precise manipulation of layout, pose, and fine details. The model exhibits strong error-correction capabilities, enabling it to adapt to out-of-distribution class labels even when early-stage structure/content is fixed. Figure 6

Figure 6: Stage-wise controlled generation: top—reference reconstruction; middle—content and structure fixed; bottom—structure only fixed, with in-domain and OOD class guidance.

Theoretical and Practical Implications

NVG addresses a key limitation of prior generative models: the lack of explicit, hierarchical structure modeling. By decomposing images into sequences of increasing granularity and guiding generation through structure-aware mechanisms, NVG enables:

  • Fine-grained, interpretable control over the generative process at multiple levels of abstraction.
  • Improved codebook utilization and more semantically meaningful tokenization compared to scale-based approaches.
  • Efficient, few-step generation with strong scaling properties, making it suitable for high-resolution synthesis.
  • Error correction and flexible adaptation to new conditions, surpassing the rigidity of autoregressive models.

Future Directions

The explicit structure modeling in NVG opens several avenues for further research:

  • Region-aware generation: Leveraging domain-specific annotations to define custom granularity sequences for targeted control.
  • Physically-aware video generation: Tracking structured regions over time to enforce temporal coherence and physical constraints.
  • Hierarchical spatial reasoning: Integrating NVG with spatial reasoning models to enable divide-and-conquer inference in visual tasks.

Conclusion

The NVG framework represents a significant advance in structured image generation, providing a principled approach to hierarchical decomposition and control. Its empirical performance, scalability, and flexibility suggest broad applicability in domains where structure and interpretability are paramount. The explicit separation of structure and content, combined with stage-wise generation, offers a new foundation for controllable, efficient, and semantically meaningful generative modeling.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 6 tweets and received 150 likes.

Upgrade to Pro to view all of the tweets about this paper:

Youtube Logo Streamline Icon: https://streamlinehq.com
Reddit Logo Streamline Icon: https://streamlinehq.com

alphaXiv

  1. Next Visual Granularity Generation (41 likes, 0 questions)