Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
139 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Compiling Crossing-free Geometric Graphs with Connectivity Constraint for Fast Enumeration, Random Sampling, and Optimization (2001.08899v1)

Published 24 Jan 2020 in cs.CG

Abstract: Given $n$ points in the plane, we propose algorithms to compile connected crossing-free geometric graphs into directed acyclic graphs (DAGs). The DAGs allow efficient counting, enumeration, random sampling, and optimization. Our algorithms rely on Wettstein's framework to compile several crossing-free geometric graphs. One of the remarkable contributions of Wettstein is to allow dealing with geometric graphs with connectivity, since it is known to be difficult to efficiently represent geometric graphs with such global property. To achieve this, Wettstein proposed specialized techniques for crossing-free spanning trees and crossing-free spanning cycles and invented compiling algorithms running in $\mathrm{O}(7.044n)$ time and $\mathrm{O}(5.619n)$ time, respectively. Our first contribution is to propose a technique to deal with the connectivity constraint more simply and efficiently. It makes the design and analysis of algorithms easier, and yields improved time complexity. Our algorithms achieve $\mathrm{O}(6n)$ time and $\mathrm{O}(4n)$ time for compiling crossing-free spanning trees and crossing-free spanning cycles, respectively. As the second contribution, we propose an algorithm to optimize the area surrounded by crossing-free spanning cycles. To achieve this, we modify the DAG so that it has additional information. Our algorithm runs in $\mathrm{O}(4.829n)$ time to find an area-minimized (or maximized) crossing-free spanning cycle of a given point set. Although the problem was shown to be NP-complete in 2000, as far as we know, there were no known algorithms faster than the obvious $\mathrm{O}(n!)$ time algorithm for 20 years.

Citations (1)

Summary

We haven't generated a summary for this paper yet.