Papers
Topics
Authors
Recent
2000 character limit reached

GOGH: Correlation-Guided Orchestration of GPUs in Heterogeneous Clusters (2510.15652v1)

Published 17 Oct 2025 in cs.DC and cs.LG

Abstract: The growing demand for computational resources in machine learning has made efficient resource allocation a critical challenge, especially in heterogeneous hardware clusters where devices vary in capability, age, and energy efficiency. Upgrading to the latest hardware is often infeasible, making sustainable use of existing, mixed-generation resources essential. In this paper, we propose a learning-based architecture for managing machine learning workloads in heterogeneous clusters. The system operates online, allocating resources to incoming training or inference requests while minimizing energy consumption and meeting performance requirements. It uses two neural networks: the first provides initial estimates of how well a new model will utilize different hardware types and how it will affect co-located models. An optimizer then allocates resources based on these estimates. After deployment, the system monitors real performance and uses this data to refine its predictions via a second neural network. This updated model improves estimates not only for the current hardware but also for hardware not initially allocated and for co-location scenarios not yet observed. The result is an adaptive, iterative approach that learns over time to make more effective resource allocation decisions in heterogeneous deep learning clusters.

Summary

  • The paper introduces GOGH, a framework that uses neural networks for initial throughput estimation and dynamic GPU scheduling in heterogeneous clusters.
  • It integrates an ILP-based optimizer with real-time performance feedback to minimize energy consumption while ensuring throughput guarantees.
  • Experimental results show that pairing an RNN for initial estimation with a Feedforward model for refinement significantly improves scheduling accuracy and resource utilization.

GOGH: Efficient GPU Orchestration in Heterogeneous Clusters

Introduction

The paper "GOGH: Correlation-Guided Orchestration of GPUs in Heterogeneous Clusters" (2510.15652) proposes a framework named GOGH to address the challenge of efficiently managing computational workloads in heterogeneous clusters. It provides a solution for dynamic resource allocation by leveraging historical data to inform scheduling decisions and optimize both performance and energy efficiency.

Framework Overview

GOGH orchestrates machine learning jobs across heterogeneous GPU clusters by utilizing the predictive power of neural networks. It makes informed scheduling decisions by considering the correlations between jobs and GPU types in terms of throughput performance.

Architectural Design

GOGH's architecture consists of three core modules: initial throughput estimation, optimization of job assignments, and refinement of throughput predictions. Figure 1

Figure 1: The architecture of GOGH.

Initial Throughput Estimation: This module employs a neural network, referred to as P1P_1, which predicts the throughput of a newly arrived job based on historical performance data of similar jobs. The estimates are refined iteratively using real-time measurements captured during the job's execution.

Optimizer: The optimization module formulates the GPU allocation problem as an integer linear programming (ILP) challenge. It aims to assign jobs to GPUs such that overall system energy consumption is minimized while ensuring throughput guarantees.

Estimation Refinement: Real performance data gathered during job execution is utilized by another neural network, P2P_2, to update and refine throughput estimates. This enables continuous improvement in scheduling accuracy.

Notations and Formulations

The paper introduces several essential notations to define the resource allocation problem formally. These include sets representing servers, GPU types, jobs, and job combinations. The objective is to minimize the total energy consumption while satisfying job-specific and system-wide constraints.

Problem 1 outlines the ILP formulation that guides resource allocation, considering throughput requirements and accelerator capacity constraints.

Evaluation

The effectiveness of GOGH was validated using a comprehensive dataset featuring a variety of machine learning workloads on heterogeneous GPU configurations. The experimental setup illustrated GOGH's ability to improve job scheduling efficiency and throughput prediction accuracy significantly.

Neural Network Models: Three neural network architectures—Feedforward (FF), Recurrent Neural Network (RNN), and Transformer—were employed to analyze their impact on prediction accuracy. Figure 2

Figure 2

Figure 2: Initial NN (P1).

Performance Results: The RNN architecture demonstrated superior predictive performance in the initial estimation phase (P1) with lower mean absolute error (MAE) than the transformer and FF models. For the refinement phase (P2), the Feedforward model exhibited consistent accuracy improvements over the RNN and Transformer. Figure 3

Figure 3: Combined validation results of P1-P2 model pairs.

End-to-End Scheduling: The combination of RNN for initial estimation and Feedforward for refinement resulted in the most accurate scheduling predictions, lowering energy consumption and enhancing resource utilization efficiency compared to alternative model pairings.

The paper situates GOGH within the broader context of recent literature on managing heterogeneous hardware. It reviews approaches for heterogeneity-aware scheduling, resource sharing, and elastic training platforms, among others.

Key contributions from works such as Gavel [narayanan2020heterogeneity], Pollux [qiao2021pollux], and ElasticFlow [gu2023elasticflow] are contrasted with GOGH's approach, particularly in terms of adaptive scheduling and dynamic resource management.

Conclusion

GOGH introduces a powerful framework for optimizing GPU orchestration in heterogeneous clusters, leveraging neural networks for data-driven scheduling decisions. It effectively balances throughput maximization with energy efficiency, addressing the fundamental challenges posed by mixed hardware generations. Future work may extend the ILP solution with approximation or reinforcement learning techniques to further enhance computational efficiency and scalability.

The insights and methodologies provided by GOGH hold substantial potential for deployment in real-world heterogeneous environments, facilitating cost-effective and sustainable machine learning operations.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We found no open problems mentioned in this paper.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.