Skip to content

Conversation

@isazi
Copy link
Collaborator

@isazi isazi commented Aug 13, 2025

Working on a simple parallel runner that uses Ray to distribute the benchmarking of different configurations to remote Ray workers.

@isazi isazi self-assigned this Aug 13, 2025
@isazi isazi marked this pull request as draft August 13, 2025 09:39
@sonarqubecloud
Copy link

@stijnh stijnh changed the title Simple parallel runner Add parallel tuning on multiple remote GPUs using Ray Jan 19, 2026
@stijnh stijnh self-assigned this Jan 19, 2026
@sonarqubecloud
Copy link

@stijnh
Copy link
Member

stijnh commented Jan 20, 2026

The current parallel runner works. I've been able to run on multiple GPUs on DAS6-VU and DAS6-Leiden.

There are several remaining problems:

  • The timings are incorrect as the host assumes that the total time is just the sum over individual configurations
  • Use of tuning_options need to be refactored, as now it the entire object is sent to every node for each benchmark job
  • Logging information can be improved
  • The strategies are not parallel-aware yet (except brute-force)
  • A guide needs to be added to the docs explaining how to launch a Ray cluster on DAS6

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants