Anyscale Connect

Fast and efficient hyperparameter tuning with Ray Tune

Wednesday, October 20, 4:00PM UTC

Hyperparameter tuning or optimization is used to find the best performing machine learning (ML) model by exploring and optimizing the model hyperparameters (eg. learning rate, tree depth, etc). It is a compute-intensive problem that lends itself well to distributed execution.

Ray Tune is a Python library, built on Ray, that allows you to easily run distributed hyperparameter tuning at scale. Ray Tune is framework-agnostic and supports all the popular training frameworks including PyTorch, TensorFlow, XGBoost, LightGBM, and Keras.

Join this webinar with Will Drevo, product manager for Ray machine learning libraries, for an overview of Ray Tune and demo of using it for tuning a deep learning model.

Register to Watch

We will showcase many Ray Tune highlights, including how to: 

  • Set up distributed hyperparameter search in under 10 lines of code

  • Scale from a single machine to a cluster with minimal code changes

  • Trial leading search methods (ASHA, BOHB, PBT, etc) with built-in access

  • Visualize results with TensorBoard or MLflow

We will also share stories of users that are finding the most-performant models, while saving compute costs and maximizing CPU/GPU utilization with Ray Tune. 


Will Drevo

Will Drevo

Product Manager, Anyscale