Ray Deep Dives

Hugging Face + Ray AIR

Ray Summit 2022

Hugging Face Transformers is a popular open-source project that features state-of-the-art machine learning for PyTorch, TensorFlow, and JAX. It allows for easy access to state-of-the-art NLP, computer vision, audio, and multimodal models.

The computational and memory requirements for fine-tuning and training those models can be significant. To deal with this issue, the Ray team has developed a Hugging Face integration for Ray AI Runtime (AIR), allowing Transformers model training to be easily parallelized across multiple CPUs or GPUs in a Ray cluster, saving time and money, all the while allowing to take advantage of the rich Ray ML ecosystem thanks to a common API.

In this session, we explore the integration between Hugging Face and Ray AIR, allowing users to scale their model training and data loading seamlessly. We will dive deep into the implementation and API and explore how we can use Ray AIR to create an end-to-end Hugging Face workflow, from data ingest through fine tuning and HPO to inference and serving.

About Antoni

Antoni Baum is a software engineer at Anyscale, working on Ray Tune, XGBoost-Ray, and other ML libraries. In his spare time, he contributes to various open source projects, trying to make machine learning more accessible and approachable.

Antoni Baum

Software Engineer, Anyscale
chucks
Ray Summit 2022 horizontal logo

Ready to Register?

Come connect with the global community of thinkers and disruptors who are building and deploying the next generation of AI and ML applications.

Save your spot
register-bottom-mobile
beanbags

Join the Conversation

Ready to get involved in the Ray community before the conference? Ask a question in the forums. Open a pull request. Or share why you’re excited with the hashtag #RaySummit on Twitter.