Call for Papers for the vLLM Featured Track at Ray Summit is now Open!
Submit your presentation by July 30
Anyscale
Toggle menu
Product
Solutions
Examples
Resources
Company
Pricing
Get Started with $100 Credit
Featured Posts and News
05 . 30 . 2025
Ray Summit 2025: Call for Proposals Closes on July 14th, 2025
By
Christina Zhu
04 . 09 . 2025
Simplifying AI Development at Scale: Google Cloud Integrates Anyscale's RayTurbo with GKE
By
Gabe Monroy, VP & GM Cloud Runtimes
and
By
Keerti Melkote, CEO of Anyscale
10 . 03 . 2024
Ray Summit 2024: Breaking Through the AI Complexity Wall
By
The Anyscale Team
All Types
05 . 20 . 2025
New: Joins & Hash-Shuffle in Ray Data
By
Alexey Kudinkin
,
Praveen Gorthy
and
Richard Liaw
05 . 20 . 2025
Streamline Distributed AI Monitoring and Debugging with New Ray Train & Ray Data Dashboards in Anyscale
By
Alan Guo
,
Cuong Nguyen
,
Justin Yu
,
Matthew Deng
,
Matthew Owen
and
Richard Liaw
05 . 20 . 2025
RayTurbo Data Improvements Deliver Up to 5x Faster Data Processing for AI Workloads
By
Alexey Kudinkin
,
Hao Chen
,
Praveen Gorthy
and
Richard Liaw
10 . 01 . 2024
Ray Data GA
By
Hao Chen
,
Richard Liaw
and
Praveen Gorthy
10 . 01 . 2024
Accelerated Metadata Fetching in Ray Data up to 4.5x Faster on Anyscale
By
Balaji Veeramani
,
Hao Chen
,
Richard Liaw
,
Matthew Connor
and
Praveen Gorthy
10 . 01 . 2024
Autoscaling Large AI Models up to 5.1x Faster on Anyscale
By
Christopher Chou
,
Austin Kuo
,
Richard Liaw
,
Edward Oakes
and
Chris Sivanich
10 . 01 . 2024
Announcing RayTurbo
By
Akshay Malik
,
Praveen Gorthy
and
Richard Liaw
10 . 01 . 2024
Batch LLM Inference on Anyscale slashes AWS Bedrock costs by up to 6x
By
Cody Yu
,
Scott Lee
,
Ricky Xu
,
William Lin
,
Praveen Gorthy
and
Richard Liaw
11 . 07 . 2023
Ray 2.8 features Ray Data extensions, AWS Neuron cores support, and Dashboard improvements
By
Jules S. Damji
and
Richard Liaw
09 . 18 . 2023
Ray 2.7 features major stability improvements to Ray AI Libraries and KubeRay and introduces RayLLM
By
Jules S. Damji
and
Richard Liaw
07 . 25 . 2023
Ray 2.6 features streaming for Serve and Train and new Multi-GPU Learner API
By
Jules S. Damji
and
Richard Liaw
06 . 22 . 2023
How continuous batching enables 23x throughput in LLM inference while reducing p50 latency
By
Cade Daniel
,
Chen Shen
,
Eric Liang
and
Richard Liaw
1
2
Next