Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Ray Serve docs - reorg Examples page #44457

Open
anyscalesam opened this issue Apr 3, 2024 · 2 comments · May be fixed by #45231
Open

Ray Serve docs - reorg Examples page #44457

anyscalesam opened this issue Apr 3, 2024 · 2 comments · May be fixed by #45231
Assignees
Labels
docs An issue or change related to documentation P2 Important issue, but not time-critical

Comments

@anyscalesam
Copy link
Collaborator

anyscalesam commented Apr 3, 2024

Description

@peytondmurray, @akshay-anyscale @angelinalg and I had a discussion on how better to organize the Ray Serve examples page.

Instead of Beginner, Intermediate, Advanced in the current hierarchy of:

Beginner
Serve ML Models
Serve a Stable Diffusion Model
Serve a Text Classification Model
Serve an Object Detection Model

Intermediate
Serve an Inference Model on AWS NeuronCores Using FastAPI
Serve an Inference with Stable Diffusion Model on AWS NeuronCores Using FastAPI
Serve a model on Intel Habana Gaudi
Scale a Gradio App with Ray Serve
Serve a Text Generator with Request Batching
Serve a Chatbot with Request and Response Streaming
Serving models with Triton Server in Ray Serve

Advanced
Serve a Java App

We want to organize it into:

ML Applications
Serve ML Models
Serve a Stable Diffusion Model
Serve a Text Classification Model
Serve an Object Detection Model
Serve a Chatbot with Request and Response Streaming

Integrations
Scale a Gradio App with Ray Serve
Serve a Text Generator with Request Batching
Serving models with Triton Server in Ray Serve
Serve a Java App

AI Accelerators
Serve an Inference Model on AWS NeuronCores Using FastAPI
Serve an Inference with Stable Diffusion Model on AWS NeuronCores Using FastAPI
Serve a model on Intel Habana Gaudi
*Eventual AMD examples and further examples for different AI Accelerator Types

Link

https://docs.ray.io/en/master/serve/examples.html

@anyscalesam anyscalesam added docs An issue or change related to documentation P1.5 Issues that will be fixed in a couple releases. It will be bumped once all P1s are cleared labels Apr 3, 2024
@anyscalesam anyscalesam changed the title [<Ray component: Core|RLlib|etc...>] Ray Serve docs - reorg Examples page Apr 3, 2024
@akshay-anyscale
Copy link
Contributor

Maybe instead of Model types we call it ML Applications .
"Serve a Text Generator with Request Batching
Serve a Chatbot with Request and Response Streaming"
should go under that

@anyscalesam anyscalesam added P2 Important issue, but not time-critical and removed P1.5 Issues that will be fixed in a couple releases. It will be bumped once all P1s are cleared labels Apr 23, 2024
@anyscalesam
Copy link
Collaborator Author

Discussed with @peytondmurray @angelinalg > currently higher priority projects with fixing internal links and raydocs logging

@peytondmurray peytondmurray linked a pull request May 9, 2024 that will close this issue
8 tasks
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
docs An issue or change related to documentation P2 Important issue, but not time-critical
Projects
None yet
Development

Successfully merging a pull request may close this issue.

3 participants