Skip to content

Commit b3759c8

Browse files
authored
Add link to iterative scheduling tutorial (#94)
* Add link to iterative scheduling tutorial * Review comments
1 parent 30dea5c commit b3759c8

File tree

3 files changed

+6
-3
lines changed

3 files changed

+6
-3
lines changed

Conceptual_Guide/Part_6-building_complex_pipelines/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -28,8 +28,8 @@
2828

2929
# Building Complex Pipelines: Stable Diffusion
3030

31-
| Navigate to | [Part 5: Building Model Ensembles](../Part_5-Model_Ensembles/) | [Documentation: BLS](https://github.com/triton-inference-server/python_backend#business-logic-scripting) |
32-
| ------------ | --------------- | --------------- |
31+
| Navigate to | [Part 5: Building Model Ensembles](../Part_5-Model_Ensembles/) | [Part 7: Iterative Scheduling Tutorial](./Part_7-iterative_scheduling) | [Documentation: BLS](https://github.com/triton-inference-server/python_backend#business-logic-scripting) |
32+
| ------------ | --------------- | --------------- | --------------- |
3333

3434
**Watch [this explainer video](https://youtu.be/JgP2WgNIq_w) with discusses the pipeline, before proceeding with the example**. This example focuses on showcasing two of Triton Inference Server's features:
3535
* Using multiple frameworks in the same inference pipeline. Refer [this for more information](https://github.com/triton-inference-server/backend#where-can-i-find-all-the-backends-that-are-available-for-triton) about supported frameworks.

Conceptual_Guide/Part_7-iterative_scheduling/README.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -28,6 +28,9 @@
2828

2929
# Deploying a GPT-2 Model using Python Backend and Iterative Scheduling
3030

31+
| Navigate to | [Part 6: Building Complex Pipelines: Stable Diffusion](../Part_6-building_complex_pipelines) | [Documentation: Iterative Scheduling](https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/user_guide/model_configuration.html#iterative-sequences) |
32+
| ------------ | --------------- | --------------- |
33+
3134
In this tutorial, we will deploy a GPT-2 model using the Python backend and
3235
demonstrate the
3336
[iterative scheduling](https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/user_guide/model_configuration.html#iterative-sequences)

Conceptual_Guide/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,4 +39,4 @@ Conceptual guides have been designed as an onboarding experience to Triton Infer
3939
* [Part 4: Accelerating Models](Part_4-inference_acceleration/): Another path towards achieving higher throughput is to accelerate the underlying models. This guide covers SDKs and tools which can be used to accelerate the models.
4040
* [Part 5: Building Model Ensembles](./Part_5-Model_Ensembles/): Models are rarely used standalone. This guide will cover "how to build a deep learning inference pipeline?"
4141
* [Part 6: Using the BLS API to build complex pipelines](Part_6-building_complex_pipelines/): Often times there are scenarios where the pipeline requires control flows. Learn how to work with complex pipelines with models deployed on different backends.
42-
42+
* [Part 7: Iterative Scheduling Tutorial](./Part_7-iterative_scheduling): Shows how to use the Triton Iterative Scheduler with a GPT2 model using HuggingFace Transformers.

0 commit comments

Comments
 (0)