Skip to content

Conversation

@MelissaFlinn
Copy link
Contributor

@MelissaFlinn MelissaFlinn commented Nov 7, 2025

Summary by CodeRabbit

  • Documentation
    • Expanded model customization docs: new end-to-end workflow, environment setup, data preparation, and training guidance.
    • New guides for synthetic data generation (SDG Hub), example notebooks, and exploring Training Hub/SDGHub examples.
    • Added how-tos for building custom workbench images, cloning example repositories, mirroring the Python index, OSFT vs SFT comparison, memory estimation, and support philosophy.

@coderabbitai
Copy link

coderabbitai bot commented Nov 7, 2025

Walkthrough

Adds multiple AsciiDoc assembly and module documents for a model customization workflow: 4 new assemblies and ~12 new modules covering environment setup, Python index/mirroring, container image builds, data preparation (including synthetic data), training, examples, and support guidance. Integrates these into the main customization guide.

Changes

Cohort / File(s) Summary
New Assembly Modules
assemblies/generate-synthetic-data-to-augment-real-data.adoc, assemblies/prepare-your-data-for-ai-consumption.adoc, assemblies/set-up-your-working-environment.adoc, assemblies/train-the-model-by-using-your-prepared-data.adoc
Adds four assembly documents that define ASSEMBLY metadata, contextual parent guards, ids/titles, and include sequences for nested modules (leveloffsets and conditional preview/content blocks).
Environment & Python Index Modules
modules/about-the-python-index.adoc, modules/mirror-the-python-index.adoc, modules/build-a-custom-container-image.adoc
New concept pages describing the org Python index, how to mirror it for disconnected environments, and instructions/examples for building custom container images preconfigured to use the org Python index (including package install examples).
Example & Exploration Modules
modules/clone-an-example-git-repository.adoc, modules/explore-the-sdg-hub-examples.adoc, modules/explore-the-training-hub-examples.adoc, modules/import-example-notebooks.adoc
Adds procedural/guide pages for cloning example repos in the workbench, exploring SDG Hub and Training Hub example notebooks and YAML flows, and importing example notebooks with repository/branch paths.
Training & Workflow Modules
modules/compare-the-performance-of-osft-and-sft.adoc, modules/estimate-memory-usage.adoc, modules/end-to-end-model-customization-workflow.adoc, modules/overview-of-the-model-customization-workflow.adoc, modules/support-philosophy.adoc
Introduces conceptual guidance on OSFT vs SFT comparisons, memory estimation references, an end-to-end customization workflow, an overall workflow overview, and a support philosophy document for secure/tooling scope.
Main Documentation Updates
customize-models-to-build-gen-ai-applications.adoc
Replaces commented/legacy assembly includes with active includes that reference the new assemblies and modules, reorganizing the customization guide to reference the new workflow content.

Estimated code review effort

🎯 2 (Simple) | ⏱️ ~12 minutes

  • Documentation-only changes with consistent AsciiDoc patterns (assembly/module metadata, includes, conditional blocks).
  • Pay attention to correctness of include paths, ids, and conditional parent-context blocks when reviewing.

Pre-merge checks and finishing touches

✅ Passed checks (3 passed)
Check name Status Explanation
Description Check ✅ Passed Check skipped - CodeRabbit’s high-level summary is enabled.
Title check ✅ Passed The title 'odh-adding customize models content' accurately reflects the main change: adding comprehensive documentation content for customizing models, including 14+ new assembly and module files organized around model customization workflows.
Docstring Coverage ✅ Passed No functions found in the changed files to evaluate docstring coverage. Skipping docstring coverage check.
✨ Finishing touches
🧪 Generate unit tests (beta)
  • Create PR with unit tests
  • Post copyable unit tests in a comment

Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share

Comment @coderabbitai help to get the list of available commands and usage tips.

Copy link

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 6

🧹 Nitpick comments (4)
modules/about-the-python-index.adoc (1)

12-12: Minor: Inconsistent capitalization in table header.

Line 12 uses "UBi9" which appears inconsistent with standard naming conventions. Consider standardizing to "UBI9" or "UBI 9" for clarity.

modules/estimate-memory-usage.adoc (1)

8-9: Consider converting file references to actual links.

Lines 8-9 reference code files as inline code but don't provide clickable links. Consider whether these should be external links to the repository or if they serve as documentation references only.

modules/import-example-notebooks.adoc (1)

15-34: Inconsistent URL formatting in table.

Line 16 uses proper AsciiDoc link: syntax, but lines 21, 26, and 31 use backticks without link: syntax for repository URLs. For consistency and to ensure clickable links in rendered documentation, apply the same formatting pattern to all repository URLs. Consider whether all should use the link: syntax or if backticks are intentional.

Apply this diff to standardize URL formatting:

-|`https://github.com/Red-Hat-AI-Innovation-Team/sdg_hub.git`
+|link:https://github.com/Red-Hat-AI-Innovation-Team/sdg_hub.git[https://github.com/Red-Hat-AI-Innovation-Team/sdg_hub.git]

-|`https://github.com/Red-Hat-AI-Innovation-Team/training_hub.git`
+|link:https://github.com/Red-Hat-AI-Innovation-Team/training_hub.git[https://github.com/Red-Hat-AI-Innovation-Team/training_hub.git]

-|`https://github.com/red-hat-data-services/red-hat-ai-examples.git`
+|link:https://github.com/red-hat-data-services/red-hat-ai-examples.git[https://github.com/red-hat-data-services/red-hat-ai-examples.git]
modules/compare-the-performance-of-osft-and-sft.adoc (1)

24-26: Improve readability of performance comparison explanations.

Lines 24-26 contain dense, complex technical information with overly long sentences (each exceeds 200 characters). Break these into shorter, more digestible segments with better structural formatting. Consider using separate bullet points or definition lists for each metric.

For example, for the memory scaling explanation, separate the concept definition from the formula and the practical implication:

 * *Memory scaling:* OSFT memory scales linearly with the unfreeze rank ratio (URR) which is a hyperparameter for OSFT that is a value between 0 and 1 representing the fraction of the matrix rank that is unfrozen and updated during fine-tuning. A rough comparison can be expressed as OSFT Memory ~ 3r times SFT Memory where r is the URR unfreeze rank ratio — the fraction of the matrix being fine-tuned. At URR = 1/3, OSFT and SFT have similar memory usage. But in most post-training setups, URR values below 1/3 are sufficient for learning new tasks, making OSFT notably lighter in memory.
+* *Memory scaling:* OSFT memory scales linearly with the unfreeze rank ratio (URR), a hyperparameter representing the fraction of the matrix rank that is unfrozen and updated during fine-tuning. URR is a value between 0 and 1.
+
+The rough memory comparison is: OSFT Memory ≈ 3r × SFT Memory, where r is the URR value.
+
+At URR = 1/3, OSFT and SFT have similar memory usage. However, in most post-training setups, URR values below 1/3 are sufficient for learning new tasks, making OSFT notably lighter in memory.

Similarly, restructure the training time explanation for clarity.

📜 Review details

Configuration used: Path: .coderabbit.yaml

Review profile: CHILL

Plan: Pro

📥 Commits

Reviewing files that changed from the base of the PR and between 18bde6b and 6c501da.

📒 Files selected for processing (17)
  • assemblies/generate-synthetic-data-to-augment-real-data.adoc (1 hunks)
  • assemblies/prepare-your-data-for-ai-consumption.adoc (1 hunks)
  • assemblies/set-up-your-working-environment.adoc (1 hunks)
  • assemblies/train-the-model-by-using-your-prepared-data.adoc (1 hunks)
  • customize-models-to-build-gen-ai-applications.adoc (1 hunks)
  • modules/about-the-python-index.adoc (1 hunks)
  • modules/build-a-custom-container-image.adoc (1 hunks)
  • modules/clone-an-example-git-repository.adoc (1 hunks)
  • modules/compare-the-performance-of-osft-and-sft.adoc (1 hunks)
  • modules/end-to-end-model-customization-workflow.adoc (1 hunks)
  • modules/estimate-memory-usage.adoc (1 hunks)
  • modules/explore-the-sdg-hub-examples.adoc (1 hunks)
  • modules/explore-the-training-hub-examples.adoc (1 hunks)
  • modules/import-example-notebooks.adoc (1 hunks)
  • modules/mirror-the-python-index.adoc (1 hunks)
  • modules/overview-of-the-model-customization-workflow.adoc (1 hunks)
  • modules/support-philosophy.adoc (1 hunks)
🔇 Additional comments (12)
assemblies/prepare-your-data-for-ai-consumption.adoc (1)

1-20: Assembly structure looks good.

The module declaration, context handling, conditional blocks, and includes follow the proper AsciiDoc assembly patterns. The leveloffset=+1 for nested modules is correct.

modules/mirror-the-python-index.adoc (1)

12-12: Verify test URL is intentional.

Line 12 uses cuda-ubi9-test in the URL. Confirm whether this test endpoint is correct for production documentation, or if it should match the production endpoint (cuda-ubi9) referenced in modules/about-the-python-index.adoc.

modules/support-philosophy.adoc (1)

1-30: Well-structured support philosophy document.

The document clearly articulates support scope, benefits, and encourages appropriate community engagement. Proper use of AsciiDoc formatting with clear sections and bullet points. All attribute references appear consistent with documentation standards.

modules/overview-of-the-model-customization-workflow.adoc (1)

12-15: Clarify commented-out xref and add missing cross-reference.

Line 13 has a commented-out xref for "prepare-your-data-for-ai-consumption". Either activate this xref or explain why it's commented out. Additionally, line 15 mentions "Automate data processing steps by building AI pipelines" but lacks a cross-reference—consider whether this should link to relevant documentation.

Verify that:

  1. The xref at line 13 should be active (referencing prepare-your-data-for-ai-consumption)
  2. The text at line 15 should have a corresponding xref or link
  3. All xref IDs exist in the referenced assemblies/modules
assemblies/generate-synthetic-data-to-augment-real-data.adoc (1)

1-15: Assembly structure is well-formed.

The module declaration, context handling, and include directive follow proper AsciiDoc assembly patterns. The description clearly explains SDG Hub functionality, and the leveloffset is correctly set for nested content inclusion.

modules/end-to-end-model-customization-workflow.adoc (1)

1-8: LGTM. Appropriate introductory concept module with well-formed external links.

modules/clone-an-example-git-repository.adoc (1)

11-11: Verify Table 2 reference. Line 11 references "Table 2" but the table definition is not visible in this module. Confirm this table exists in the source document or clarify the intent.

modules/explore-the-sdg-hub-examples.adoc (1)

1-45: LGTM. Well-structured procedure module with clear prerequisites, comprehensive steps, and proper cross-references. Code examples and resource links are properly formatted.

assemblies/set-up-your-working-environment.adoc (1)

1-19: LGTM. Assembly structure is correct with proper context handling and module includes. The intentional leveloffset=+2 on line 16 establishes the correct document hierarchy.

modules/build-a-custom-container-image.adoc (1)

6-6: Verify Table 1 reference. Line 6 references "Table 1" but the table definition is not visible in this module. Confirm this table exists in the source document or clarify the intent.

customize-models-to-build-gen-ai-applications.adoc (1)

25-26: Clarify data processing assembly placeholder. Lines 25-26 contain a commented-out include for a data processing assembly. Confirm whether this is intentionally deferred or if a reference should be added now. The comment should be removed in the final version if no assembly is planned for this location.

modules/compare-the-performance-of-osft-and-sft.adoc (1)

20-20: Clarify the reference to the examples/docs directory.

The reference to "The examples/docs directory" is vague. Please provide either the full path or a more specific reference to improve findability. Consider linking to specific documentation modules if these are available.

Copy link

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 0

♻️ Duplicate comments (1)
modules/build-a-custom-container-image.adoc (1)

33-53: Fix incomplete and incorrect placeholders in code example to make it functional.

The container building example contains multiple incomplete placeholders and syntax errors that prevent users from following it successfully:

  1. Line 37: podman pull <link to registry.redhat> is a non-functional placeholder. Must specify the actual registry URL.
  2. Line 41: FROM <image name> needs a concrete base image reference.
  3. Line 52: podman build <new_image_name> is incomplete—missing the required -t flag for tagging and . for build context.
  4. Line 51 comment: The #Build the custom image comment uses Dockerfile syntax but appears after the Dockerfile block, creating confusion about command flow.

Apply this diff to provide complete, functional example code:

 [source, bash]
 ----
-podman pull <link to registry.redhat>
+podman pull registry.redhat.io/ubi9/python-39:latest
 
 # Log in via `podman login registry.redhat.io`
+
+# Create a Dockerfile
+cat > Dockerfile << 'EOF'
-FROM <image name>
+FROM registry.redhat.io/ubi9/python-39:latest
 
 # Install software
 
 # Install Python, pip, and then the docling library
 RUN pip3 install docling
 
 # Define the default command to run when the container starts
 CMD ["docling", "--help"]
-
-#Build the custom image
-podman build <new_image_name>
+EOF
+
+# Build the custom image
+podman build -t my-custom-docling-image:latest .

This ensures users have:

  • A working pull command with the correct Red Hat registry
  • A complete, valid Dockerfile with proper FROM and RUN instructions
  • A complete podman build command with tag (-t) and build context (.)
🧹 Nitpick comments (1)
modules/compare-the-performance-of-osft-and-sft.adoc (1)

24-28: Slightly awkward phrasing — optional improvement.

Line 24 contains redundant attribution: "URR which is a hyperparameter for OSFT that is a value between 0 and 1." The phrase combines "which is a hyperparameter" with "that is a value," which is slightly repetitive.

Consider simplifying to: "The unfreeze rank ratio (URR), a hyperparameter for OSFT with a value between 0 and 1, represents..."

📜 Review details

Configuration used: Path: .coderabbit.yaml

Review profile: CHILL

Plan: Pro

📥 Commits

Reviewing files that changed from the base of the PR and between 6c501da and d5df8d5.

📒 Files selected for processing (17)
  • assemblies/generate-synthetic-data-to-augment-real-data.adoc (1 hunks)
  • assemblies/prepare-your-data-for-ai-consumption.adoc (1 hunks)
  • assemblies/set-up-your-working-environment.adoc (1 hunks)
  • assemblies/train-the-model-by-using-your-prepared-data.adoc (1 hunks)
  • customize-models-to-build-gen-ai-applications.adoc (1 hunks)
  • modules/about-the-python-index.adoc (1 hunks)
  • modules/build-a-custom-container-image.adoc (1 hunks)
  • modules/clone-an-example-git-repository.adoc (1 hunks)
  • modules/compare-the-performance-of-osft-and-sft.adoc (1 hunks)
  • modules/end-to-end-model-customization-workflow.adoc (1 hunks)
  • modules/estimate-memory-usage.adoc (1 hunks)
  • modules/explore-the-sdg-hub-examples.adoc (1 hunks)
  • modules/explore-the-training-hub-examples.adoc (1 hunks)
  • modules/import-example-notebooks.adoc (1 hunks)
  • modules/mirror-the-python-index.adoc (1 hunks)
  • modules/overview-of-the-model-customization-workflow.adoc (1 hunks)
  • modules/support-philosophy.adoc (1 hunks)
🚧 Files skipped from review as they are similar to previous changes (12)
  • modules/overview-of-the-model-customization-workflow.adoc
  • modules/estimate-memory-usage.adoc
  • assemblies/prepare-your-data-for-ai-consumption.adoc
  • customize-models-to-build-gen-ai-applications.adoc
  • modules/explore-the-sdg-hub-examples.adoc
  • modules/clone-an-example-git-repository.adoc
  • modules/mirror-the-python-index.adoc
  • modules/support-philosophy.adoc
  • modules/import-example-notebooks.adoc
  • assemblies/set-up-your-working-environment.adoc
  • modules/about-the-python-index.adoc
  • assemblies/generate-synthetic-data-to-augment-real-data.adoc
🔇 Additional comments (7)
modules/build-a-custom-container-image.adoc (1)

1-31: LGTM—Module structure and basic examples are clear and well-formatted.

The metadata, introduction, and simple pip install examples follow AsciiDoc conventions and provide helpful context on pre-configured Python index and system trust store usage. The three concrete library examples (docling, sdg-hub, training-hub) are clear and actionable.

modules/end-to-end-model-customization-workflow.adoc (1)

1-8: LGTM!

The module is well-structured with clear reference to the Knowledge Tuning example. The external links are properly formatted with HTTPS.

assemblies/train-the-model-by-using-your-prepared-data.adoc (2)

20-24: Previous issue resolved.

The duplicate word "the the" mentioned in past review comments has been corrected. Line 24 now correctly reads with a single "the" as expected.


14-18: Includes are properly structured.

The module includes at lines 14, 16, and 18 use consistent leveloffset syntax and reference valid module paths.

modules/explore-the-training-hub-examples.adoc (2)

16-16: Previous xref issue resolved.

The broken cross-reference that used section number "2.4.1" has been corrected to use the proper ID format xref:clone-an-example-git-repository[...].


32-37: Previous security best practice applied.

The insecure HTTP link to red.ht has been updated to use HTTPS (line 36), aligning with documentation security best practices.

modules/compare-the-performance-of-osft-and-sft.adoc (1)

1-32: Overall structure and technical content look good.

The module clearly explains both algorithms and provides a meaningful performance comparison. The mathematical relationships (memory scaling, training time trade-offs) and continual learning benefits are well-articulated for the documentation audience.


* Red Hat AI documentation:

** link:https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.25/html/working_with_distributed_workloads/running-kfto-based-distributed-training-workloads_distributed-workloads[Chapter 4. Running Training Operator-based distributed training workloads] in the *Working with distributed workloads* guide
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
** link:https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.25/html/working_with_distributed_workloads/running-kfto-based-distributed-training-workloads_distributed-workloads[Chapter 4. Running Training Operator-based distributed training workloads] in the *Working with distributed workloads* guide
** link:{rhoaidocshome}{default-format-url}/working_with_distributed_workloads/running-kfto-based-distributed-training-workloads_distributed-workloads[Chapter 4. Running Training Operator-based distributed training workloads] in the *Working with distributed workloads* guide


* Example:

** link:https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.25/html/openshift_ai_tutorial_-_fraud_detection_example/running-a-distributed-workload#distributing-training-jobs-with-kfto[Distributing training jobs with the Training Operator] in the *Red Hat OpenShift AI tutorial: Fraud Detection example*
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
** link:https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.25/html/openshift_ai_tutorial_-_fraud_detection_example/running-a-distributed-workload#distributing-training-jobs-with-kfto[Distributing training jobs with the Training Operator] in the *Red Hat OpenShift AI tutorial: Fraud Detection example*
** link:{rhoaidocshome}{default-format-url}/openshift_ai_tutorial_-_fraud_detection_example/running-a-distributed-workload#distributing-training-jobs-with-kfto[Distributing training jobs with the Training Operator] in the *Red Hat OpenShift AI tutorial: Fraud Detection example*

Copy link
Contributor

@smccarthy-ie smccarthy-ie Nov 9, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

What about adding conditions and links to equivalent upstream docs?

[id='about-the-python-index_{context}']
= About the {org-name} Python Index

{org-name} AI includes a maintained Python package index that provides secure and reliable access to supported libraries, with full support for disconnected environments. For details about {org-name} support for the Python package index, see Support philosophy: A secure platform.
Copy link
Contributor

@smccarthy-ie smccarthy-ie Nov 9, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Add a link to "Support philosophy: A secure platform"?


*Additional resources*

* link:https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.25/html/managing_openshift_ai/creating-custom-workbench-images[Creating custom workbench images] No newline at end of file
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
* link:https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.25/html/managing_openshift_ai/creating-custom-workbench-images[Creating custom workbench images]
* link:{rhoaidocshome}{default-format-url}/managing_openshift_ai/creating-custom-workbench-images[Creating custom workbench images]

Copy link
Contributor

@smccarthy-ie smccarthy-ie Nov 9, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

What about adding conditions and links for equivalent upstream docs?

+
The file-browser window shows the files and directories that are saved inside your own personal space in {productname-short} .

. Bring the content of an example Git repo inside your JupyterLab environment:
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
. Bring the content of an example Git repo inside your JupyterLab environment:
. Bring the content of an example Git repository inside your JupyterLab environment:

+
In most post-training setups, URR values below 1/3 are sufficient for learning new tasks, making OSFT notably lighter in memory.

* *Training time:* On datasets of equal size, OSFT typically takes about 2x longer per phase. However, since OSFT does not require replay buffers from past tasks (unlike SFT), the total training time across multiple phases or tasks is lower with clear benefits as the number of tasks grows.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
* *Training time:* On datasets of equal size, OSFT typically takes about 2x longer per phase. However, since OSFT does not require replay buffers from past tasks (unlike SFT), the total training time across multiple phases or tasks is lower with clear benefits as the number of tasks grows.
* *Training time:* On datasets of equal size, OSFT typically takes about 2x longer per phase. However, because OSFT does not require replay buffers from past tasks (unlike SFT), the total training time across multiple phases or tasks is lower with clear benefits as the number of tasks grows.


*Prerequisites*

* Install the Synthetic Data Generation (SDG) Hub library, as described in xref:set-up-your-working-environment[Set up your working environment].
Copy link
Contributor

@smccarthy-ie smccarthy-ie Nov 9, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Use correct docs link format


. To access the SDG Hub examples, clone the `sdg_hub` Git repository:
+
* To clone the repository from JupyterLab, follow the steps in xref:clone-an-example-git-repository[Clone an example Git repository].
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Use correct docs link format


*Additional resources*

* Upstream documentation: link:https://github.com/instructlab/sdg/tree/main/docs[https://github.com/instructlab/sdg/tree/main/docs]
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
* Upstream documentation: link:https://github.com/instructlab/sdg/tree/main/docs[https://github.com/instructlab/sdg/tree/main/docs]
* SDG community documentation: link:https://github.com/instructlab/sdg/tree/main/docs[https://github.com/instructlab/sdg/tree/main/docs]

*Additional resources*

* Upstream documentation: link:https://github.com/instructlab/sdg/tree/main/docs[https://github.com/instructlab/sdg/tree/main/docs]
* GitHub repository: link:https://github.com/instructlab/sdg[https://github.com/instructlab/sdg] No newline at end of file
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
* GitHub repository: link:https://github.com/instructlab/sdg[https://github.com/instructlab/sdg]
* SDG GitHub repository: link:https://github.com/instructlab/sdg[https://github.com/instructlab/sdg]


*Prerequisites*

* Install the Training Hub library, as described in xref:set-up-your-working-environment[Set up your working environment].
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Use correct docs link format

+
You can extend a base notebook to use distributed training across multiple nodes by using the KubeFlow Trainer Operator (KFTO). The KFTO, abstracts the underlying infrastructure complexity of distributed training and fine-tuning of models. The iterative process of fine-tuning significantly reduces the time and resources required compared to training models from scratch.
+
For details, see xref:train-the-model-by-using-your-prepared-data[Train the model by using your prepared data].
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Use correct link format


Serve and consume a customized model:: After you customize a model, you can serve your customized models as APIs (Application Programming Interfaces). Serving a model as an API enables seamless integration into existing or newly developed applications.
+
Learn more about serving and consuming a customized model link:https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.24/html/deploying_models/deploying_models_on_the_single_model_serving_platform[Chapter 2: Deploying a model on the Single Model Serving platform] in the Deploying models guide. No newline at end of file
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
Learn more about serving and consuming a customized model link:https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.24/html/deploying_models/deploying_models_on_the_single_model_serving_platform[Chapter 2: Deploying a model on the Single Model Serving platform] in the Deploying models guide.
Learn more about serving and consuming a customized model link:{rhoaidocshome}{default-format-url}/deploying_models/deploying_models_on_the_single_model_serving_platform[Chapter 2: Deploying a model on the Single Model Serving platform] in the Deploying models guide.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

What about adding links to equivalent upstream docs?

Prepare your data for AI consumption:: To prepare your data, use Docling, a powerful Python library to transform unstructured data (such as text documents, images, and audio files) into structured formats that models can consume.
//For details, see xref:prepare-your-data-for-ai-consumption[Prepare your data for AI consumption].
+
To automate data processing tasks, you can build Kubeflow Pipelines (KFP), see Automate data processing steps by building AI pipelines
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
To automate data processing tasks, you can build Kubeflow Pipelines (KFP), see Automate data processing steps by building AI pipelines
To automate data processing tasks, you can build Kubeflow Pipelines (KFP), see Automate data processing steps by building AI pipelines.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Add link to Automate data processing steps by building AI pipelines?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants