@@ -57,7 +57,8 @@ def test_rag_basic_inference(self, lls_client: LlamaStackClient) -> None:
5757 Validates that the server can perform text generation using the chat completions API
5858 and provides factually correct responses.
5959
60- Based on the example available at https://llama-stack.readthedocs.io/en/latest/getting_started/detailed_tutorial.html#step-4-run-the-demos
60+ Based on the example available at
61+ https://llama-stack.readthedocs.io/en/latest/getting_started/detailed_tutorial.html#step-4-run-the-demos
6162 """
6263 models = lls_client .models .list ()
6364 model_id = next (m for m in models if m .api_model_type == "llm" ).identifier
@@ -84,7 +85,8 @@ def test_rag_simple_agent(self, lls_client: LlamaStackClient) -> None:
8485 Validates agent creation, session management, and turn-based interactions
8586 with both identity and capability questions.
8687
87- Based on the example available at https://llama-stack.readthedocs.io/en/latest/getting_started/detailed_tutorial.html#step-4-run-the-demos
88+ Based on the example available at
89+ https://llama-stack.readthedocs.io/en/latest/getting_started/detailed_tutorial.html#step-4-run-the-demos
8890 """
8991 models = lls_client .models .list ()
9092 model_id = next (m for m in models if m .api_model_type == "llm" ).identifier
@@ -120,7 +122,8 @@ def test_rag_build_rag_agent(self, lls_client: LlamaStackClient) -> None:
120122 about fine-tuning techniques (LoRA, QAT, memory optimizations), and validates
121123 that responses contain expected technical keywords.
122124
123- Based on the example available at https://llama-stack.readthedocs.io/en/latest/getting_started/detailed_tutorial.html#step-4-run-the-demos
125+ Based on the example available at
126+ https://llama-stack.readthedocs.io/en/latest/getting_started/detailed_tutorial.html#step-4-run-the-demos
124127 """
125128 models = lls_client .models .list ()
126129 model_id = next (m for m in models if m .api_model_type == "llm" ).identifier
@@ -164,7 +167,7 @@ def test_rag_build_rag_agent(self, lls_client: LlamaStackClient) -> None:
164167 documents = [
165168 RAGDocument (
166169 document_id = f"num-{ i } " ,
167- content = f"https://raw.githubusercontent.com/pytorch/torchtune/refs/tags/v0.6.1/docs/source/tutorials/{ url } " ,
170+ content = f"https://raw.githubusercontent.com/pytorch/torchtune/refs/tags/v0.6.1/docs/source/tutorials/{ url } " , # noqa
168171 mime_type = "text/plain" ,
169172 metadata = {},
170173 )
0 commit comments