From 0ac6b292cff32a0d0f1e6136aa9faf0fbe129c27 Mon Sep 17 00:00:00 2001 From: Michael Abashian Date: Tue, 11 Feb 2025 16:11:27 -0500 Subject: [PATCH] Fix more tests --- .../ai/api/model_pipelines/langchain/pipelines.py | 3 +++ .../ai/api/model_pipelines/ollama/pipelines.py | 14 ++++++++++++++ 2 files changed, 17 insertions(+) diff --git a/ansible_ai_connect/ai/api/model_pipelines/langchain/pipelines.py b/ansible_ai_connect/ai/api/model_pipelines/langchain/pipelines.py index 27588c9b0..c2f206923 100644 --- a/ansible_ai_connect/ai/api/model_pipelines/langchain/pipelines.py +++ b/ansible_ai_connect/ai/api/model_pipelines/langchain/pipelines.py @@ -459,5 +459,8 @@ class LangchainRoleExplanationPipeline( def __init__(self, config: LANGCHAIN_PIPELINE_CONFIGURATION): super().__init__(config=config) + def self_test(self) -> Optional[HealthCheckSummary]: + raise NotImplementedError + def invoke(self, params: RoleExplanationParameters) -> RoleExplanationResponse: raise NotImplementedError diff --git a/ansible_ai_connect/ai/api/model_pipelines/ollama/pipelines.py b/ansible_ai_connect/ai/api/model_pipelines/ollama/pipelines.py index eab3de1a0..3ff21543f 100644 --- a/ansible_ai_connect/ai/api/model_pipelines/ollama/pipelines.py +++ b/ansible_ai_connect/ai/api/model_pipelines/ollama/pipelines.py @@ -21,6 +21,7 @@ LangchainMetaData, LangchainPlaybookExplanationPipeline, LangchainPlaybookGenerationPipeline, + LangchainRoleExplanationPipeline, LangchainRoleGenerationPipeline, ) from ansible_ai_connect.ai.api.model_pipelines.ollama.configuration import ( @@ -80,6 +81,19 @@ def get_chat_model(self, model_id): ) +@Register(api_type="ollama") +class OllamaRoleExplanationPipeline(LangchainRoleExplanationPipeline[OllamaConfiguration]): + + def __init__(self, config: OllamaConfiguration): + super().__init__(config=config) + + def get_chat_model(self, model_id): + return OllamaLLM( + base_url=self.config.inference_url, + model=model_id, + ) + + @Register(api_type="ollama") class OllamaPlaybookExplanationPipeline(LangchainPlaybookExplanationPipeline[OllamaConfiguration]):