Skip to content

Commit 75e45de

Browse files
Merge pull request #6 from IBM/feature/mcp-support
Feature/mcp support
2 parents cf11f74 + 2d231b1 commit 75e45de

18 files changed

+1117
-278
lines changed

.gitignore

+2-1
Original file line numberDiff line numberDiff line change
@@ -5,5 +5,6 @@
55
*.pyc
66
*__pycache__
77
sandbox.ipynb
8+
.webui_secret_key
89
venv/
9-
.venv/
10+
.venv/

CHANGELOG.md

+27-2
Original file line numberDiff line numberDiff line change
@@ -4,6 +4,30 @@ All notable changes to Flexo will be documented in this file.
44
The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/),
55
and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html).
66

7+
## [v0.2.2] - 2025-03-25
8+
9+
### Features & Improvements
10+
- Added `/models` endpoint and enabled CORS support.
11+
- Initiated MCP client and tool registry functionality (WIP).
12+
- Updated tool patterns for enhanced consistency.
13+
14+
### Fixes & Updates
15+
- Fixed multi tool accumulation issue.
16+
- Fixed mid-response tool buffer leak.
17+
- Updated chat completions data models: updated `FunctionDetail` and removed `name` from `ToolCall`.
18+
19+
### Notes
20+
- MCP configuration in `agent.yaml` is now commented out by default.
21+
22+
[v0.2.2]: https://github.com/ibm/flexo/releases/tag/v0.2.2
23+
24+
## [v0.2.1] - 2025-03-14
25+
26+
### Fixes & Updates
27+
- Fixed issue with LLMFactory not recognizing openai-compat vendor names
28+
29+
[v0.2.1]: https://github.com/ibm/flexo/releases/tag/v0.2.1
30+
731
## [v0.2.0] - 2025-03-10
832

933
### Features & Improvements
@@ -38,6 +62,8 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
3862
- Updated tool configuration and path structures
3963
- Added Elasticsearch SSL certificate documentation
4064

65+
[v0.1.1]: https://github.com/ibm/flexo/releases/tag/v0.1.1
66+
4167
## [v0.1.0] - 2024-01-31
4268

4369
### Initial Release
@@ -52,5 +78,4 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
5278
- Robust prompt building and parsing systems
5379
- Comprehensive LLM integration components
5480

55-
[v0.1.1]: https://github.com/ibm/flexo/releases/tag/v0.1.1
56-
[v0.1.0]: https://github.com/ibm/flexo/releases/tag/v0.1.0
81+
[v0.1.0]: https://github.com/ibm/flexo/releases/tag/v0.1.0

src/agent/chat_agent_streaming.py

+15-12
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
# src/agent/chat_streaming_agent.py
22

33
import os
4-
import json
4+
import json5
55
import yaml
66
import asyncio
77
import logging
@@ -99,7 +99,7 @@ def __init__(self, config: Dict) -> None:
9999

100100
# Load parser config if needed for manual detection
101101
self.logger.info("\n\n" + "=" * 60 + "\n" + "Agent Configuration Summary" + "\n" + "=" * 60 + "\n" +
102-
f"Main Chat Model Config:\n{json.dumps(self.main_chat_model_config, indent=4)}\n" +
102+
f"Main Chat Model Config:\n{json5.dumps(self.main_chat_model_config, indent=4)}\n" +
103103
f"Tool Detection Mode: {self.detection_mode}\n" +
104104
f"Vendor Chat API Mode: {self.use_vendor_chat_completions}\n" +
105105
"\n" + "=" * 60 + "\n")
@@ -122,9 +122,12 @@ def __init__(self, config: Dict) -> None:
122122
vendor=self.main_chat_model_config.get('vendor')
123123
)
124124

125-
# Initialize ToolRegistry
126-
self.tool_registry = ToolRegistry()
127-
self.tool_registry.load_from_config(tool_configs=self.config.get("tools_config"))
125+
# Initialize tool registry
126+
self.tool_registry = ToolRegistry(
127+
tools_config=self.config.get("tools_config"),
128+
mcp_config=self.config.get("mcp_config")
129+
)
130+
asyncio.create_task(self.tool_registry.initialize_all_tools())
128131

129132
@handle_streaming_errors
130133
async def stream_step(
@@ -160,7 +163,6 @@ async def stream_step(
160163

161164
case StreamState.STREAMING:
162165
self.logger.info(f"--- Entering Streaming State ---")
163-
self.detection_strategy.reset()
164166
async for item in self._handle_streaming(context):
165167
yield item
166168

@@ -205,6 +207,7 @@ async def _handle_streaming(self, context: StreamContext) -> AsyncGenerator[SSEC
205207
Raises:
206208
Exception: If maximum streaming iterations are exceeded
207209
"""
210+
self.detection_strategy.reset()
208211
context.streaming_entry_count += 1
209212
if context.streaming_entry_count > context.max_streaming_iterations:
210213
self.logger.error("Maximum streaming iterations reached. Aborting further streaming.")
@@ -218,7 +221,6 @@ async def _handle_streaming(self, context: StreamContext) -> AsyncGenerator[SSEC
218221
conversation_history=context.conversation_history,
219222
tool_definitions=context.tool_definitions if self.detection_mode == "manual" else None
220223
)
221-
self.logger.debug(f"Prompt payload: {prompt_payload}")
222224

223225
prompt_output: PromptBuilderOutput = (
224226
await self.prompt_builder.build_chat(prompt_payload) if self.use_vendor_chat_completions
@@ -253,6 +255,7 @@ async def _handle_streaming(self, context: StreamContext) -> AsyncGenerator[SSEC
253255
return
254256

255257
final_result = await self.detection_strategy.finalize_detection(context)
258+
self.logger.debug(f"Final detection result: {final_result}")
256259

257260
if final_result.state == DetectionState.COMPLETE_MATCH:
258261
async for chunk in self._handle_complete_match(context, final_result, accumulated_content):
@@ -328,7 +331,6 @@ async def _handle_tool_execution(
328331
tool_results.append(result)
329332
context.conversation_history.append(
330333
ToolMessage(
331-
name=call.function.name,
332334
content=result["result"],
333335
tool_call_id=call.id
334336
)
@@ -405,7 +407,7 @@ async def _initialize_context(
405407

406408
return StreamContext(
407409
conversation_history=selected_history,
408-
tool_definitions=self.tool_registry.get_tool_definitions(),
410+
tool_definitions=await self.tool_registry.get_tool_definitions(),
409411
context=api_passed_context,
410412
llm_factory=self.llm_factory,
411413
current_state=StreamState.STREAMING,
@@ -458,13 +460,14 @@ async def _execute_tools_concurrently(self, context: StreamContext) -> List[Any]
458460
"""
459461
async def run_tool(tool_call: ToolCall):
460462
try:
461-
tool = self.tool_registry.get_tool(tool_call.function.name)
463+
tool = await self.tool_registry.get_tool(tool_call.function.name)
462464
if not tool:
463465
raise RuntimeError(f"Tool {tool_call.function.name} not found")
464-
466+
tool_args = json5.loads(tool_call.function.arguments)
467+
self.logger.info(f"Running tool {tool_call.function.name} with arguments: {tool_args}")
465468
result = await tool.execute(
466469
context=context,
467-
**tool_call.function.arguments
470+
**tool_args
468471
)
469472
return {"tool_name": tool_call.function.name, "result": result.result}
470473

src/api/routes/chat_completions_api.py

+55-7
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
# src/api/routes/chat_completions_api.py
2-
31
"""Chat Completions API Module.
42
53
This module implements a FastAPI router for streaming chat completion functionality,
@@ -10,14 +8,11 @@
108
- Message format conversion
119
- Streaming chat completions
1210
- SSE (Server-Sent Events) response handling
13-
14-
Dependencies:
15-
- FastAPI for route handling
16-
- StreamingChatAgent for chat processing
17-
- Pydantic models for request/response validation
11+
- Agent info endpoint (compatible with OpenAI's /models format)
1812
"""
1913

2014
import os
15+
import time
2116
import yaml
2217
import logging
2318
from typing import List, Optional
@@ -37,6 +32,9 @@
3732
logger = logging.getLogger(__name__)
3833
router = APIRouter()
3934

35+
# Capture the agent's start time once when the app starts.
36+
AGENT_START_TIME = int(time.time())
37+
4038
# Security setup
4139
API_KEY_NAME = "X-API-KEY"
4240
api_key_header = APIKeyHeader(name=API_KEY_NAME, auto_error=False)
@@ -118,6 +116,22 @@ def convert_message_content(messages: List[TextChatMessage]) -> List[TextChatMes
118116
return converted
119117

120118

119+
def get_non_sensitive_config(config: dict) -> dict:
120+
"""Filter the agent configuration to include only non-sensitive parameters.
121+
122+
Args:
123+
config (dict): The full agent configuration.
124+
125+
Returns:
126+
dict: Filtered configuration excluding keys that may be sensitive.
127+
"""
128+
# Exclude keys that include 'key', 'secret', or 'password' (case insensitive)
129+
return {
130+
k: v for k, v in config.items()
131+
if not any(sensitive in k.lower() for sensitive in ["key", "secret", "password"])
132+
}
133+
134+
121135
@router.post(
122136
"/chat/completions",
123137
summary="Generate streaming chat completions",
@@ -193,3 +207,37 @@ async def sse_generator():
193207
except Exception as e:
194208
logger.error("Error in /chat/completions: %s", str(e), exc_info=True)
195209
raise HTTPException(status_code=500, detail=str(e))
210+
211+
212+
@router.get(
213+
"/models",
214+
summary="Agent information",
215+
description="Returns information about the running agent including its default model and non-sensitive configuration parameters.",
216+
tags=["Models"],
217+
operation_id="getModels"
218+
)
219+
async def get_models(
220+
agent: StreamingChatAgent = Depends(get_streaming_agent),
221+
api_key: Optional[str] = Depends(get_api_key)
222+
):
223+
"""Retrieve agent information in a format similar to OpenAI's models endpoint.
224+
225+
This endpoint returns the agent's name, default model, creation time,
226+
owner, and a selection of non-sensitive configuration parameters.
227+
"""
228+
agent_name = agent.config.get("name", "default-agent")
229+
default_model = agent.config.get("model", "default-model")
230+
created = agent.config.get("created", AGENT_START_TIME)
231+
owned_by = agent.config.get("owned_by", "user")
232+
non_sensitive_params = get_non_sensitive_config(agent.config)
233+
234+
agent_info = {
235+
"id": agent_name,
236+
"object": "agent",
237+
"default_model": default_model,
238+
"created": created,
239+
"owned_by": owned_by,
240+
"params": non_sensitive_params
241+
}
242+
243+
return {"data": [agent_info], "object": "list"}

src/configs/agent.yaml

+18
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,10 @@ max_streaming_iterations: 4 # Number of times the agent is allowed to return to
1010
timeouts:
1111
model_response_timeout: 60
1212

13+
# CORS allowed origins (optional)
14+
allowed_origins:
15+
- http://localhost:8080 # example for local Open WebUI
16+
1317
# System prompt defining agent behavior and capabilities
1418
system_prompt: |
1519
I am a helpful AI assistant focused on clear, accurate, and direct communication.
@@ -86,6 +90,20 @@ models_config:
8690
# max_tokens: 2000
8791
# temperature: 0.7
8892

93+
# MCP configuration
94+
#mcp_config:
95+
# # Example SSE
96+
# transport: sse
97+
# sse_url: "http://localhost:8001/sse"
98+
# sampling_enabled: false
99+
100+
# Example stdio
101+
# transport: stdio
102+
# command: python
103+
# args: ["weather/weather.py"]
104+
# env: null
105+
# sampling_enabled: false
106+
89107
# Tool Configurations
90108
tools_config:
91109
# Weather API Integration

src/configs/prompt_builders.yaml

+26-8
Original file line numberDiff line numberDiff line change
@@ -5,17 +5,25 @@ watsonx-llama:
55
begin_text: "<|begin_of_text|>"
66
system_prompt:
77
header: "You have access to the following tools: {tools}. Current date: {date}"
8-
tool_instructions: 'Each tool call must contain "name" and "parameters". To use a tool, emit a call in the following JSON format:\n<|tool_call|>[{\"name\": \"tool_name\", \"parameters\": {\"arg1\": \"value1\"}}]\n\n### Example:\n<|tool_call|>[{\"name\": \"tool_name_1\", \"parameters\": {\"param_1\": \"value_1\", \"param_2\": \"value_2\"}}]\n\nEnsure that all tool calls strictly follow this JSON format.'
8+
tool_instructions: 'Each tool call must contain "name" and "parameters". To use a tool, emit a call in the following JSON format:
9+
<|tool_call|>[{"name": "tool_name", "parameters": {"arg1": "value1"}}]
10+
11+
### Example:
12+
<|tool_call|>[{"name": "tool_name_1", "parameters": {"param_1": "value_1", "param_2": "value_2"}}]
13+
14+
Ensure that all tool calls strictly follow this JSON format.'
915

1016
watsonx-granite:
1117
system_prompt:
1218
header: "You have access to the following tools: {tools}. Current date: {date}"
13-
tool_instructions: 'Each tool call must contain "name" and "arguments". To use a tool, emit a function call in the following JSON format:\n<|tool_call|>[{\"name\": \"tool_name\", \"arguments\": {\"arg1\": \"value1\"}}]'
19+
tool_instructions: 'Each tool call must contain "name" and "arguments". To use a tool, emit a function call in the following JSON format:
20+
<|tool_call|>[{"name": "tool_name", "arguments": {"arg1": "value1"}}]'
1421

1522
watsonx-mistral:
1623
system_prompt:
1724
header: "You have access to the following tools: {tools}. Current date: {date}"
18-
tool_instructions: 'Each tool call must contain "name" and "arguments". To use a tool, emit a function call in the following JSON format:\n<|tool_call|>[{\"name\": \"tool_name\", \"arguments\": {\"arg1\": \"value1\"}}]'
25+
tool_instructions: 'Each tool call must contain "name" and "arguments". To use a tool, emit a function call in the following JSON format:
26+
<|tool_call|>[{"name": "tool_name", "arguments": {"arg1": "value1"}}]'
1927

2028
openai:
2129
system_prompt:
@@ -25,26 +33,36 @@ openai:
2533
anthropic:
2634
system_prompt:
2735
header: "You can use the following tools: {tools}. Current date: {date}"
28-
tool_instructions: 'Each tool call must contain "name" and "arguments". To use a tool, emit a function call in the following JSON format:\n<|tool_call|>[{\"name\": \"tool_name\", \"arguments\": {\"arg1\": \"value1\"}}]'
36+
tool_instructions: 'Each tool call must contain "name" and "arguments". To use a tool, emit a function call in the following JSON format:
37+
<|tool_call|>[{"name": "tool_name", "arguments": {"arg1": "value1"}}]'
2938

3039
mistralai:
3140
system_prompt:
3241
header: "You can use the following tools: {tools}. Current date: {date}"
33-
tool_instructions: 'Each tool call must contain "name" and "arguments". To use a tool, emit a function call in the following JSON format:\n<|tool_call|>[{\"name\": \"tool_name\", \"arguments\": {\"arg1\": \"value1\"}}]'
42+
tool_instructions: 'Each tool call must contain "name" and "arguments". To use a tool, emit a function call in the following JSON format:
43+
<|tool_call|>[{"name": "tool_name", "arguments": {"arg1": "value1"}}]'
3444

3545
xai:
3646
system_prompt:
3747
header: "You can use the following tools: {tools}. Current date: {date}"
38-
tool_instructions: 'Each tool call must contain "name" and "arguments". To use a tool, emit a function call in the following JSON format:\n<|tool_call|>[{\"name\": \"tool_name\", \"arguments\": {\"arg1\": \"value1\"}}]'
48+
tool_instructions: 'Each tool call must contain "name" and "arguments". To use a tool, emit a function call in the following JSON format:
49+
<|tool_call|>[{"name": "tool_name", "arguments": {"arg1": "value1"}}]'
3950

4051
openai-compat-llama:
4152
tokens:
4253
begin_text: "<|begin_of_text|>"
4354
system_prompt:
4455
header: "You have access to the following tools: {tools}. Current date: {date}"
45-
tool_instructions: 'Each tool call must contain "name" and "parameters". To use a tool, emit a call in the following JSON format:\n<|tool_call|>[{\"name\": \"tool_name\", \"parameters\": {\"arg1\": \"value1\"}}]\n\n### Example:\n<|tool_call|>[{\"name\": \"tool_name_1\", \"parameters\": {\"param_1\": \"value_1\", \"param_2\": \"value_2\"}}]\n\nEnsure that all tool calls strictly follow this JSON format.'
56+
tool_instructions: 'Each tool call must contain "name" and "parameters". To use a tool, emit a call in the following JSON format:
57+
<|tool_call|>[{"name": "tool_name", "parameters": {"arg1": "value1"}}]
58+
59+
### Example:
60+
<|tool_call|>[{"name": "tool_name_1", "parameters": {"param_1": "value_1", "param_2": "value_2"}}]
61+
62+
Ensure that all tool calls strictly follow this JSON format.'
4663

4764
openai-compat-granite:
4865
system_prompt:
4966
header: "You have access to the following tools: {tools}. Current date: {date}"
50-
tool_instructions: 'Each tool call must contain "name" and "arguments". To use a tool, emit a function call in the following JSON format:\n<|tool_call|>[{\"name\": \"tool_name\", \"arguments\": {\"arg1\": \"value1\"}}]'
67+
tool_instructions: 'Each tool call must contain "name" and "arguments". To use a tool, emit a function call in the following JSON format:
68+
<|tool_call|>[{"name": "tool_name", "arguments": {"arg1": "value1"}}]'

0 commit comments

Comments
 (0)