@@ -13,13 +13,14 @@ class LLMBackend(ABC):
1313 def __init__ (self , config : dict [str , Any ], seed : int | None = None ) -> None :
1414 self .config = config
1515 self .seed = seed
16+ # Set temperature based on seed - 0.0 for reproducibility, 0.7 otherwise
17+ self .temperature = 0.0 if seed is not None else 0.7
1618
1719 @abstractmethod
1820 def generate (
1921 self ,
2022 prompt : str ,
2123 system_prompt : str | None = None ,
22- temperature : float = 0.7 ,
2324 max_tokens : int | None = None ,
2425 stream : bool = False ,
2526 ) -> ModelResponse :
@@ -30,7 +31,6 @@ def generate(
3031 def chat (
3132 self ,
3233 messages : list [dict [str , str ]],
33- temperature : float = 0.7 ,
3434 max_tokens : int | None = None ,
3535 ) -> ModelResponse :
3636 """Multi-turn chat conversation."""
@@ -79,37 +79,27 @@ def generate(
7979 self ,
8080 prompt : str ,
8181 system_prompt : str | None = None ,
82- temperature : float = 0.7 ,
8382 max_tokens : int | None = None ,
8483 stream : bool = False ,
8584 ) -> ModelResponse :
8685 """Generate response from Ollama model."""
87- # For reproducibility, use temperature=0 when seed is set
88- if self .seed is not None :
89- temperature = 0.0
90-
9186 return self .client .generate (
9287 prompt = prompt ,
9388 system_prompt = system_prompt ,
94- temperature = temperature ,
89+ temperature = self . temperature ,
9590 max_tokens = max_tokens ,
9691 stream = stream ,
9792 )
9893
9994 def chat (
10095 self ,
10196 messages : list [dict [str , str ]],
102- temperature : float = 0.7 ,
10397 max_tokens : int | None = None ,
10498 ) -> ModelResponse :
10599 """Multi-turn chat conversation with Ollama."""
106- # For reproducibility, use temperature=0 when seed is set
107- if self .seed is not None :
108- temperature = 0.0
109-
110100 return self .client .chat (
111101 messages = messages ,
112- temperature = temperature ,
102+ temperature = self . temperature ,
113103 max_tokens = max_tokens ,
114104 )
115105
@@ -168,14 +158,10 @@ def generate(
168158 self ,
169159 prompt : str ,
170160 system_prompt : str | None = None ,
171- temperature : float = 0.7 ,
172161 max_tokens : int | None = None ,
173162 stream : bool = False ,
174163 ) -> ModelResponse :
175164 """Generate response from OpenRouter model."""
176- # For reproducibility, use temperature=0 when seed is set
177- if self .seed is not None :
178- temperature = 0.0
179165
180166 start_time = time .time ()
181167
@@ -189,7 +175,7 @@ def generate(
189175 request_params = {
190176 "model" : self .model ,
191177 "messages" : messages ,
192- "temperature" : temperature ,
178+ "temperature" : self . temperature ,
193179 "stream" : stream ,
194180 "timeout" : self .timeout ,
195181 "extra_headers" : self ._get_headers (),
@@ -238,13 +224,9 @@ def generate(
238224 def chat (
239225 self ,
240226 messages : list [dict [str , str ]],
241- temperature : float = 0.7 ,
242227 max_tokens : int | None = None ,
243228 ) -> ModelResponse :
244229 """Multi-turn chat conversation with OpenRouter."""
245- # For reproducibility, use temperature=0 when seed is set
246- if self .seed is not None :
247- temperature = 0.0
248230
249231 start_time = time .time ()
250232
@@ -253,7 +235,7 @@ def chat(
253235 request_params = {
254236 "model" : self .model ,
255237 "messages" : messages ,
256- "temperature" : temperature ,
238+ "temperature" : self . temperature ,
257239 "timeout" : self .timeout ,
258240 "extra_headers" : self ._get_headers (),
259241 }
0 commit comments