Temperature and seed LLM

This commit is contained in:
Luciano Gervasoni
2025-04-23 17:46:47 +02:00
parent cf55c586f7
commit b8fdcae5ec

View File

@@ -9,6 +9,7 @@ class OllamaClient():
def __init__(self):
self.host = os.getenv("ENDPOINT_OLLAMA", "https://ollamamodel.matitos.org")
self.client = ollama.Client(host=self.host)
self.options = {"temperature": 0, "seed": 13579}
def _get_default_model(self):
return os.getenv("OLLAMA_MODEL_DEFAULT", "llama3.2:3b")
@@ -44,7 +45,7 @@ class OllamaClient():
def generate(self, model, prompt, format=None):
try:
# Generate response
response = self.client.generate(model=model, prompt=prompt, format=format)
response = self.client.generate(model=model, prompt=prompt, format=format, options=self.options)
# Extract response
response = response.response
# Json? -> Dict
@@ -65,7 +66,7 @@ class OllamaClient():
def generate_stream(self, model, prompt):
try:
# Generate response
response = self.client.generate(model=model, prompt=prompt, format="json", stream=True)
response = self.client.generate(model=model, prompt=prompt, format="json", stream=True, options=self.options)
# Streamed chunks
for chunk in response:
yield chunk.response