Temperature and seed LLM
This commit is contained in:
@@ -9,6 +9,7 @@ class OllamaClient():
|
||||
def __init__(self):
|
||||
self.host = os.getenv("ENDPOINT_OLLAMA", "https://ollamamodel.matitos.org")
|
||||
self.client = ollama.Client(host=self.host)
|
||||
self.options = {"temperature": 0, "seed": 13579}
|
||||
|
||||
def _get_default_model(self):
|
||||
return os.getenv("OLLAMA_MODEL_DEFAULT", "llama3.2:3b")
|
||||
@@ -44,7 +45,7 @@ class OllamaClient():
|
||||
def generate(self, model, prompt, format=None):
|
||||
try:
|
||||
# Generate response
|
||||
response = self.client.generate(model=model, prompt=prompt, format=format)
|
||||
response = self.client.generate(model=model, prompt=prompt, format=format, options=self.options)
|
||||
# Extract response
|
||||
response = response.response
|
||||
# Json? -> Dict
|
||||
@@ -65,7 +66,7 @@ class OllamaClient():
|
||||
def generate_stream(self, model, prompt):
|
||||
try:
|
||||
# Generate response
|
||||
response = self.client.generate(model=model, prompt=prompt, format="json", stream=True)
|
||||
response = self.client.generate(model=model, prompt=prompt, format="json", stream=True, options=self.options)
|
||||
# Streamed chunks
|
||||
for chunk in response:
|
||||
yield chunk.response
|
||||
|
||||
Reference in New Issue
Block a user