feat(config): enhance Ollama configuration with dynamic path support
This commit is contained in:
@@ -2,23 +2,26 @@
|
|||||||
"audio_file": "/tmp/pyvtt_recording.wav",
|
"audio_file": "/tmp/pyvtt_recording.wav",
|
||||||
"output_file": "/tmp/pyvtt_transcript.txt",
|
"output_file": "/tmp/pyvtt_transcript.txt",
|
||||||
"whisper_path": "/path/to/whisper-cli",
|
"whisper_path": "/path/to/whisper-cli",
|
||||||
"language": "en",
|
|
||||||
"socket_path": "/tmp/pyvtt.sock",
|
"socket_path": "/tmp/pyvtt.sock",
|
||||||
"ollama_url": "http://localhost",
|
"ollama_url": "http://localhost",
|
||||||
|
"ollama_path": "/api/chat",
|
||||||
"ollama_port": 12345,
|
"ollama_port": 12345,
|
||||||
"presets": [
|
"presets": [
|
||||||
{
|
{
|
||||||
"name": "Default",
|
"name": "Default",
|
||||||
"language": "en",
|
"language": "en",
|
||||||
"whisper_model": "/path/to/default-whisper-model.bin",
|
"whisper_model": "/path/to/default-whisper-model.bin",
|
||||||
"ollama_model": "default-model",
|
"ollama": "disable"
|
||||||
"ollama_prompt": "Provide a detailed response to the following text:\n\n"
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name": "Quick English",
|
"name": "Quick English",
|
||||||
"whisper_model": "/path/to/quick-whisper-model.bin",
|
"whisper_model": "/path/to/quick-whisper-model.bin",
|
||||||
"ollama_model": "quick-model",
|
"ollama_model": "gemma3:4b",
|
||||||
"ollama_prompt": "Quickly correct the following English text for grammar and punctuation:\n\n"
|
"ollama_context": 131072,
|
||||||
|
"ollama_prompt": [
|
||||||
|
"Quickly correct the following English text for grammar and punctuation:\n",
|
||||||
|
"\n"
|
||||||
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name": "German Correction",
|
"name": "German Correction",
|
||||||
|
|||||||
@@ -14,6 +14,7 @@ class OllamaClient:
|
|||||||
"""
|
"""
|
||||||
self.base_url = config.ollama_url
|
self.base_url = config.ollama_url
|
||||||
self.port = config.ollama_port
|
self.port = config.ollama_port
|
||||||
|
self.path = config.ollama_path
|
||||||
|
|
||||||
def send_chat(
|
def send_chat(
|
||||||
self,
|
self,
|
||||||
@@ -52,7 +53,7 @@ class OllamaClient:
|
|||||||
"stream": False
|
"stream": False
|
||||||
}
|
}
|
||||||
|
|
||||||
endpoint = f"{self.base_url}:{self.port}/api/chat"
|
endpoint = f"{self.base_url}:{self.port}{self.path}"
|
||||||
|
|
||||||
# Anfrage an Ollama senden und Antwort extrahieren
|
# Anfrage an Ollama senden und Antwort extrahieren
|
||||||
try:
|
try:
|
||||||
|
|||||||
@@ -20,6 +20,7 @@ class AppConfig(BaseModel):
|
|||||||
whisper_path: str
|
whisper_path: str
|
||||||
socket_path: str
|
socket_path: str
|
||||||
ollama_url: str
|
ollama_url: str
|
||||||
|
ollama_path: str
|
||||||
ollama_port: int
|
ollama_port: int
|
||||||
journal_path: str
|
journal_path: str
|
||||||
presets: List[PresetConfig]
|
presets: List[PresetConfig]
|
||||||
|
|||||||
Reference in New Issue
Block a user