diff --git a/src/connection_handler.py b/src/connection_handler.py index 67afe95..dc38497 100644 --- a/src/connection_handler.py +++ b/src/connection_handler.py @@ -23,9 +23,9 @@ def log_output(pipe): window.show_toast(_("Model request too large for system"), window.main_overlay) elif 'msg="amdgpu detected, but no compatible rocm library found.' in line: if bool(os.getenv("FLATPAK_ID")): - window.ollama_information_label.set_label(_("AMD GPU detected but the extension is missing, Ollama will use CPU.\n{}Alpaca support{}").format("", "")) + window.ollama_information_label.set_label(_("AMD GPU detected but the extension is missing, Ollama will use CPU.\n{}").format("{}".format(_('Alpaca Support')))) else: - window.ollama_information_label.set_label(_("AMD GPU detected but ROCm is missing, Ollama will use CPU.\n{}Alpaca support{}").format("", "")) + window.ollama_information_label.set_label(_("AMD GPU detected but ROCm is missing, Ollama will use CPU.\n{}").format("{}".format(_('Alpaca Support')))) window.ollama_information_label.set_css_classes(['dim-label', 'error']) elif 'msg="amdgpu is supported"' in line: window.ollama_information_label.set_label(_("Using AMD GPU type '{}'").format(line.split('=')[-1]))