Fixed reconnection dialog not selecting 'use local instance'
This commit is contained in:
@@ -22,7 +22,10 @@ def log_output(pipe):
|
||||
if 'msg="model request too large for system"' in line:
|
||||
window.show_toast(_("Model request too large for system"), window.main_overlay)
|
||||
elif 'msg="amdgpu detected, but no compatible rocm library found.' in line:
|
||||
window.ollama_information_label.set_label(_("AMD GPU detected but the extension is missing, Ollama will use CPU"))
|
||||
if bool(os.getenv("FLATPAK_ID")):
|
||||
window.ollama_information_label.set_label(_("AMD GPU detected but <a href='appstream://com.jeffser.Alpaca.Plugins.AMD'>the extension</a> is missing, Ollama will use CPU"))
|
||||
else:
|
||||
window.ollama_information_label.set_label(_("AMD GPU detected but ROCm is missing, Ollama will use CPU"))
|
||||
window.ollama_information_label.set_css_classes(['dim-label', 'error'])
|
||||
elif 'msg="amdgpu is supported"' in line:
|
||||
window.ollama_information_label.set_label(_("Using AMD GPU type '{}'").format(line.split('=')[-1]))
|
||||
|
||||
Reference in New Issue
Block a user