--- a/OllamaInterface/i18n/ollama_de.ts Tue Sep 03 15:51:28 2024 +0200 +++ b/OllamaInterface/i18n/ollama_de.ts Tue Sep 03 17:06:05 2024 +0200 @@ -43,7 +43,7 @@ <translation><p>Fehler: Der lokale Server auf <b>{0}</b> antwortet nicht.</p></translation> </message> <message> - <location filename="../OllamaClient.py" line="591" /> + <location filename="../OllamaClient.py" line="589" /> <source><p>Error: The configured server at <b>{0}</b> is not responding.</p></source> <translation><p>Fehler: Der konfigurierte Server auf <b>{0}</b> antwortet nicht.</p></translation> </message> @@ -164,18 +164,38 @@ </message> <message> <location filename="../ConfigurationPage/OllamaPage.ui" line="0" /> - <source>Model Library</source> - <translation>Modellbibliothek</translation> + <source>URLs</source> + <translation>URLs</translation> + </message> + <message> + <location filename="../ConfigurationPage/OllamaPage.ui" line="0" /> + <source>Model Library:</source> + <translation>Modellbibliothek:</translation> + </message> + <message> + <location filename="../ConfigurationPage/OllamaPage.ui" line="0" /> + <source>Enter the URL of the 'ollama' model library. Leave empty to use the default URL.</source> + <translation>Gib die URL der 'ollama' Modellbibliothek ein. Leer lassen zur Verwendung der Standard-URL.</translation> </message> <message> <location filename="../ConfigurationPage/OllamaPage.ui" line="0" /> - <source>URL:</source> - <translation>URL:</translation> + <source>Download:</source> + <translation>Download:</translation> </message> <message> <location filename="../ConfigurationPage/OllamaPage.ui" line="0" /> - <source>Enter the URL of the 'ollama' model library.</source> - <translation>Gib die URL der 'ollama' Modellbibliothek ein.</translation> + <source>Enter the URL of the 'ollama' download page. Leave empty to use the default URL.</source> + <translation>Gib die URL der 'ollama' Downloadseite ein. Leer lassen zur Verwendung der Standard-URL.</translation> + </message> + <message> + <location filename="../ConfigurationPage/OllamaPage.ui" line="0" /> + <source>Blog:</source> + <translation>Blog:</translation> + </message> + <message> + <location filename="../ConfigurationPage/OllamaPage.ui" line="0" /> + <source>Enter the URL of the 'ollama' blog. Leave empty to use the default URL.</source> + <translation>Gib die URL des 'ollama' Blogs ein. Leer lassen zur Verwendung der Standard-URL.</translation> </message> <message> <location filename="../ConfigurationPage/OllamaPage.ui" line="0" /> @@ -340,13 +360,13 @@ <translation>Modellverwaltung</translation> </message> <message> - <location filename="../OllamaWidget.py" line="958" /> + <location filename="../OllamaWidget.py" line="979" /> <location filename="../OllamaWidget.py" line="690" /> <source>List Models</source> <translation>Modelle auflisten</translation> </message> <message> - <location filename="../OllamaWidget.py" line="976" /> + <location filename="../OllamaWidget.py" line="997" /> <location filename="../OllamaWidget.py" line="692" /> <source>List Running Models</source> <translation>Laufende Modelle auflisten</translation> @@ -357,123 +377,143 @@ <translation>Modell Bibliothek anzeigen</translation> </message> <message> - <location filename="../OllamaWidget.py" line="1002" /> - <location filename="../OllamaWidget.py" line="700" /> + <location filename="../OllamaWidget.py" line="1014" /> + <location filename="../OllamaWidget.py" line="701" /> <source>Install Model</source> <translation>Modell installieren</translation> </message> <message> - <location filename="../OllamaWidget.py" line="1075" /> - <location filename="../OllamaWidget.py" line="1066" /> - <location filename="../OllamaWidget.py" line="1055" /> - <location filename="../OllamaWidget.py" line="703" /> + <location filename="../OllamaWidget.py" line="1087" /> + <location filename="../OllamaWidget.py" line="1078" /> + <location filename="../OllamaWidget.py" line="1067" /> + <location filename="../OllamaWidget.py" line="704" /> <source>Remove Model</source> <translation>Modell löschen</translation> </message> <message> - <location filename="../OllamaWidget.py" line="710" /> + <location filename="../OllamaWidget.py" line="711" /> <source>Local Server</source> <translation>Lokaler Server</translation> </message> <message> - <location filename="../OllamaWidget.py" line="712" /> + <location filename="../OllamaWidget.py" line="713" /> <source>Start with Monitoring</source> <translation>Mit Monitoring starten</translation> </message> <message> - <location filename="../OllamaWidget.py" line="716" /> + <location filename="../OllamaWidget.py" line="717" /> <source>Start</source> <translation>Starten</translation> </message> <message> - <location filename="../OllamaWidget.py" line="719" /> + <location filename="../OllamaWidget.py" line="720" /> <source>Stop</source> <translation>Stoppen</translation> </message> <message> + <location filename="../OllamaWidget.py" line="727" /> + <source>ollama URLs</source> + <translation>ollama URLs</translation> + </message> + <message> + <location filename="../OllamaWidget.py" line="729" /> + <source>Model Library</source> + <translation>Modellbibliothek</translation> + </message> + <message> <location filename="../OllamaWidget.py" line="733" /> + <source>Download</source> + <translation>Download</translation> + </message> + <message> + <location filename="../OllamaWidget.py" line="737" /> + <source>Blog</source> + <translation>Blog</translation> + </message> + <message> + <location filename="../OllamaWidget.py" line="754" /> <source>Configure...</source> <translation>Einstellungen...</translation> </message> <message> - <location filename="../OllamaWidget.py" line="773" /> + <location filename="../OllamaWidget.py" line="794" /> <source>Clear All Chat Histories</source> <translation>Alle Chat Verläufe löschen</translation> </message> <message> - <location filename="../OllamaWidget.py" line="774" /> + <location filename="../OllamaWidget.py" line="795" /> <source><p>Do you really want to delete all chat histories? This is <b>irreversible</b>.</p></source> <translation><p>Sollen wirklich alle Chat Verläufe gelöscht werden? Dies ist <b>unumkehrbar</b>.</p></translation> </message> <message> - <location filename="../OllamaWidget.py" line="789" /> + <location filename="../OllamaWidget.py" line="810" /> <source>Import Chat History</source> <translation>Chat Verlauf importieren</translation> </message> <message> - <location filename="../OllamaWidget.py" line="791" /> + <location filename="../OllamaWidget.py" line="812" /> <source>Chat History Files (*.json);;All Files (*)</source> <translation>Chat Verlauf Dateien (*.json);;Alle Dateien (*)</translation> </message> <message> - <location filename="../OllamaWidget.py" line="824" /> - <location filename="../OllamaWidget.py" line="792" /> + <location filename="../OllamaWidget.py" line="845" /> + <location filename="../OllamaWidget.py" line="813" /> <source>Chat History Files (*.json)</source> <translation>Chat Verlauf Dateien (*.json)</translation> </message> <message> - <location filename="../OllamaWidget.py" line="822" /> - <location filename="../OllamaWidget.py" line="812" /> + <location filename="../OllamaWidget.py" line="843" /> + <location filename="../OllamaWidget.py" line="833" /> <source>Export Chat History</source> <translation>Chat Verlauf exportieren</translation> </message> <message> - <location filename="../OllamaWidget.py" line="813" /> + <location filename="../OllamaWidget.py" line="834" /> <source>Select the chats to be exported:</source> <translation>Wähle die zu exportierenden Chats:</translation> </message> <message> - <location filename="../OllamaWidget.py" line="911" /> + <location filename="../OllamaWidget.py" line="932" /> <source>Run Local 'ollama' Server</source> <translation>Lokalen 'ollama' Server ausführen</translation> </message> <message> - <location filename="../OllamaWidget.py" line="912" /> + <location filename="../OllamaWidget.py" line="933" /> <source>The loacl 'ollama' server process could not be started.</source> <translation>Der lokale 'ollama' Serverprozess konnte nicht gestartet werden.</translation> </message> <message> - <location filename="../OllamaWidget.py" line="959" /> + <location filename="../OllamaWidget.py" line="980" /> <source>There are no models available.</source> <translation>Es sind keine Modelle verfügbar.</translation> </message> <message> - <location filename="../OllamaWidget.py" line="977" /> + <location filename="../OllamaWidget.py" line="998" /> <source>There are no models running.</source> <translation>Es werden keine Modelle ausgeführt.</translation> </message> <message> - <location filename="../OllamaWidget.py" line="1003" /> + <location filename="../OllamaWidget.py" line="1015" /> <source>Enter the name of the model to be installed:</source> <translation>Gib den Namen des zu installierenden Modells ein:</translation> </message> <message> - <location filename="../OllamaWidget.py" line="1056" /> + <location filename="../OllamaWidget.py" line="1068" /> <source>Select the model to be removed by the 'ollama' server:</source> <translation>Wähle das vom 'ollama' Server zu entfernende Modell aus:</translation> </message> <message> - <location filename="../OllamaWidget.py" line="1067" /> + <location filename="../OllamaWidget.py" line="1079" /> <source><p>The model <b>{0}</b> was deleted successfully.</p></source> <translation><p>Das Modell <b>{0}</b> wurde erfolgreich entfernt.</p></translation> </message> <message> - <location filename="../OllamaWidget.py" line="1076" /> + <location filename="../OllamaWidget.py" line="1088" /> <source><p>The model <b>{0}</b> could not be removed from the 'ollama' server.</p></source> <translation><p>Das Modell <b>{0}</b> konnte nicht vom 'ollama' Server entfernt werden.</p></translation> </message> <message> - <location filename="../OllamaWidget.py" line="1092" /> + <location filename="../OllamaWidget.py" line="1115" /> <source>Network Error</source> <translation>Netzwerkfehler</translation> </message> @@ -511,25 +551,25 @@ <context> <name>PluginOllamaInterface</name> <message> - <location filename="../../PluginAiOllama.py" line="181" /> - <location filename="../../PluginAiOllama.py" line="180" /> - <location filename="../../PluginAiOllama.py" line="176" /> - <location filename="../../PluginAiOllama.py" line="82" /> + <location filename="../../PluginAiOllama.py" line="184" /> + <location filename="../../PluginAiOllama.py" line="183" /> + <location filename="../../PluginAiOllama.py" line="179" /> + <location filename="../../PluginAiOllama.py" line="83" /> <source>ollama AI Interface</source> <translation>ollama KI Schnittstelle</translation> </message> <message> - <location filename="../../PluginAiOllama.py" line="182" /> + <location filename="../../PluginAiOllama.py" line="185" /> <source>Ctrl+Alt+Shift+O</source> <translation>Ctrl+Alt+Shift+O</translation> </message> <message> - <location filename="../../PluginAiOllama.py" line="188" /> + <location filename="../../PluginAiOllama.py" line="191" /> <source>Switch the input focus to the ollama AI window.</source> <translation>Schaltet den Eingabefokus auf das ollama KI Fenster um.</translation> </message> <message> - <location filename="../../PluginAiOllama.py" line="191" /> + <location filename="../../PluginAiOllama.py" line="194" /> <source><b>Activate ollama AI Interface</b><p>This switches the input focus to the ollama AI window.</p></source> <translation><b>Aktiviere ollama KI Schnittstelle</b><p>Dies schaltet den Eingabefokus auf das ollama KI Fenster um.</p></translation> </message>