{"id":7853,"date":"2024-11-04T08:42:18","date_gmt":"2024-11-04T09:42:18","guid":{"rendered":"https:\/\/ciit-software.com\/?p=7853"},"modified":"2024-12-22T10:22:29","modified_gmt":"2024-12-22T11:22:29","slug":"lokale-large-language-models-mit-ollama-auf-linux-installieren","status":"publish","type":"post","link":"https:\/\/ciit-software.com\/en\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/","title":{"rendered":"Installing Local Large Language Models with Ollama on Linux"},"content":{"rendered":"\t\t<div data-elementor-type=\"wp-post\" data-elementor-id=\"7853\" class=\"elementor elementor-7853\" data-elementor-post-type=\"post\">\n\t\t\t\t<div class=\"elementor-element elementor-element-0bebc08 e-flex e-con-boxed e-con e-parent\" data-id=\"0bebc08\" data-element_type=\"container\" data-e-type=\"container\">\n\t\t\t\t\t<div class=\"e-con-inner\">\n\t\t\t\t<div class=\"elementor-element elementor-element-22da60f elementor-widget elementor-widget-text-editor\" data-id=\"22da60f\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<p>In der heutigen digitalen Welt spielen Large Language Models (LLMs) eine immer gr\u00f6\u00dfere Rolle. Viele Unternehmen stehen jedoch vor der Herausforderung, ihre sensiblen Daten sicher zu halten und gleichzeitig leistungsf\u00e4hige KI-Modelle zu nutzen. Eine L\u00f6sung: lokale Large Language Models installieren. In diesem Beitrag zeigen wir Ihnen, wie Sie mit der Software Ollama ein lokales LLM auf einem Linux-Server einrichten. Dies erm\u00f6glicht Ihnen, die Vorteile leistungsstarker Sprachmodelle zu nutzen, ohne sensible Daten in die Cloud zu senden.<\/p>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-19a0594 elementor-widget elementor-widget-heading\" data-id=\"19a0594\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h4 class=\"elementor-heading-title elementor-size-default\">Warum lokale LLMs?<\/h4>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-f60f333 elementor-widget elementor-widget-text-editor\" data-id=\"f60f333\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<p>Viele Unternehmen bevorzugen lokale LLMs, um die Kontrolle \u00fcber ihre Daten zu behalten. Cloud-basierte L\u00f6sungen wie Microsoft Azure oder AWS bieten zwar immense Rechenleistung, doch die Datenhoheit bleibt oft ein kritischer Punkt. Lokale Installationen erm\u00f6glichen es, hochsensible Informationen intern zu verarbeiten und gleichzeitig die Leistungsf\u00e4higkeit moderner Sprachmodelle auszusch\u00f6pfen.\u00a0<span style=\"background-color: var( --e-global-color-primary ); color: var( --e-global-color-text ); font-family: var( --e-global-typography-text-font-family ), Sans-serif; font-weight: var( --e-global-typography-text-font-weight ); text-align: var(--text-align);\">F\u00fcr die Installation von Ollama auf einem Linux-Server ben\u00f6tigen Sie:<\/span><\/p><ul><li><strong>Linux-Distribution<\/strong>: beliebige Linux Distribution, wir verwenden hier Ubuntu Server 24.04 LTS\u00a0<\/li><li><strong>Nvidia-Grafikkarte<\/strong>: Eine leistungsf\u00e4hige Karte wie die Nvidia RTX A5000 sorgt f\u00fcr die n\u00f6tige Rechenleistung, die LLMs ben\u00f6tigen<\/li><li><strong>Docker<\/strong>: Zum Starten von Open WebUI als Docker-Container.<\/li><\/ul><div>\u00a0<\/div>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-065c1d7 elementor-widget elementor-widget-heading\" data-id=\"065c1d7\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h4 class=\"elementor-heading-title elementor-size-default\">Schritt 1: Ollama installieren<\/h4>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-bb7af3b elementor-widget elementor-widget-text-editor\" data-id=\"bb7af3b\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<p>Ollama erm\u00f6glicht die Verwaltung und Nutzung lokaler LLMs. Die Installation ist unkompliziert:<\/p>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-0725a81 elementor-widget elementor-widget-code-highlight\" data-id=\"0725a81\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"code-highlight.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"prismjs-default copy-to-clipboard \">\n\t\t\t<pre data-line=\"\" class=\"highlight-height language-javascript line-numbers\">\n\t\t\t\t<code readonly=\"true\" class=\"language-javascript\">\n\t\t\t\t\t<xmp>sudo curl -fsSL https:\/\/ollama.com\/install.sh | sh\n<\/xmp>\n\t\t\t\t<\/code>\n\t\t\t<\/pre>\n\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-6006a48 elementor-widget elementor-widget-text-editor\" data-id=\"6006a48\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<p>Nach der Installation sollten Sie den Server neu starten, um sicherzustellen, dass alle Kernel-Komponenten korrekt geladen werden.<\/p>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-b4ddd31 elementor-widget elementor-widget-heading\" data-id=\"b4ddd31\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h4 class=\"elementor-heading-title elementor-size-default\">Schritt 2: \u00dcberpr\u00fcfen der Nvidia-Grafikkarte<\/h4>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-5101e1c elementor-widget elementor-widget-text-editor\" data-id=\"5101e1c\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<p>Nutzen Sie <code>nvidia-smi<\/code>, um den Status Ihrer Grafikkarte zu \u00fcberwachen:<\/p>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-e7af0d9 elementor-widget elementor-widget-code-highlight\" data-id=\"e7af0d9\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"code-highlight.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"prismjs-default copy-to-clipboard \">\n\t\t\t<pre data-line=\"\" class=\"highlight-height language-javascript line-numbers\">\n\t\t\t\t<code readonly=\"true\" class=\"language-javascript\">\n\t\t\t\t\t<xmp>nvidia-smi -l 1<\/xmp>\n\t\t\t\t<\/code>\n\t\t\t<\/pre>\n\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-521edba elementor-widget elementor-widget-heading\" data-id=\"521edba\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h4 class=\"elementor-heading-title elementor-size-default\"><br>Schritt 3: Docker installieren<\/h4>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-25f310e elementor-widget elementor-widget-text-editor\" data-id=\"25f310e\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<p>Docker wird ben\u00f6tigt, um Open WebUI zu betreiben. Die Installationsanleitung finden Sie <a href=\"https:\/\/docs.docker.com\/engine\/install\/\" target=\"_new\" rel=\"noopener\">hier<\/a>.<\/p>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-69626ad elementor-widget elementor-widget-heading\" data-id=\"69626ad\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h4 class=\"elementor-heading-title elementor-size-default\">Schritt 4: Open WebUI starten<\/h4>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-322fd15 elementor-widget elementor-widget-text-editor\" data-id=\"322fd15\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<div class=\"contain-inline-size rounded-md border-[0.5px] border-token-border-medium relative bg-token-sidebar-surface-primary dark:bg-gray-950\"><div class=\"flex items-center text-token-text-secondary px-4 py-2 text-xs font-sans justify-between rounded-t-md h-9 bg-token-sidebar-surface-primary dark:bg-token-main-surface-secondary\">\u00a0<span style=\"color: var( --e-global-color-text );font-family: var( --e-global-typography-text-font-family ), Sans-serif;font-weight: var( --e-global-typography-text-font-weight );text-align: var(--text-align);background-color: var( --e-global-color-primary )\">Starten Sie Open WebUI als Docker-Container, um die Benutzeroberfl\u00e4che f\u00fcr Ihre LLMs zu nutzen:<\/span><\/div><\/div>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-d61fe3c elementor-widget elementor-widget-code-highlight\" data-id=\"d61fe3c\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"code-highlight.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"prismjs-default copy-to-clipboard \">\n\t\t\t<pre data-line=\"\" class=\"highlight-height language-javascript line-numbers\">\n\t\t\t\t<code readonly=\"true\" class=\"language-javascript\">\n\t\t\t\t\t<xmp>docker run -d --network=host -v open-webui:\/app\/backend\/data -e OLLAMA_BASE_URL=http:\/\/127.0.0.1:11434 --name open-webui --restart always ghcr.io\/open-webui\/open-webui:main\n<\/xmp>\n\t\t\t\t<\/code>\n\t\t\t<\/pre>\n\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-a13febb elementor-widget elementor-widget-heading\" data-id=\"a13febb\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h4 class=\"elementor-heading-title elementor-size-default\"><br>Schritt 5: Zugriff auf die Benutzeroberfl\u00e4che<\/h4>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-a94d997 elementor-widget elementor-widget-text-editor\" data-id=\"a94d997\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<div class=\"contain-inline-size rounded-md border-[0.5px] border-token-border-medium relative bg-token-sidebar-surface-primary dark:bg-gray-950\"><div class=\"flex items-center text-token-text-secondary px-4 py-2 text-xs font-sans justify-between rounded-t-md h-9 bg-token-sidebar-surface-primary dark:bg-token-main-surface-secondary\">Rufen Sie Open WebUI im Browser auf, indem Sie die IP-Adresse Ihres Servers eingeben, z. B.:<\/div><\/div>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-0f90426 elementor-widget elementor-widget-code-highlight\" data-id=\"0f90426\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"code-highlight.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"prismjs-default copy-to-clipboard \">\n\t\t\t<pre data-line=\"\" class=\"highlight-height language-javascript line-numbers\">\n\t\t\t\t<code readonly=\"true\" class=\"language-javascript\">\n\t\t\t\t\t<xmp>http:\/\/192.168.0.5:8080\n<\/xmp>\n\t\t\t\t<\/code>\n\t\t\t<\/pre>\n\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-44e7d06 elementor-widget elementor-widget-text-editor\" data-id=\"44e7d06\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<div class=\"contain-inline-size rounded-md border-[0.5px] border-token-border-medium relative bg-token-sidebar-surface-primary dark:bg-gray-950\"><div class=\"flex items-center text-token-text-secondary px-4 py-2 text-xs font-sans justify-between rounded-t-md h-9 bg-token-sidebar-surface-primary dark:bg-token-main-surface-secondary\">Erstellen Sie einen Admin-Account, um Modelle zu installieren und Abfragen an die Sprachmodelle zu stellen.<\/div><\/div>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-d6341a3 elementor-widget elementor-widget-heading\" data-id=\"d6341a3\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h4 class=\"elementor-heading-title elementor-size-default\"><br>Schritt 6: Modelle installieren und nutzen<\/h4>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-df389aa elementor-widget elementor-widget-text-editor\" data-id=\"df389aa\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<div class=\"contain-inline-size rounded-md border-[0.5px] border-token-border-medium relative bg-token-sidebar-surface-primary dark:bg-gray-950\"><div class=\"flex items-center text-token-text-secondary px-4 py-2 text-xs font-sans justify-between rounded-t-md h-9 bg-token-sidebar-surface-primary dark:bg-token-main-surface-secondary\">Modelle k\u00f6nnen entweder direkt \u00fcber die Kommandozeile oder die WebUI installiert werden. Zum Beispiel:<\/div><\/div>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-c3eda1e elementor-widget elementor-widget-code-highlight\" data-id=\"c3eda1e\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"code-highlight.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"prismjs-default copy-to-clipboard \">\n\t\t\t<pre data-line=\"\" class=\"highlight-height language-javascript line-numbers\">\n\t\t\t\t<code readonly=\"true\" class=\"language-javascript\">\n\t\t\t\t\t<xmp>ollama pull llama3\n<\/xmp>\n\t\t\t\t<\/code>\n\t\t\t<\/pre>\n\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-154f15e elementor-widget elementor-widget-text-editor\" data-id=\"154f15e\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<div class=\"contain-inline-size rounded-md border-[0.5px] border-token-border-medium relative bg-token-sidebar-surface-primary dark:bg-gray-950\"><div class=\"flex items-center text-token-text-secondary px-4 py-2 text-xs font-sans justify-between rounded-t-md h-9 bg-token-sidebar-surface-primary dark:bg-token-main-surface-secondary\">Die WebUI bietet eine benutzerfreundliche M\u00f6glichkeit, Modelle zu verwalten und Abfragen zu stellen, \u00e4hnlich wie bei ChatGPT.<\/div><\/div>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-a0ddc60 elementor-widget elementor-widget-heading\" data-id=\"a0ddc60\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h4 class=\"elementor-heading-title elementor-size-default\"><br>Tipps zur Hardware-Optimierung<\/h4>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-e42c653 elementor-widget elementor-widget-text-editor\" data-id=\"e42c653\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<div>F\u00fcr den produktiven Einsatz von LLMs sind leistungsf\u00e4hige Hardware-Ressourcen entscheidend. Eingesetzte Grafikkarten und Server-Boards mit ausreichendem RAM und guter Erweiterbarkeit bzgl. PCI Steckpl\u00e4tzen f\u00fcr die Grafikkarten sind ideal, um anspruchsvolle Modelle effizient zu betreiben. Die Gr\u00f6\u00dfe der verwendbaren Modelle h\u00e4ngt entscheiden von Anzahl und Leistungsf\u00e4higkeit der Grafikkarten ab.<\/div><div>\u00a0<\/div><h3>Beispiel-Konfigurationen<\/h3><p>Einstiegskonfiguration (f\u00fcr Llama 7B und einfache Anwendungen)<\/p><ul><li>CPU: AMD Ryzen 9 oder Intel i9<\/li><li>GPU: NVIDIA RTX 3060 mit 12 GB VRAM<\/li><li>RAM: 32 GB<\/li><li>Speicher: 1 TB NVMe SSD<\/li><\/ul><p>\u00a0<\/p><p>Fortgeschrittene Konfiguration (f\u00fcr Llama 13B bis 30B)<\/p><ul><li>CPU: AMD Threadripper oder Intel Xeon<\/li><li>GPU: NVIDIA RTX 3090 oder A6000 mit mindestens 24 GB VRAM<\/li><li>RAM: 64-128 GB<\/li><li>Speicher: 2 TB NVMe SSD<\/li><\/ul><p>\u00a0<\/p><p>High-End-Konfiguration (f\u00fcr Llama 65B und anspruchsvolle Anwendungen)<\/p><ul><li>CPU: Dual AMD EPYC oder Intel Xeon<\/li><li>GPU: NVIDIA A100 oder H100 (40 GB oder mehr VRAM) oder ein Cluster aus mehreren GPUs<\/li><li>RAM: 128 GB oder mehr<\/li><li>Speicher: 4 TB NVMe SSD<\/li><\/ul>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-60f0125 elementor-widget elementor-widget-heading\" data-id=\"60f0125\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"heading.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t<h4 class=\"elementor-heading-title elementor-size-default\"><br>Fazit<\/h4>\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-41b40d9 elementor-widget elementor-widget-text-editor\" data-id=\"41b40d9\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<div class=\"contain-inline-size rounded-md border-[0.5px] border-token-border-medium relative bg-token-sidebar-surface-primary dark:bg-gray-950\"><div class=\"flex items-center text-token-text-secondary px-4 py-2 text-xs font-sans justify-between rounded-t-md h-9 bg-token-sidebar-surface-primary dark:bg-token-main-surface-secondary\"><p>Mit Ollama k\u00f6nnen Sie lokale LLMs effizient auf Ihrem Linux-Server betreiben und dabei die volle Kontrolle \u00fcber Ihre Daten behalten. Diese L\u00f6sung ist besonders f\u00fcr Unternehmen geeignet, die sensible Informationen verarbeiten und dennoch die Vorteile moderner Sprachmodelle nutzen m\u00f6chten. Zum praktischen Video Tutorial geht es hier:<\/p><\/div><\/div>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-1e2897c elementor-widget elementor-widget-video\" data-id=\"1e2897c\" data-element_type=\"widget\" data-e-type=\"widget\" data-settings=\"{&quot;youtube_url&quot;:&quot;https:\\\/\\\/www.youtube.com\\\/watch?v=6IsuEWmyQfs&quot;,&quot;video_type&quot;:&quot;youtube&quot;,&quot;controls&quot;:&quot;yes&quot;}\" data-widget_type=\"video.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<div class=\"elementor-wrapper elementor-open-inline\">\n\t\t\t<div class=\"elementor-video\"><\/div>\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t","protected":false},"excerpt":{"rendered":"<p>In der heutigen digitalen Welt spielen Large Language Models (LLMs) eine immer gr\u00f6\u00dfere Rolle. Viele Unternehmen stehen jedoch vor der Herausforderung, ihre sensiblen Daten sicher zu halten und gleichzeitig leistungsf\u00e4hige KI-Modelle zu nutzen. Eine L\u00f6sung: lokale Large Language Models installieren. In diesem Beitrag zeigen wir Ihnen, wie Sie mit der Software Ollama ein lokales LLM [&hellip;]<\/p>\n","protected":false},"author":8,"featured_media":8070,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"inline_featured_image":false,"footnotes":""},"categories":[1],"tags":[57,115,114,112,119,121,83,109,108,116,74,111,110,113,52,120,118,117],"class_list":["post-7853","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-uncategorized","tag-ai","tag-cloud","tag-datenschutz","tag-docker","tag-grafikkarte","tag-hardware","tag-ki","tag-ki-modell","tag-large-language-models","tag-linux","tag-llm","tag-nvidia","tag-ollama","tag-open-webui","tag-software","tag-sprachmodell","tag-tech","tag-unternehmen"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.3 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>Lokale Large Language Models mit Ollama auf Linux installieren - CIIT<\/title>\n<meta name=\"description\" content=\"Installieren Sie lokale Large Language Models mit Ollama auf Linux, um KI-Modelle zu nutzen und die Kontrolle \u00fcber Ihre Daten zu behalten.\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/ciit-software.com\/en\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Lokale Large Language Models mit Ollama auf Linux installieren - CIIT\" \/>\n<meta property=\"og:description\" content=\"Installieren Sie lokale Large Language Models mit Ollama auf Linux, um KI-Modelle zu nutzen und die Kontrolle \u00fcber Ihre Daten zu behalten.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/ciit-software.com\/en\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/\" \/>\n<meta property=\"og:site_name\" content=\"CIIT\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/CIITIndividualsoftware\" \/>\n<meta property=\"article:published_time\" content=\"2024-11-04T09:42:18+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2024-12-22T11:22:29+00:00\" \/>\n<meta name=\"author\" content=\"Michael Schaffler-Gl\u00f6\u00dfl\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@CIIT_at\" \/>\n<meta name=\"twitter:site\" content=\"@CIIT_at\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Michael Schaffler-Gl\u00f6\u00dfl\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"4 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\\\/\\\/ciit-software.com\\\/lokale-large-language-models-mit-ollama-auf-linux-installieren\\\/#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/ciit-software.com\\\/lokale-large-language-models-mit-ollama-auf-linux-installieren\\\/\"},\"author\":{\"name\":\"Michael Schaffler-Gl\u00f6\u00dfl\",\"@id\":\"https:\\\/\\\/ciit-software.com\\\/#\\\/schema\\\/person\\\/a4b77611f602115013aca7ecdfa1cde2\"},\"headline\":\"Lokale Large Language Models mit Ollama auf Linux installieren\",\"datePublished\":\"2024-11-04T09:42:18+00:00\",\"dateModified\":\"2024-12-22T11:22:29+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/ciit-software.com\\\/lokale-large-language-models-mit-ollama-auf-linux-installieren\\\/\"},\"wordCount\":565,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\\\/\\\/ciit-software.com\\\/#organization\"},\"image\":{\"@id\":\"https:\\\/\\\/ciit-software.com\\\/lokale-large-language-models-mit-ollama-auf-linux-installieren\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/ciit-software.com\\\/wp-content\\\/uploads\\\/2024\\\/11\\\/7-scaled.avif\",\"keywords\":[\"ai\",\"Cloud\",\"Datenschutz\",\"Docker\",\"Grafikkarte\",\"Hardware\",\"KI\",\"KI Modell\",\"Large Language Models\",\"Linux\",\"LLM\",\"Nvidia\",\"Ollama\",\"Open WebUI\",\"software\",\"Sprachmodell\",\"Tech\",\"Unternehmen\"],\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\\\/\\\/ciit-software.com\\\/lokale-large-language-models-mit-ollama-auf-linux-installieren\\\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/ciit-software.com\\\/lokale-large-language-models-mit-ollama-auf-linux-installieren\\\/\",\"url\":\"https:\\\/\\\/ciit-software.com\\\/lokale-large-language-models-mit-ollama-auf-linux-installieren\\\/\",\"name\":\"Lokale Large Language Models mit Ollama auf Linux installieren - CIIT\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/ciit-software.com\\\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\\\/\\\/ciit-software.com\\\/lokale-large-language-models-mit-ollama-auf-linux-installieren\\\/#primaryimage\"},\"image\":{\"@id\":\"https:\\\/\\\/ciit-software.com\\\/lokale-large-language-models-mit-ollama-auf-linux-installieren\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/ciit-software.com\\\/wp-content\\\/uploads\\\/2024\\\/11\\\/7-scaled.avif\",\"datePublished\":\"2024-11-04T09:42:18+00:00\",\"dateModified\":\"2024-12-22T11:22:29+00:00\",\"description\":\"Installieren Sie lokale Large Language Models mit Ollama auf Linux, um KI-Modelle zu nutzen und die Kontrolle \u00fcber Ihre Daten zu behalten.\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/ciit-software.com\\\/lokale-large-language-models-mit-ollama-auf-linux-installieren\\\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/ciit-software.com\\\/lokale-large-language-models-mit-ollama-auf-linux-installieren\\\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/ciit-software.com\\\/lokale-large-language-models-mit-ollama-auf-linux-installieren\\\/#primaryimage\",\"url\":\"https:\\\/\\\/ciit-software.com\\\/wp-content\\\/uploads\\\/2024\\\/11\\\/7-scaled.avif\",\"contentUrl\":\"https:\\\/\\\/ciit-software.com\\\/wp-content\\\/uploads\\\/2024\\\/11\\\/7-scaled.avif\",\"width\":2560,\"height\":1440},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/ciit-software.com\\\/lokale-large-language-models-mit-ollama-auf-linux-installieren\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/ciit-software.com\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Lokale Large Language Models mit Ollama auf Linux installieren\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/ciit-software.com\\\/#website\",\"url\":\"https:\\\/\\\/ciit-software.com\\\/\",\"name\":\"CIIT\",\"description\":\"Individual Software\",\"publisher\":{\"@id\":\"https:\\\/\\\/ciit-software.com\\\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/ciit-software.com\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\\\/\\\/ciit-software.com\\\/#organization\",\"name\":\"CIIT GmbH\",\"url\":\"https:\\\/\\\/ciit-software.com\\\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/ciit-software.com\\\/#\\\/schema\\\/logo\\\/image\\\/\",\"url\":\"https:\\\/\\\/ciit-software.com\\\/wp-content\\\/uploads\\\/2024\\\/01\\\/cropped-CIIT_Logo_Software-Develoment.png\",\"contentUrl\":\"https:\\\/\\\/ciit-software.com\\\/wp-content\\\/uploads\\\/2024\\\/01\\\/cropped-CIIT_Logo_Software-Develoment.png\",\"width\":512,\"height\":512,\"caption\":\"CIIT GmbH\"},\"image\":{\"@id\":\"https:\\\/\\\/ciit-software.com\\\/#\\\/schema\\\/logo\\\/image\\\/\"},\"sameAs\":[\"https:\\\/\\\/www.facebook.com\\\/CIITIndividualsoftware\",\"https:\\\/\\\/x.com\\\/CIIT_at\"]},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/ciit-software.com\\\/#\\\/schema\\\/person\\\/a4b77611f602115013aca7ecdfa1cde2\",\"name\":\"Michael Schaffler-Gl\u00f6\u00dfl\",\"url\":\"https:\\\/\\\/ciit-software.com\\\/en\\\/author\\\/michael\\\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Lokale Large Language Models mit Ollama auf Linux installieren - CIIT","description":"Installieren Sie lokale Large Language Models mit Ollama auf Linux, um KI-Modelle zu nutzen und die Kontrolle \u00fcber Ihre Daten zu behalten.","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/ciit-software.com\/en\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/","og_locale":"en_US","og_type":"article","og_title":"Lokale Large Language Models mit Ollama auf Linux installieren - CIIT","og_description":"Installieren Sie lokale Large Language Models mit Ollama auf Linux, um KI-Modelle zu nutzen und die Kontrolle \u00fcber Ihre Daten zu behalten.","og_url":"https:\/\/ciit-software.com\/en\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/","og_site_name":"CIIT","article_publisher":"https:\/\/www.facebook.com\/CIITIndividualsoftware","article_published_time":"2024-11-04T09:42:18+00:00","article_modified_time":"2024-12-22T11:22:29+00:00","author":"Michael Schaffler-Gl\u00f6\u00dfl","twitter_card":"summary_large_image","twitter_creator":"@CIIT_at","twitter_site":"@CIIT_at","twitter_misc":{"Written by":"Michael Schaffler-Gl\u00f6\u00dfl","Est. reading time":"4 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/ciit-software.com\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/#article","isPartOf":{"@id":"https:\/\/ciit-software.com\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/"},"author":{"name":"Michael Schaffler-Gl\u00f6\u00dfl","@id":"https:\/\/ciit-software.com\/#\/schema\/person\/a4b77611f602115013aca7ecdfa1cde2"},"headline":"Lokale Large Language Models mit Ollama auf Linux installieren","datePublished":"2024-11-04T09:42:18+00:00","dateModified":"2024-12-22T11:22:29+00:00","mainEntityOfPage":{"@id":"https:\/\/ciit-software.com\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/"},"wordCount":565,"commentCount":0,"publisher":{"@id":"https:\/\/ciit-software.com\/#organization"},"image":{"@id":"https:\/\/ciit-software.com\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/#primaryimage"},"thumbnailUrl":"https:\/\/ciit-software.com\/wp-content\/uploads\/2024\/11\/7-scaled.avif","keywords":["ai","Cloud","Datenschutz","Docker","Grafikkarte","Hardware","KI","KI Modell","Large Language Models","Linux","LLM","Nvidia","Ollama","Open WebUI","software","Sprachmodell","Tech","Unternehmen"],"inLanguage":"en-US","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/ciit-software.com\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/ciit-software.com\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/","url":"https:\/\/ciit-software.com\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/","name":"Lokale Large Language Models mit Ollama auf Linux installieren - CIIT","isPartOf":{"@id":"https:\/\/ciit-software.com\/#website"},"primaryImageOfPage":{"@id":"https:\/\/ciit-software.com\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/#primaryimage"},"image":{"@id":"https:\/\/ciit-software.com\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/#primaryimage"},"thumbnailUrl":"https:\/\/ciit-software.com\/wp-content\/uploads\/2024\/11\/7-scaled.avif","datePublished":"2024-11-04T09:42:18+00:00","dateModified":"2024-12-22T11:22:29+00:00","description":"Installieren Sie lokale Large Language Models mit Ollama auf Linux, um KI-Modelle zu nutzen und die Kontrolle \u00fcber Ihre Daten zu behalten.","breadcrumb":{"@id":"https:\/\/ciit-software.com\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/ciit-software.com\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/ciit-software.com\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/#primaryimage","url":"https:\/\/ciit-software.com\/wp-content\/uploads\/2024\/11\/7-scaled.avif","contentUrl":"https:\/\/ciit-software.com\/wp-content\/uploads\/2024\/11\/7-scaled.avif","width":2560,"height":1440},{"@type":"BreadcrumbList","@id":"https:\/\/ciit-software.com\/lokale-large-language-models-mit-ollama-auf-linux-installieren\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/ciit-software.com\/"},{"@type":"ListItem","position":2,"name":"Lokale Large Language Models mit Ollama auf Linux installieren"}]},{"@type":"WebSite","@id":"https:\/\/ciit-software.com\/#website","url":"https:\/\/ciit-software.com\/","name":"CIIT","description":"Individual software","publisher":{"@id":"https:\/\/ciit-software.com\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/ciit-software.com\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/ciit-software.com\/#organization","name":"CIIT GmbH","url":"https:\/\/ciit-software.com\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/ciit-software.com\/#\/schema\/logo\/image\/","url":"https:\/\/ciit-software.com\/wp-content\/uploads\/2024\/01\/cropped-CIIT_Logo_Software-Develoment.png","contentUrl":"https:\/\/ciit-software.com\/wp-content\/uploads\/2024\/01\/cropped-CIIT_Logo_Software-Develoment.png","width":512,"height":512,"caption":"CIIT GmbH"},"image":{"@id":"https:\/\/ciit-software.com\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/CIITIndividualsoftware","https:\/\/x.com\/CIIT_at"]},{"@type":"Person","@id":"https:\/\/ciit-software.com\/#\/schema\/person\/a4b77611f602115013aca7ecdfa1cde2","name":"Michael Schaffler-Gl\u00f6\u00dfl","url":"https:\/\/ciit-software.com\/en\/author\/michael\/"}]}},"_links":{"self":[{"href":"https:\/\/ciit-software.com\/en\/wp-json\/wp\/v2\/posts\/7853","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/ciit-software.com\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/ciit-software.com\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/ciit-software.com\/en\/wp-json\/wp\/v2\/users\/8"}],"replies":[{"embeddable":true,"href":"https:\/\/ciit-software.com\/en\/wp-json\/wp\/v2\/comments?post=7853"}],"version-history":[{"count":0,"href":"https:\/\/ciit-software.com\/en\/wp-json\/wp\/v2\/posts\/7853\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/ciit-software.com\/en\/wp-json\/wp\/v2\/media\/8070"}],"wp:attachment":[{"href":"https:\/\/ciit-software.com\/en\/wp-json\/wp\/v2\/media?parent=7853"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/ciit-software.com\/en\/wp-json\/wp\/v2\/categories?post=7853"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/ciit-software.com\/en\/wp-json\/wp\/v2\/tags?post=7853"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}