Mit lokalen LLMs via n8n und Ollama chatten
Dies ist ein AI-Bereich Automatisierungsworkflow mit 5 Nodes. Hauptsächlich werden ChainLlm, ChatTrigger, LmChatOllama und andere Nodes verwendet, kombiniert mit KI-Technologie für intelligente Automatisierung. Mit lokalen Large Language Models über n8n und Ollama sprechen
- •KI-Service API Key (z.B. OpenAI, Anthropic)
Verwendete Nodes (5)
Kategorie
{
"id": "af8RV5b2TWB2LclA",
"meta": {
"instanceId": "95f2ab28b3dabb8da5d47aa5145b95fe3845f47b20d6343dd5256b6a28ba8fab",
"templateCredsSetupCompleted": true
},
"name": "Chat with local LLMs using n8n and Ollama",
"tags": [],
"nodes": [
{
"id": "475385fa-28f3-45c4-bd1a-10dde79f74f2",
"name": "Bei Chatnachricht-Empfang",
"type": "@n8n/n8n-nodes-langchain.chatTrigger",
"position": [
700,
460
],
"webhookId": "ebdeba3f-6b4f-49f3-ba0a-8253dd226161",
"parameters": {
"options": {}
},
"typeVersion": 1.1
},
{
"id": "61133dc6-dcd9-44ff-85f2-5d8cc2ce813e",
"name": "Ollama Chat Model",
"type": "@n8n/n8n-nodes-langchain.lmChatOllama",
"position": [
900,
680
],
"parameters": {
"options": {}
},
"credentials": {
"ollamaApi": {
"id": "MyYvr1tcNQ4e7M6l",
"name": "Local Ollama"
}
},
"typeVersion": 1
},
{
"id": "3e89571f-7c87-44c6-8cfd-4903d5e1cdc5",
"name": "Notizzettel",
"type": "n8n-nodes-base.stickyNote",
"position": [
160,
80
],
"parameters": {
"width": 485,
"height": 473,
"content": "## Chat with local LLMs using n8n and Ollama\nThis n8n workflow allows you to seamlessly interact with your self-hosted Large Language Models (LLMs) through a user-friendly chat interface. By connecting to Ollama, a powerful tool for managing local LLMs, you can send prompts and receive AI-generated responses directly within n8n.\n\n### How it works\n1. When chat message received: Captures the user's input from the chat interface.\n2. Chat LLM Chain: Sends the input to the Ollama server and receives the AI-generated response.\n3. Delivers the LLM's response back to the chat interface.\n\n### Set up steps\n* Make sure Ollama is installed and running on your machine before executing this workflow.\n* Edit the Ollama address if different from the default.\n"
},
"typeVersion": 1
},
{
"id": "9345cadf-a72e-4d3d-b9f0-d670744065fe",
"name": "Notizzettel1",
"type": "n8n-nodes-base.stickyNote",
"position": [
1040,
660
],
"parameters": {
"color": 6,
"width": 368,
"height": 258,
"content": "## Ollama setup\n* Connect to your local Ollama, usually on http://localhost:11434\n* If running in Docker, make sure that the n8n container has access to the host's network in order to connect to Ollama. You can do this by passing `--net=host` option when starting the n8n Docker container"
},
"typeVersion": 1
},
{
"id": "eeffdd4e-6795-4ebc-84f7-87b5ac4167d9",
"name": "Chat LLM Kette",
"type": "@n8n/n8n-nodes-langchain.chainLlm",
"position": [
920,
460
],
"parameters": {},
"typeVersion": 1.4
}
],
"active": false,
"pinData": {},
"settings": {
"executionOrder": "v1"
},
"versionId": "3af03daa-e085-4774-8676-41578a4cba2d",
"connections": {
"61133dc6-dcd9-44ff-85f2-5d8cc2ce813e": {
"ai_languageModel": [
[
{
"node": "Chat LLM Chain",
"type": "ai_languageModel",
"index": 0
}
]
]
},
"475385fa-28f3-45c4-bd1a-10dde79f74f2": {
"main": [
[
{
"node": "Chat LLM Chain",
"type": "main",
"index": 0
}
]
]
}
}
}Wie verwende ich diesen Workflow?
Kopieren Sie den obigen JSON-Code, erstellen Sie einen neuen Workflow in Ihrer n8n-Instanz und wählen Sie "Aus JSON importieren". Fügen Sie die Konfiguration ein und passen Sie die Anmeldedaten nach Bedarf an.
Für welche Szenarien ist dieser Workflow geeignet?
Anfänger - Künstliche Intelligenz
Ist es kostenpflichtig?
Dieser Workflow ist völlig kostenlos. Beachten Sie jedoch, dass Drittanbieterdienste (wie OpenAI API), die im Workflow verwendet werden, möglicherweise kostenpflichtig sind.
Verwandte Workflows
Mihai Farcas
@mihailtdFull-stack developer with 5+ years streamlining healthcare processes. Proficient in NodeJS, VueJS, MongoDB, PostgreSQL, Kubernetes, and n8n. Ready to optimize your workflows – book a consult via my link.
Diesen Workflow teilen