{ "meta": { "instanceId": "workflow-65f089a6", "versionId": "1.0.0", "createdAt": "2025-09-29T07:07:54.801346", "updatedAt": "2025-09-29T07:07:54.801395", "owner": "n8n-user", "license": "MIT", "category": "automation", "status": "active", "priority": "high", "environment": "production" }, "nodes": [ { "id": "trigger-3373165d", "name": "Manual Trigger", "type": "n8n-nodes-base.manualTrigger", "typeVersion": 1, "position": [ 100, 100 ], "parameters": {} }, { "id": "27e5f0c0-ba88-4c28-b3be-99c973be15cb", "name": "Sticky Note", "type": "n8n-nodes-base.stickyNote", "position": [ -480, -140 ], "parameters": { "width": 1083, "height": 357, "content": "## This is an example of basic LLM Chain connected to an open-source model\n### The Chain is connected to the Mistral-7B-Instruct-v0.1 model, but you can change this\n\nPlease note the initial prompt that guides the model:\n```\nYou are a helpful assistant.\nPlease reply politely to the users.\nUse emojis and a text.\nQ: {{ $json.input }}\nA: \n```\n\nThis way the model \"knows\" that it needs to answer the question right after the `A: `.\n\nSince Hugging Face node is this is an inference mode, it does not support LangChain Agents at the moment. Please use [Ollama Chat Model]({{ $env.WEBHOOK_URL }} node for that" }, "typeVersion": 1, "notes": "This stickyNote node performs automated tasks as part of the workflow." }, { "id": "4756d5a8-7027-4942-b214-a5ff8310869a", "name": "When chat message received", "type": "n8n-nodes-base.noOp", "position": [ -200, 280 ], "webhookId": "bf2e38b8-566a-4aeb-8efe-28240f4a6991", "parameters": { "options": {} }, "typeVersion": 1.1, "notes": "This chatTrigger node performs automated tasks as part of the workflow." }, { "id": "20a36351-8579-4ac6-9746-526b072aeaa6", "name": "Basic LLM Chain", "type": "n8n-nodes-base.noOp", "position": [ 20, 280 ], "parameters": { "messages": { "messageValues": [ { "message": "=You are a helpful assistant. Please reply politely to the users. Use emojis and a text." } ] } }, "typeVersion": 1.5, "notes": "This chainLlm node performs automated tasks as part of the workflow." }, { "id": "9b88e307-3ad5-4167-8c5f-e5827f7444ac", "name": "Hugging Face Inference Model", "type": "n8n-nodes-base.noOp", "position": [ 120, 440 ], "parameters": { "model": "mistralai/Mistral-7B-Instruct-v0.1", "options": { "maxTokens": "YOUR_TOKEN_HERE", "temperature": 0.8, "frequencyPenalty": 2 } }, "credentials": { "huggingFaceApi": { "id": "ARQ5mOhvBxi283Qk", "name": "HuggingFaceApi account" } }, "typeVersion": 1, "notes": "This lmOpenHuggingFaceInference node performs automated tasks as part of the workflow." }, { "id": "error-f9679ada", "name": "Error Handler", "type": "n8n-nodes-base.stopAndError", "typeVersion": 1, "position": [ 1000, 400 ], "parameters": { "message": "Workflow execution error", "options": {} } } ], "pinData": {}, "connections": {}, "name": "Stickynote Workflow", "settings": { "executionOrder": "v1", "saveManualExecutions": true, "callerPolicy": "workflowsFromSameOwner", "errorWorkflow": null, "timezone": "UTC", "executionTimeout": 3600, "maxExecutions": 1000, "retryOnFail": true, "retryCount": 3, "retryDelay": 1000 }, "description": "Automated workflow: Stickynote Workflow. This workflow integrates 4 different services: lmOpenHuggingFaceInference, stickyNote, chainLlm, chatTrigger. It contains 4 nodes and follows best practices for error handling and security.", "tags": [ "automation", "n8n", "production-ready", "excellent", "optimized" ], "notes": "Excellent quality workflow: Stickynote Workflow. This workflow has been optimized for production use with comprehensive error handling, security, and documentation." }