Chat with Local LLMs Using n8n and Ollama

This workflow allows users to engage in real-time conversations with AI through a locally deployed large language model, ensuring data security and privacy. Users can input text in the chat interface, and the system will utilize the powerful local model to generate intelligent responses, enhancing interaction efficiency. It is suitable for internal customer service in enterprises, model testing by researchers, and natural language processing tasks that require high response speed, helping users achieve a secure and convenient automated chat system.

Tags

Local LLMn8n Integration

Workflow Name

Chat with Local LLMs Using n8n and Ollama

Key Features and Highlights

This workflow enables seamless interaction with locally deployed large language models (LLMs) through the n8n platform. Leveraging Ollama, a powerful local language model management tool, users can send text prompts directly within the n8n chat interface and receive AI-generated intelligent responses in real time, ensuring data privacy while enhancing interaction efficiency.

Core Problems Addressed

It addresses the dual needs of data security and response speed when using large language models by avoiding the transmission of sensitive data to the cloud. The workflow improves the convenience of invoking local models and facilitates automated integration.

Application Scenarios

  • Internal intelligent customer service within enterprises, ensuring customer data security
  • Local model testing and debugging for researchers and developers
  • Integration of intelligent Q&A features in automated workflows
  • Natural language processing tasks requiring high response speed and low dependency on network conditions

Main Workflow Steps

  1. Receive Chat Messages: Listen for and capture messages sent by users through the chat interface.
  2. Invoke Ollama Local Model: Forward user input to the local Ollama server to process using pre-configured language models.
  3. Return Model Response: Receive the model-generated reply and deliver it back to the chat interface for real-time interaction.

Involved Systems or Services

  • n8n: Workflow automation and trigger platform
  • Ollama: Local large language model management and invocation tool
  • Webhook: Used to receive chat message trigger events
  • LangChain Node: Manages and invokes conversation chains

Target Users and Value Proposition

This workflow is ideal for enterprises and individual users with high data privacy requirements, especially those capable of local deployment and seeking to combine automation platforms for intelligent chat interactions. It enables rapid development of secure, stable, and scalable local intelligent dialogue systems, enhancing business automation and user experience.

Recommend Templates

Automated Speech Recognition Workflow

This workflow automates the reading of local WAV format audio files and calls the Wit.ai speech recognition API for intelligent transcription, simplifying the process of converting speech to text. Through automation, it addresses the need for converting audio files to text, enhancing processing efficiency and accuracy. It is suitable for scenarios such as customer service and meeting management, significantly reducing labor costs and promoting intelligent office practices and data applications.

Speech RecognitionAuto Transcription

AI-Based Automatic Image Title and Watermark Generation

This workflow utilizes the Google Gemini multimodal visual language model to automatically generate structured titles and descriptions for input images, intelligently overlaying them as watermarks. The entire process includes steps such as image downloading, resizing, text generation, format parsing, and image editing, achieving intelligent understanding and automated annotation of visual content. This significantly enhances content production efficiency and image protection capabilities. It is applicable in various scenarios, including media publishing, social media management, and copyright protection.

AI Image GenerationAuto Watermark

Use Any LLM Model via OpenRouter

This workflow enables flexible invocation and management of various large language models through the OpenRouter platform. Users can dynamically select models and input content simply by triggering chat messages, enhancing the efficiency of interactions. Its built-in chat memory function ensures contextual coherence, preventing information loss. This makes it suitable for scenarios such as intelligent customer service, content generation, and automated office tasks, greatly simplifying the integration and management of multiple models, making it ideal for AI developers and teams.

Multi-modelChat Memory

Chinese Translator

This workflow automatically translates text or image content sent by users into Chinese by receiving messages from the Line chat bot, and provides pinyin and English definitions. It supports intelligent processing of various message types and leverages a powerful AI language model to achieve high-quality bidirectional translation between Chinese and English, as well as image text recognition. This tool is not only suitable for language learners but also provides convenient cross-language communication solutions for businesses and travelers, enhancing the user interaction experience.

Chinese TranslationSmart Translation

Chinese Vocabulary Intelligent Practice Assistant

This workflow builds an intelligent Chinese vocabulary practice assistant that interacts via Telegram, provides vocabulary support through Google Sheets, and uses AI technology to generate multiple-choice questions. It not only evaluates users' answers in real-time and provides feedback but also features multi-turn conversation memory to ensure a personalized learning experience. It is suitable for Chinese learners, educational institutions, and individual self-learners, significantly enhancing the interactivity and efficiency of learning.

Chinese VocabularySmart Practice

Calendly Invitation Intelligent Analysis and Notion Data Synchronization Workflow

This workflow automates the connection between Calendly invitation events and Humantic AI's personality analysis, allowing for real-time access to personalized data about invitees. The analysis results are structured and synchronized to a Notion database. This enables businesses to gain deeper insights into the personality traits of clients or candidates, enhancing the quality of recruitment and sales decisions. Additionally, it eliminates data silos, achieves centralized information management, optimizes communication strategies, and significantly improves work efficiency.

Personality AnalysisNotion Sync

LangChain - Example - Code Node Example

This workflow utilizes custom code nodes and the LangChain framework to demonstrate flexible interactions with OpenAI language models. By manually triggering and inputting natural language queries, users can generate intelligent responses and integrate external knowledge bases (such as Wikipedia), enabling the automation of complex tasks. It is suitable for scenarios such as intelligent Q&A chatbots, natural language interfaces, and educational assistance systems, enhancing the capabilities of automated intelligent Q&A and tool invocation to meet diverse customization needs.

LangChainSmart QA

Flux AI Image Generator

This workflow automatically invokes multiple advanced image generation models to quickly produce high-quality artistic images based on user-inputted text descriptions and selected art styles. It supports a variety of unique styles, and the generated images are automatically uploaded to cloud storage and displayed through a customized webpage, ensuring a smooth user experience. This process simplifies the complexity of traditional image generation, making artistic creation, marketing content production, and personalized design more convenient and efficient, catering to the needs of different users.

AI Image GenerationHuggingface