{ "name": "n8n-3-3: Perplexity", "nodes": [ { "parameters": { "content": "## Create Article from Perplexity Research", "height": 465.2493344282225, "width": 625.851492623043, "color": 3 }, "id": "a4106938-68c9-43b7-bc4b-9c39b8151add", "name": "Sticky Note", "type": "n8n-nodes-base.stickyNote", "position": [ 0, 600 ], "typeVersion": 1 }, { "parameters": { "content": "## Convert Article into HTML", "height": 464.14912719677955, "width": 479.02028317328745, "color": 4 }, "id": "5b937528-3801-40ca-a43a-e2c9805401f1", "name": "Sticky Note1", "type": "n8n-nodes-base.stickyNote", "position": [ 920, 600 ], "typeVersion": 1 }, { "parameters": { "model": "gpt-4o-mini-2024-07-18", "options": { "responseFormat": "text" } }, "id": "00d55ec6-ef21-439d-886d-72c9b2806df2", "name": "gpt-4o-mini", "type": "@n8n/n8n-nodes-langchain.lmChatOpenAi", "position": [ 1060, 900 ], "typeVersion": 1, "credentials": { "openAiApi": { "id": "qKk5wVYPDLRt63R9", "name": "OpenAi account" } } }, { "parameters": { "options": { "frequencyPenalty": 0, "maxTokens": -1, "responseFormat": "text", "presencePenalty": 0, "temperature": 0, "timeout": 60000, "maxRetries": 2, "topP": 1 } }, "id": "b96d7e3e-a5d3-445f-8ccc-238ca73efbfa", "name": "gpt-4o-mini1", "type": "@n8n/n8n-nodes-langchain.lmChatOpenAi", "position": [ 1240, 280 ], "typeVersion": 1, "credentials": { "openAiApi": { "id": "qKk5wVYPDLRt63R9", "name": "OpenAi account" } } }, { "parameters": { "schemaType": "manual", "inputSchema": "{\n \"type\": \"object\",\n \"properties\": {\n \"article\": {\n \"type\": \"object\",\n \"required\": [\"category\", \"title\", \"metadata\", \"content\", \"hashtags\"],\n \"properties\": {\n \"category\": {\n \"type\": \"string\",\n \"description\": \"Article category\"\n },\n \"title\": {\n \"type\": \"string\",\n \"description\": \"Article title\"\n },\n \"metadata\": {\n \"type\": \"object\",\n \"properties\": {\n \"timePosted\": {\n \"type\": \"string\",\n \"description\": \"Time since article was posted\"\n },\n \"author\": {\n \"type\": \"string\",\n \"description\": \"Article author name\"\n },\n \"tag\": {\n \"type\": \"string\",\n \"description\": \"Article primary tag\"\n }\n },\n \"required\": [\"timePosted\", \"author\", \"tag\"]\n },\n \"content\": {\n \"type\": \"object\",\n \"properties\": {\n \"mainText\": {\n \"type\": \"string\",\n \"description\": \"Main article content\"\n },\n \"sections\": {\n \"type\": \"array\",\n \"items\": {\n \"type\": \"object\",\n \"properties\": {\n \"title\": {\n \"type\": \"string\",\n \"description\": \"Section title\"\n },\n \"text\": {\n \"type\": \"string\",\n \"description\": \"Section content\"\n },\n \"quote\": {\n \"type\": \"string\",\n \"description\": \"Blockquote text\"\n }\n },\n \"required\": [\"title\", \"text\", \"quote\"]\n }\n }\n },\n \"required\": [\"mainText\", \"sections\"]\n },\n \"hashtags\": {\n \"type\": \"array\",\n \"items\": {\n \"type\": \"string\"\n },\n \"description\": \"Article hashtags\"\n }\n }\n }\n }\n}" }, "id": "78aeb20b-c679-444f-b200-960b0758047d", "name": "Structured Output Parser1", "type": "@n8n/n8n-nodes-langchain.outputParserStructured", "position": [ 320, 900 ], "typeVersion": 1.2 }, { "parameters": { "options": { "frequencyPenalty": 0, "maxTokens": -1, "responseFormat": "json_object", "presencePenalty": 0, "temperature": 0, "timeout": 60000, "maxRetries": 2, "topP": 1 } }, "id": "0568f2ca-838c-43c0-bd73-130aa86e3902", "name": "gpt-4o-mini2", "type": "@n8n/n8n-nodes-langchain.lmChatOpenAi", "position": [ 100, 900 ], "typeVersion": 1, "credentials": { "openAiApi": { "id": "qKk5wVYPDLRt63R9", "name": "OpenAi account" } } }, { "parameters": { "chatId": "={{ $json.telegram_chat_id }}", "text": "={{ $('Perplexity Topic Agent').item.json.output.slice(0, 300) }}", "additionalFields": { "appendAttribution": false, "parse_mode": "HTML" } }, "id": "e5521fc8-53fa-4cda-9adc-055fd6babe26", "name": "Telegram", "type": "n8n-nodes-base.telegram", "position": [ 2000, 80 ], "typeVersion": 1.2, "webhookId": "66a380b0-fc47-4922-a082-fe2aba045638", "credentials": { "telegramApi": { "id": "ANYJVKBZhKBjOxLX", "name": "perplexity_researchbot" } }, "disabled": true }, { "parameters": { "options": { "frequencyPenalty": 0, "maxTokens": -1, "responseFormat": "text", "presencePenalty": 0, "temperature": 0, "timeout": 60000, "maxRetries": 2, "topP": 1 } }, "id": "3eda42fe-c377-49b8-8fc2-7dad94333a77", "name": "gpt-4o-mini5", "type": "@n8n/n8n-nodes-langchain.lmChatOpenAi", "position": [ 1940, 900 ], "typeVersion": 1, "credentials": { "openAiApi": { "id": "qKk5wVYPDLRt63R9", "name": "OpenAi account" } } }, { "parameters": { "promptType": "define", "text": "=Create a modern, responsive single-line HTML document. Convert any markdown to Tailwind CSS classes. Replace markdown lists with proper HTML list elements. Remove all newline characters while preserving tags in content. Enhance the layout with Tailwind CSS cards where appropriate. Use the following base structure, but improve the styling and responsiveness:\n\n\n\n\n
\n \n \nTime Posted: Just now
\nAuthor: AI Research Team
\nTag: AI Models
\nDeepSeek V3 is a state-of-the-art AI model that leverages\n advanced architectures and techniques to deliver high performance across various applications.\n This overview covers its key concepts, practical applications, advantages, limitations, and best\n practices for implementation.
\n1. Mixture-of-Experts (MoE) Architecture: DeepSeek V3\n employs a Mixture-of-Experts (MoE) architecture, which consists of multiple neural networks,\n each optimized for different tasks. This architecture allows for efficient processing by\n activating only a portion of the network for each task, reducing hardware costs.
\n2. Parameters: The model boasts a total of 671\n billion\n parameters, with 37 billion active parameters for each token during processing. The addition\n of\n the Multi-Token Prediction (MTP) module increases the total parameters to 685 billion,\n making it\n significantly larger than other models like Meta's Llama 3.1 (405B).
\n3. Multi-head Latent Attention (MLA): DeepSeek V3\n uses\n Multi-head Latent Attention (MLA) to extract key details from text multiple times, improving\n its\n accuracy.
\n4. Multi-Token Prediction (MTP): The model utilizes\n Multi-Token Prediction to generate several tokens at once, speeding up inference and\n enabling\n speculative decoding.
\n\n DeepSeek V3 employs a Mixture-of-Experts architecture for efficient processing.\n
\n DeepSeek V3 democratizes AI access for smaller organizations.\n
\n DeepSeek V3 processes information at 60 tokens per second.\n
\n Deployment of DeepSeek V3 may be complex for small companies.\n
\n Engage with the open-source community for better implementation.\n
Hashtags: #DeepSeekV3 #AI #MachineLearning #OpenSource
\n