feat(editor): Auto-add LLM chain for new LLM nodes on empty canvas (#10245)
Co-authored-by: JP van Oosten <jp@n8n.io>
This commit is contained in:
@@ -74,7 +74,7 @@ export class LmChatAnthropic implements INodeType {
|
||||
codex: {
|
||||
categories: ['AI'],
|
||||
subcategories: {
|
||||
AI: ['Language Models'],
|
||||
AI: ['Language Models', 'Root Nodes'],
|
||||
'Language Models': ['Chat Models (Recommended)'],
|
||||
},
|
||||
resources: {
|
||||
|
||||
@@ -28,7 +28,7 @@ export class LmChatOllama implements INodeType {
|
||||
codex: {
|
||||
categories: ['AI'],
|
||||
subcategories: {
|
||||
AI: ['Language Models'],
|
||||
AI: ['Language Models', 'Root Nodes'],
|
||||
'Language Models': ['Chat Models (Recommended)'],
|
||||
},
|
||||
resources: {
|
||||
|
||||
@@ -26,7 +26,7 @@ export class LmChatOpenAi implements INodeType {
|
||||
codex: {
|
||||
categories: ['AI'],
|
||||
subcategories: {
|
||||
AI: ['Language Models'],
|
||||
AI: ['Language Models', 'Root Nodes'],
|
||||
'Language Models': ['Chat Models (Recommended)'],
|
||||
},
|
||||
resources: {
|
||||
|
||||
@@ -26,7 +26,7 @@ export class LmCohere implements INodeType {
|
||||
codex: {
|
||||
categories: ['AI'],
|
||||
subcategories: {
|
||||
AI: ['Language Models'],
|
||||
AI: ['Language Models', 'Root Nodes'],
|
||||
'Language Models': ['Text Completion Models'],
|
||||
},
|
||||
resources: {
|
||||
|
||||
@@ -27,7 +27,7 @@ export class LmOllama implements INodeType {
|
||||
codex: {
|
||||
categories: ['AI'],
|
||||
subcategories: {
|
||||
AI: ['Language Models'],
|
||||
AI: ['Language Models', 'Root Nodes'],
|
||||
'Language Models': ['Text Completion Models'],
|
||||
},
|
||||
resources: {
|
||||
|
||||
@@ -38,7 +38,7 @@ export class LmOpenAi implements INodeType {
|
||||
codex: {
|
||||
categories: ['AI'],
|
||||
subcategories: {
|
||||
AI: ['Language Models'],
|
||||
AI: ['Language Models', 'Root Nodes'],
|
||||
'Language Models': ['Text Completion Models'],
|
||||
},
|
||||
resources: {
|
||||
|
||||
@@ -26,7 +26,7 @@ export class LmOpenHuggingFaceInference implements INodeType {
|
||||
codex: {
|
||||
categories: ['AI'],
|
||||
subcategories: {
|
||||
AI: ['Language Models'],
|
||||
AI: ['Language Models', 'Root Nodes'],
|
||||
'Language Models': ['Text Completion Models'],
|
||||
},
|
||||
resources: {
|
||||
|
||||
@@ -29,7 +29,7 @@ export class LmChatAwsBedrock implements INodeType {
|
||||
codex: {
|
||||
categories: ['AI'],
|
||||
subcategories: {
|
||||
AI: ['Language Models'],
|
||||
AI: ['Language Models', 'Root Nodes'],
|
||||
'Language Models': ['Chat Models (Recommended)'],
|
||||
},
|
||||
resources: {
|
||||
|
||||
@@ -27,7 +27,7 @@ export class LmChatAzureOpenAi implements INodeType {
|
||||
codex: {
|
||||
categories: ['AI'],
|
||||
subcategories: {
|
||||
AI: ['Language Models'],
|
||||
AI: ['Language Models', 'Root Nodes'],
|
||||
'Language Models': ['Chat Models (Recommended)'],
|
||||
},
|
||||
resources: {
|
||||
|
||||
@@ -27,7 +27,7 @@ export class LmChatGoogleGemini implements INodeType {
|
||||
codex: {
|
||||
categories: ['AI'],
|
||||
subcategories: {
|
||||
AI: ['Language Models'],
|
||||
AI: ['Language Models', 'Root Nodes'],
|
||||
'Language Models': ['Chat Models (Recommended)'],
|
||||
},
|
||||
resources: {
|
||||
|
||||
@@ -25,7 +25,7 @@ export class LmChatGooglePalm implements INodeType {
|
||||
codex: {
|
||||
categories: ['AI'],
|
||||
subcategories: {
|
||||
AI: ['Language Models'],
|
||||
AI: ['Language Models', 'Root Nodes'],
|
||||
'Language Models': ['Chat Models (Recommended)'],
|
||||
},
|
||||
resources: {
|
||||
|
||||
@@ -32,7 +32,7 @@ export class LmChatGoogleVertex implements INodeType {
|
||||
codex: {
|
||||
categories: ['AI'],
|
||||
subcategories: {
|
||||
AI: ['Language Models'],
|
||||
AI: ['Language Models', 'Root Nodes'],
|
||||
'Language Models': ['Chat Models (Recommended)'],
|
||||
},
|
||||
resources: {
|
||||
|
||||
@@ -26,7 +26,7 @@ export class LmChatGroq implements INodeType {
|
||||
codex: {
|
||||
categories: ['AI'],
|
||||
subcategories: {
|
||||
AI: ['Language Models'],
|
||||
AI: ['Language Models', 'Root Nodes'],
|
||||
'Language Models': ['Chat Models (Recommended)'],
|
||||
},
|
||||
resources: {
|
||||
|
||||
@@ -27,7 +27,7 @@ export class LmChatMistralCloud implements INodeType {
|
||||
codex: {
|
||||
categories: ['AI'],
|
||||
subcategories: {
|
||||
AI: ['Language Models'],
|
||||
AI: ['Language Models', 'Root Nodes'],
|
||||
'Language Models': ['Chat Models (Recommended)'],
|
||||
},
|
||||
resources: {
|
||||
|
||||
@@ -25,7 +25,7 @@ export class LmGooglePalm implements INodeType {
|
||||
codex: {
|
||||
categories: ['AI'],
|
||||
subcategories: {
|
||||
AI: ['Language Models'],
|
||||
AI: ['Language Models', 'Root Nodes'],
|
||||
'Language Models': ['Text Completion Models'],
|
||||
},
|
||||
resources: {
|
||||
|
||||
@@ -1808,8 +1808,8 @@ export type AddedNode = {
|
||||
} & Partial<INodeUi>;
|
||||
|
||||
export type AddedNodeConnection = {
|
||||
from: { nodeIndex: number; outputIndex?: number };
|
||||
to: { nodeIndex: number; inputIndex?: number };
|
||||
from: { nodeIndex: number; outputIndex?: number; type?: NodeConnectionType };
|
||||
to: { nodeIndex: number; inputIndex?: number; type?: NodeConnectionType };
|
||||
};
|
||||
|
||||
export type AddedNodesAndConnections = {
|
||||
|
||||
@@ -1,5 +1,10 @@
|
||||
import { computed } from 'vue';
|
||||
import type { IDataObject, INodeParameters } from 'n8n-workflow';
|
||||
import {
|
||||
CHAIN_LLM_LANGCHAIN_NODE_TYPE,
|
||||
NodeConnectionType,
|
||||
type IDataObject,
|
||||
type INodeParameters,
|
||||
} from 'n8n-workflow';
|
||||
import type {
|
||||
ActionTypeDescription,
|
||||
AddedNode,
|
||||
@@ -11,6 +16,7 @@ import type {
|
||||
} from '@/Interface';
|
||||
import {
|
||||
AGENT_NODE_TYPE,
|
||||
AI_CATEGORY_LANGUAGE_MODELS,
|
||||
BASIC_CHAIN_NODE_TYPE,
|
||||
CHAT_TRIGGER_NODE_TYPE,
|
||||
MANUAL_CHAT_TRIGGER_NODE_TYPE,
|
||||
@@ -37,11 +43,12 @@ import { useExternalHooks } from '@/composables/useExternalHooks';
|
||||
|
||||
import { sortNodeCreateElements, transformNodeType } from '../utils';
|
||||
import { useI18n } from '@/composables/useI18n';
|
||||
import { useCanvasStore } from '@/stores/canvas.store';
|
||||
|
||||
export const useActions = () => {
|
||||
const nodeCreatorStore = useNodeCreatorStore();
|
||||
const nodeTypesStore = useNodeTypesStore();
|
||||
const i18n = useI18n();
|
||||
|
||||
const singleNodeOpenSources = [
|
||||
NODE_CREATOR_OPEN_SOURCES.PLUS_ENDPOINT,
|
||||
NODE_CREATOR_OPEN_SOURCES.NODE_CONNECTION_ACTION,
|
||||
@@ -216,6 +223,19 @@ export const useActions = () => {
|
||||
return isCompatibleNode && isChatTriggerMissing;
|
||||
}
|
||||
|
||||
// AI-226: Prepend LLM Chain node when adding a language model
|
||||
function shouldPrependLLMChain(addedNodes: AddedNode[]): boolean {
|
||||
const canvasHasAINodes = useCanvasStore().aiNodes.length > 0;
|
||||
if (canvasHasAINodes) return false;
|
||||
|
||||
return addedNodes.some((node) => {
|
||||
const nodeType = nodeTypesStore.getNodeType(node.type);
|
||||
return Object.keys(nodeType?.codex?.subcategories ?? {}).includes(
|
||||
AI_CATEGORY_LANGUAGE_MODELS,
|
||||
);
|
||||
});
|
||||
}
|
||||
|
||||
function getAddedNodesAndConnections(addedNodes: AddedNode[]): AddedNodesAndConnections {
|
||||
if (addedNodes.length === 0) {
|
||||
return { nodes: [], connections: [] };
|
||||
@@ -230,7 +250,14 @@ export const useActions = () => {
|
||||
nodeToAutoOpen.openDetail = true;
|
||||
}
|
||||
|
||||
if (shouldPrependChatTrigger(addedNodes)) {
|
||||
if (shouldPrependLLMChain(addedNodes) || shouldPrependChatTrigger(addedNodes)) {
|
||||
if (shouldPrependLLMChain(addedNodes)) {
|
||||
addedNodes.unshift({ type: CHAIN_LLM_LANGCHAIN_NODE_TYPE, isAutoAdd: true });
|
||||
connections.push({
|
||||
from: { nodeIndex: 2, type: NodeConnectionType.AiLanguageModel },
|
||||
to: { nodeIndex: 1 },
|
||||
});
|
||||
}
|
||||
addedNodes.unshift({ type: CHAT_TRIGGER_NODE_TYPE, isAutoAdd: true });
|
||||
connections.push({
|
||||
from: { nodeIndex: 0 },
|
||||
|
||||
@@ -610,7 +610,7 @@ export default defineComponent({
|
||||
return this.workflowsStore.getWorkflowExecution;
|
||||
},
|
||||
workflowRunning(): boolean {
|
||||
return this.uiStore.isActionActive['workflowRunning'];
|
||||
return this.uiStore.isActionActive.workflowRunning;
|
||||
},
|
||||
currentWorkflow(): string {
|
||||
return this.$route.params.name?.toString() || this.workflowsStore.workflowId;
|
||||
@@ -4428,7 +4428,7 @@ export default defineComponent({
|
||||
from.outputIndex ?? 0,
|
||||
toNode.name,
|
||||
to.inputIndex ?? 0,
|
||||
NodeConnectionType.Main,
|
||||
from.type ?? NodeConnectionType.Main,
|
||||
);
|
||||
}
|
||||
|
||||
@@ -4449,6 +4449,22 @@ export default defineComponent({
|
||||
});
|
||||
}
|
||||
|
||||
const lastNodeType = this.nodeTypesStore.getNodeType(lastAddedNode.type);
|
||||
const isSubNode = NodeHelpers.isSubNodeType(lastNodeType);
|
||||
|
||||
// When adding a sub-node and there's more than one node added at the time, it must mean that it's
|
||||
// connected to a root node, so we adjust the position of the sub-node to make it appear in the correct
|
||||
// in relation to the root node
|
||||
if (isSubNode && nodes.length > 1) {
|
||||
this.onMoveNode({
|
||||
nodeName: lastAddedNode.name,
|
||||
position: [
|
||||
lastAddedNode.position[0] - NodeViewUtils.NODE_SIZE * 2.5,
|
||||
lastAddedNode.position[1] + NodeViewUtils.NODE_SIZE * 1.5,
|
||||
],
|
||||
});
|
||||
}
|
||||
|
||||
this.nodeHelpers.addPinDataConnections(this.workflowsStore.pinnedWorkflowData);
|
||||
},
|
||||
|
||||
|
||||
Reference in New Issue
Block a user