import copy import json import pytest from langflow.graph import Graph from langflow.graph.graph.utils import ( find_last_node, process_flow, set_new_target_handle, ungroup_node, update_source_handle, update_target_handle, update_template, ) from langflow.graph.vertex.base import Vertex from langflow.initial_setup.setup import load_starter_projects # Test cases for the graph module # now we have three types of graph: # BASIC_EXAMPLE_PATH, COMPLEX_EXAMPLE_PATH, OPENAPI_EXAMPLE_PATH @pytest.fixture def sample_template(): return { "field1": {"proxy": {"field": "some_field", "id": "node1"}}, "field2": {"proxy": {"field": "other_field", "id": "node2"}}, } @pytest.fixture def sample_nodes(): return [ { "id": "node1", "data": {"node": {"template": {"some_field": {"show": True, "advanced": False, "name": "Name1"}}}}, }, { "id": "node2", "data": { "node": { "template": { "other_field": { "show": False, "advanced": True, "display_name": "DisplayName2", } } } }, }, { "id": "node3", "data": {"node": {"template": {"unrelated_field": {"show": True, "advanced": True}}}}, }, ] def get_node_by_type(graph, node_type: type[Vertex]) -> Vertex | None: """Get a node by type.""" return next((node for node in graph.vertices if isinstance(node, node_type)), None) def test_invalid_node_types(): graph_data = { "nodes": [ { "id": "1", "data": { "node": { "base_classes": ["BaseClass"], "template": { "_type": "InvalidNodeType", }, }, }, }, ], "edges": [], } g = Graph() with pytest.raises(KeyError): g.add_nodes_and_edges(graph_data["nodes"], graph_data["edges"]) def test_find_last_node(grouped_chat_json_flow): grouped_chat_data = json.loads(grouped_chat_json_flow).get("data") nodes, edges = grouped_chat_data["nodes"], grouped_chat_data["edges"] last_node = find_last_node(nodes, edges) assert last_node is not None # Replace with the actual expected value assert last_node["id"] == "LLMChain-pimAb" # Replace with the actual expected value def test_ungroup_node(grouped_chat_json_flow): grouped_chat_data = json.loads(grouped_chat_json_flow).get("data") group_node = grouped_chat_data["nodes"][2] # Assuming the first node is a group node base_flow = copy.deepcopy(grouped_chat_data) ungroup_node(group_node["data"], base_flow) # after ungroup_node is called, the base_flow and grouped_chat_data should be different assert base_flow != grouped_chat_data # assert node 2 is not a group node anymore assert base_flow["nodes"][2]["data"]["node"].get("flow") is None # assert the edges are updated assert len(base_flow["edges"]) > len(grouped_chat_data["edges"]) assert base_flow["edges"][0]["source"] == "ConversationBufferMemory-kUMif" assert base_flow["edges"][0]["target"] == "LLMChain-2P369" assert base_flow["edges"][1]["source"] == "PromptTemplate-Wjk4g" assert base_flow["edges"][1]["target"] == "LLMChain-2P369" assert base_flow["edges"][2]["source"] == "ChatOpenAI-rUJ1b" assert base_flow["edges"][2]["target"] == "LLMChain-2P369" def test_process_flow(grouped_chat_json_flow): grouped_chat_data = json.loads(grouped_chat_json_flow).get("data") processed_flow = process_flow(grouped_chat_data) assert processed_flow is not None assert isinstance(processed_flow, dict) assert "nodes" in processed_flow assert "edges" in processed_flow def test_process_flow_one_group(one_grouped_chat_json_flow): grouped_chat_data = json.loads(one_grouped_chat_json_flow).get("data") # There should be only one node assert len(grouped_chat_data["nodes"]) == 1 # Get the node, it should be a group node group_node = grouped_chat_data["nodes"][0] node_data = group_node["data"]["node"] assert node_data.get("flow") is not None template_data = node_data["template"] assert any("openai_api_key" in key for key in template_data) # Get the openai_api_key dict openai_api_key = next( (template_data[key] for key in template_data if "openai_api_key" in key), None, ) assert openai_api_key is not None assert openai_api_key["value"] == "test" processed_flow = process_flow(grouped_chat_data) assert processed_flow is not None assert isinstance(processed_flow, dict) assert "nodes" in processed_flow assert "edges" in processed_flow # Now get the node that has ChatOpenAI in its id chat_openai_node = next((node for node in processed_flow["nodes"] if "ChatOpenAI" in node["id"]), None) assert chat_openai_node is not None assert chat_openai_node["data"]["node"]["template"]["openai_api_key"]["value"] == "test" def test_process_flow_vector_store_grouped(vector_store_grouped_json_flow): grouped_chat_data = json.loads(vector_store_grouped_json_flow).get("data") nodes = grouped_chat_data["nodes"] assert len(nodes) == 4 # There are two group nodes in this flow # One of them is inside the other totalling 7 nodes # 4 nodes grouped, one of these turns into 1 normal node and 1 group node # This group node has 2 nodes inside it processed_flow = process_flow(grouped_chat_data) assert processed_flow is not None processed_nodes = processed_flow["nodes"] assert len(processed_nodes) == 7 assert isinstance(processed_flow, dict) assert "nodes" in processed_flow assert "edges" in processed_flow edges = processed_flow["edges"] # Expected keywords in source and target fields expected_keywords = [ {"source": "VectorStoreInfo", "target": "VectorStoreAgent"}, {"source": "ChatOpenAI", "target": "VectorStoreAgent"}, {"source": "OpenAIEmbeddings", "target": "Chroma"}, {"source": "Chroma", "target": "VectorStoreInfo"}, {"source": "WebBaseLoader", "target": "RecursiveCharacterTextSplitter"}, {"source": "RecursiveCharacterTextSplitter", "target": "Chroma"}, ] for idx, expected_keyword in enumerate(expected_keywords): for key, value in expected_keyword.items(): assert ( value in edges[idx][key].split("-")[0] ), f"Edge {idx}, key {key} expected to contain {value} but got {edges[idx][key]}" def test_update_template(sample_template, sample_nodes): # Making a deep copy to keep original sample_nodes unchanged nodes_copy = copy.deepcopy(sample_nodes) update_template(sample_template, nodes_copy) # Now, validate the updates. node1_updated = next((n for n in nodes_copy if n["id"] == "node1"), None) node2_updated = next((n for n in nodes_copy if n["id"] == "node2"), None) node3_updated = next((n for n in nodes_copy if n["id"] == "node3"), None) assert node1_updated["data"]["node"]["template"]["some_field"]["show"] is True assert node1_updated["data"]["node"]["template"]["some_field"]["advanced"] is False assert node1_updated["data"]["node"]["template"]["some_field"]["display_name"] == "Name1" assert node2_updated["data"]["node"]["template"]["other_field"]["show"] is False assert node2_updated["data"]["node"]["template"]["other_field"]["advanced"] is True assert node2_updated["data"]["node"]["template"]["other_field"]["display_name"] == "DisplayName2" # Ensure node3 remains unchanged assert node3_updated == sample_nodes[2] # Test `update_target_handle` def test_update_target_handle_proxy(): new_edge = { "data": { "targetHandle": { "type": "some_type", "proxy": {"id": "some_id", "field": ""}, } } } g_nodes = [{"id": "some_id", "data": {"node": {"flow": None}}}] updated_edge = update_target_handle(new_edge, g_nodes) assert updated_edge["data"]["targetHandle"] == new_edge["data"]["targetHandle"] # Test `set_new_target_handle` def test_set_new_target_handle(): proxy_id = "proxy_id" new_edge = {"target": None, "data": {"targetHandle": {}}} target_handle = {"type": "type_1", "proxy": {"field": "field_1"}} node = { "data": { "node": { "flow": True, "template": {"field_1": {"proxy": {"field": "new_field", "id": "new_id"}}}, } } } set_new_target_handle(proxy_id, new_edge, target_handle, node) assert new_edge["target"] == "proxy_id" assert new_edge["data"]["targetHandle"]["fieldName"] == "field_1" assert new_edge["data"]["targetHandle"]["proxy"] == { "field": "new_field", "id": "new_id", } # Test `update_source_handle` def test_update_source_handle(): new_edge = {"source": None, "data": {"sourceHandle": {"id": None}}} flow_data = { "nodes": [{"id": "some_node"}, {"id": "last_node"}], "edges": [{"source": "some_node"}], } updated_edge = update_source_handle(new_edge, flow_data["nodes"], flow_data["edges"]) assert updated_edge["source"] == "last_node" assert updated_edge["data"]["sourceHandle"]["id"] == "last_node" def test_serialize_graph(): starter_projects = load_starter_projects() data = starter_projects[0][1]["data"] graph = Graph.from_payload(data) assert isinstance(graph, Graph) serialized = graph.dumps() assert serialized is not None assert isinstance(serialized, str) assert len(serialized) > 0