Tools and External Integrations
Last updated
Last updated
In our workflow builder, there are several tool nodes specifically designed to handle tasks like web scraping, online searching, and making HTTP requests. These nodes offer flexibility in integrating external data sources, enabling workflows to retrieve, process, and utilize information from the web seamlessly.
The Tavily tool node is a specialized tool for web search, powered by the Tavily search service. This node allows you to execute search queries with customizable parameters, retrieve relevant web content, and integrate the results directly into your workflows.
Name: Customizable name for identifying this node.
Connection: The connection configuration for Tavily.
Description: Short description of the node’s functionality.
Optimized for Agents: When enabled, formats the output specifically for AI agents, structuring the results in a format that agents can easily parse, interpret, and display.
query: A string
containing the search query to execute.
content:
Standart (if Optimized for Agents disabled) - dict
output with the following fields:
result: Formatted string output of the search results, including source URL, title, content, and relevance score.
sources_with_url: List of sources with URLs in a clickable format, useful for quickly accessing original pages.
raw_response: Full raw JSON response from Tavily, useful for advanced data processing.
images: List of image URLs.
answer: Short answer generated by Tavily for the search query (if available).
query: Echo of the original search query.
response_time: The time taken for Tavily to respond to the query.
Example:
Optimized for Agents - formatted string
Example:
Type: Tavily
Name: Customizable name for identifying this connection.
API key: Your API key
To get API key please follow official Tavily documentation.
Add an Input node and specify query.
Drag a Tavily node into the workspace and connect it to the Input node. Set the desired configuration.
Attach a downstream node (e.g. Output) to handle the search content.
The ScaleSerp tool node enables users to perform web searches using the Scale SERP API. It is designed to retrieve search results across various search types, such as organic web search, news, images, and videos. This tool is ideal for workflows that need dynamic web search capabilities and flexible data retrieval.
Name: Customizable name for identifying this node.
Connection: The connection configuration for ScaleSerp.
Description: Short description of the node’s functionality.
Optimized for Agents: When enabled, formats the output specifically for AI agents, structuring the results in a format that agents can easily parse, interpret, and display.
query: A string
containing the search query to execute.
content:
Standart (if Optimized for Agents disabled) - dict
output with the following fields:
result: Formatted string output of the search results, including title, URL, and content snippets.
sources_with_url: List of sources in [Title: (URL)] format for quick access to original pages.
urls: List of URLs of the search results.
raw_response: Full raw JSON response from Scale SERP, useful for advanced data analysis and processing.
Example:
Optimized for Agents - formatted string
Example:
Type: ScaleSerp
Name: Customizable name for identifying this connection.
API key: Your API key
To get API key please follow official ScaleSerp documentation.
Add an Input node and specify query in input field.
Drag a ScaleSerp node into the workspace and connect it to the Input node. Set the desired configuration.
Attach a downstream node (e.g. Output) to handle the search content.
The ZenRows tool node provides a powerful solution for web scraping, allowing users to extract content from web pages using the ZenRows service and is ideal for integrating scraped content directly into workflows for further processing or analysis.
Name: Customizable name for identifying this node.
Connection: The connection configuration for Tavily.
Description: Short description of the node’s functionality.
URL: The URL of the page to scrape. This also can be dynamically provided in the Input.
Optimized for Agents: When enabled, formats the output specifically for AI agents, structuring the results in a format that agents can easily parse, interpret, and display.
url: A string
containing the URL of the page to scrape. If it missed, try to use URL from the configuration.
content:
Standart (if Optimized for Agents disabled) - dict
output with the following fields:
url: The URL of the scraped page.
content: The main content of the page in Markdown format.
Example:
Optimized for Agents - formatted string
Example:
Type: ZenRows
Name: Customizable name for identifying this connection.
API key: Your API key
To get API key please follow official ZenRows documentation.
Workflow with url specified in Input node
Add an Input node and specify url.
Drag a ZenRows node into the workspace and connect it to the Input node. Set the desired configuration.
Attach a downstream node (e.g. Output) to handle the search content.
Workflow with url specified in ZenRows node
Add an Input node.
Drag a ZenRows node into the workspace and connect it to the Input node. Set the desired configuration and URL.
Attach a downstream node (e.g. Output) to handle the search content.
The FireCrawl tool node is a web scraping tool powered by FireCrawl. It enables users to extract structured content from web pages. This tool is ideal for workflows requiring flexible web data extraction capabilities.
Name: Customizable name for identifying this node.
Connection: The connection configuration for Tavily.
Description: Short description of the node’s functionality.
URL: The URL of the page to scrape. This also can be dynamically provided in the Input.
Optimized for Agents: When enabled, formats the output specifically for AI agents, structuring the results in a format that agents can easily parse, interpret, and display.
url: A string
containing the URL of the page to scrape. If it missed, try to use URL from the configuration.
content:
Standart (if Optimized for Agents disabled) - dict
output with the following fields:
success: Boolean indicating whether the scraping was successful.
url: The URL of the scraped page.
markdown: Content in Markdown format (if applicable).
content: Main text content of the page.
html: Full HTML content if Include HTML was enabled.
raw_html: Raw HTML if extracted without additional processing.
metadata: Metadata information extracted from the page.
llm_extraction: Extracted content optimized for language model processing.
warning: Any warnings or messages returned during the scraping process.
Example:
Optimized for Agents - formatted string
Example:
Type: FireCrawl
Name: Customizable name for identifying this connection.
API key: Your API key.
To get API key please follow official FireCrawl documentation.
Workflow with url specified in Input node:
Add an Input node and specify url.
Drag a FireCrawl node into the workspace and connect it to the Input node. Set the desired configuration.
Attach a downstream node (e.g. Output) to handle the search content.
Workflow with url specified in FireCrawl node:
Add an Input node.
Drag a FireCrawl node into the workspace and connect it to the Input node. Set the desired configuration and URL.
Attach a downstream node (e.g. Output) to handle the search content.
The HTTP API Call node is a versatile tool for making HTTP API requests. This node allows for dynamic configuration of request parameters, headers, and data payload, making it suitable for interacting with various APIs. The node is flexible and can handle responses in JSON, text, or raw data format.
Name: Customizable name for identifying this node.
Connection: Connection to the HTTP service, specifying the base URL and HTTP method (GET, POST, etc.).
Description: Short description of the node’s functionality.
Response Type: Specifies the expected response format, with options:
json: Parses the response as JSON and returns it as a dict
.
text: Returns the response as a plain text string
.
raw: Returns the raw binary response (bytes
), useful for non-text responses (e.g., images).
Timeout: Sets the maximum time in seconds the node will wait for a response, with a default of 30 seconds.
Headers: Dictionary of HTTP headers to include in the request, such as Content-Type
.
Data: Dictionary containing the data to send as the request body, applicable mainly for POST, PUT, or PATCH requests.
Params: A dictionary of query parameters to append to the URL.
data: Dictionary containing the data to send as the request body, applicable mainly for POST, PUT, or PATCH requests.
headers: Dictionary containing additional headers to include in the request.
params: Dictionary of query parameters to append to the URL.
content: The main content of the response, returned in the format specified by the Response Type configuration
json: Returns the response as a dict
.
text: Returns the response as a string
.
raw: Returns the response as raw binary data (bytes
).
status_code: The HTTP status code returned by the API, indicating the success or failure of the request.
Example:
Type: Http
Name: Customizable name for identifying this connection.
URL: API URL (e.g. https://api.openai.com/v1/
for OpenAI).
Method: HTTP method (GET, POST, etc.).
Headers: Dictionary containing additional headers to include in the request.
Params: Dictionary of query parameters to append to the URL.
Data: Dictionary containing the data to send as the request body, applicable mainly for POST, PUT, or PATCH requests.
The HTTP API Call node allows flexibility in defining data, headers, and params through multiple sources with the specific order of priority:
Input: The highest priority and supplied dynamically during execution, e.g., from another node in the workflow. If data, headers, or params are specified here, they will override any values set in the node configuration or connection.
Node Configuration: Set directly in the node’s settings during workflow design. If no values are provided in the input, the node uses values defined directly in the node configuration.
Connection: Defined within the HTTP connection. Default values set in the connection (like base URL or default headers) are used only if neither the input data nor the node configuration specifies them. This is particularly useful for setting default headers, authentication tokens, or base parameters that apply to all requests made using the connection.
When merging values from multiple sources, the node combines dictionaries for headers and params across all levels. In case of conflicts, the priority order (Input > Node Configuration > Connection) determines which value to use.
Consider an HTTP API Request node configured as follows:
Connection:
Headers:
Params:
Node Configuration:
Headers:
Params:
Input (provided at runtime):
Headers:
Params:
Given this setup, the functionality will prioritize and combine the values as follows:
Headers:
Params:
Add an Input node and specify data, headers, and query parameters to provide for the request dynamically.
Drag a HTTP API Call node into the workspace and connect it to the Input node. Set the desired configuration.
Attach a downstream node (e.g. Output) to handle the API Call content and status_code.
The E2B Interpreter node enables users to interact with an E2B sandbox environment for executing Python code, running shell commands, and managing files within a secure, isolated environment. This node is ideal for workflows that require on-demand computation, filesystem operations, and API requests.
Currently, the E2B Interpreter node is optimized for use with Agents, offering full functionality for dynamic code execution, shell command processing, and file management. While it can be used in workflows, some configurations and advanced features are still under development for seamless integration in workflow settings.
Name: Customizable name for identifying this node.
Connection: Configuration for connecting to the E2B environment.
Description: Short description of the node’s functionality.
Optimized for Agents: When enabled, formats the output specifically for AI agents, structuring the results in a format that agents can easily parse, interpret, and display.
input: The main input field for specifying either a shell command, Python code, or package installations.
content:
Standart (if Optimized for Agents disabled) - dict
output with the following fields:
files_installation: Details about uploaded files.
packages_installation: Installed packages.
shell_command_execution: Shell command output.
code_execution: Python code execution result.
Example:
Optimized for Agents - formatted string
Example:
Type: e2b
Name: Customizable name for identifying this connection.
API key: Your API key.
To get API key please follow official E2B Interpreter documentation.