The first AI agent that builds third-party integrations through reverse engineering platforms' internal APIs.
An AI agent that generates integration code by reverse-engineering platforms’ internal APIs.
You use create_har.py
to generate a file containing all browser network requests, a file with the cookies, and write a prompt describing the action triggered in the browser. The agent outputs runnable Python code that hits the platform’s internal endpoints to perform the desired action.
Let’s assume we want to download utility bills:
https://www.example.com/utility-bills?accountId=123&userId=456
accountId=123 userId=456
GET https://www.example.com/get_account_id
GET https://www.example.com/get_user_id
Set up your OpenAI API Keys and add the OPENAI_API_KEY
environment variable. (We recommend using an account with access to models that are at least as capable as OpenAI o1-mini. Models on par with OpenAI o1-preview are ideal.)
Install Python requirements via poetry:
poetry install
Open a poetry shell:
poetry shell
Register the Poetry virtual environment with Jupyter:
poetry run ipython kernel install --user --name=integuru
Run the following command to spawn a browser:
poetry run python create_har.py
Log into your platform and perform the desired action (such as downloading a utility bill).
Run Integuru:
poetry run integuru --prompt "download utility bills" --model gpt-4o
You can also run it via Jupyter Notebook main.ipynb
Recommended to use gpt-4o as the model for graph generation as it supports function calling. Integuru will automatically switch to o1-preview for code generation if available in the user’s OpenAI account. ⚠️ Note: o1-preview does not support function calls.
After setting up the project, you can use Integuru to analyze and reverse-engineer API requests for external platforms. Simply provide the appropriate .har file and a prompt describing the action that you want to trigger.
poetry run integuru --help
Usage: integuru [OPTIONS]
Options:
--model TEXT The LLM model to use (default is gpt-4o)
--prompt TEXT The prompt for the model [required]
--har-path TEXT The HAR file path (default is
./network_requests.har)
--cookie-path TEXT The cookie file path (default is
./cookies.json)
--max_steps INTEGER The max_steps (default is 20)
--input_variables <TEXT TEXT>...
Input variables in the format key value
--generate-code Whether to generate the full integration
code
--help Show this message and exit.
To run unit tests using pytest
, use the following command:
poetry run pytest
This repository includes a CI workflow using GitHub Actions. The workflow is defined in the .github/workflows/ci.yml
file and is triggered on each push and pull request to the main
branch. The workflow performs the following steps:
poetry
.pytest
.When the destination site uses two-factor authentication (2FA), the workflow remains the same. Ensure that you complete the 2FA process and obtain the cookies/auth tokens/session tokens after 2FA. These tokens will be used in the workflow.
Contributions to improve Integuru are welcome. Please feel free to submit issues or pull requests on the project’s repository.
Integuru is built by Integuru.ai. Besides our work on the agent, we take custom requests for new integrations or additional features for existing supported platforms. We also offer hosting and authentication services. If you have requests or want to work with us, reach out at [email protected].
We open-source unofficial APIs that we’ve built already. You can find them here.
Collected data is stored locally in the network_requests.har
and cookies.json
files.
The tool uses a cloud-based LLM (OpenAI’s GPT-4o and o1-preview models).
The LLM is not trained or improved by the usage of this tool.