ToolHub is simple library to help your LLMs take API-based actions. This project includes an initial set of tools we provide on RapidAPI, which has over 10,000 APIs that can be accessed, and an interface to support any OpenAPI spec, as well as arbitrary Python functions that can be implemented and called locally.
The core value prop is:
- Ability to call a large number of APIs using RapidAPI; and easily integrate other APIs that have OpenAPI specs - with robust error handling / communication that allows the LLM to retry.
- Easily integrate your own Python or HTTP backends (future support for GraphQL, gRPC, Thrift, etc.).
- Easily switch LLM backends, from OpenAI chat to assistant (future support for Google Gemini, Anthropic Claude, etc.).
- Customizable Auth
Why did we build this instead of just using one of the following?
- Langchain: requires installation of Python clients (supports OpenAPI but not RapidAPI), not very easy to use.
- LlamaHub: fairly broad support for document readers; does not extend well to reading dynamic content, or writes.
- ToolBlench: RapidAPI-only, not very easy to use, not regularly updated or maintained.
- ChatGPT plugins: plugins are OpenAI-specific, and furthermore are ChatGPT-specific - they are not available to all users via the chat completions or assistant APIs.
- Use the Alpha Vantage API to access stock prices (https://rapidapi.com/danilo.nobrega88/api/alpha-vantage12)
- Convert currencies (https://rapidapi.com/principalapis/api/currency-conversion-and-exchange-rates)
- Get the latest tax rates (https://rapidapi.com/davidrates/api/taxrates-io/)
- Use the business data API to search for target customers (https://rapidapi.com/letscrape-6bRBa3QguO5/api/local-business-data/)
- Connect to Crunchbase for more research on prospects
- Validate email addresses: https://rapidapi.com/mr_admin/api/email-verifier/
- Use a local ‘random’ syscall to enable your LLM to access random numbers
- Use the Google Translate API to get translations (https://rapidapi.com/googlecloud/api/google-translate1/)
Toolhub can be installed via pip:
pip install toolhub==0.1.0
The core product of ToolHub is toolhub.lib.hub::Hub
, which manages a set of available tools.
The key functionalities of a hub are:
- Prepare a list of the available tools - in a format the LLM API can consume directly.
- Call one or more tools with arguments, and return the output - both input and output in a format the LLM API can feed/consume directly.
A hub can be constructed with a standard or curated set of tools using a toolhub.lib.registry::Registry
. At the moment, the standard tools include 10k+ RapidAPI endpoints, OpenAPI endpoints for select APIs, and select Python functions. A registry can also be constructed simply with custom tools.
At the moment, ToolHub implements hubs for the OpenAI chat completions and assistant APIs (support for other LLM APIs/formats coming soon). Examples of how to a use hub are in toolhub/demo/openai_chat.py
and toolhub/demo/openai_assistant.py
.
- Signup for an account at rapidapi.com, create an application (make sure to create a new App, use RapidAPI type), and get a key for it
- For any APIs that you wish to use, from the API’s pricing page in RapidAPI, click Subscribe
- Create a registry object with the set of tools that you want to support. This can be filtered on the URL of specific endpoints, or APIs in RapidAPI. Specify the RapidAPI key in the
AuthContext
used to initializeHub
:AuthContext(rapidapi=RapidApiAuthContext(rapidapi_key=<your key>))
from toolhub.lib import auth
from toolhub.lib import registry
from toolhub.integrations.rapidapi import provider as rapidapi_provider
from toolhub.openai import openai_assistant_hub
# A collection of the tools that we want the registry to support.
# For example, this registry supports:
# 1. All Alpha Vantage endpoints, and
# 2. The Yelp business search endpoint.
registry_ = registry.Registry([
rapidapi_provider.Provider.standard(
filter_rapidapi_api_hostnames=["https://alpha-vantage12.p.rapidapi.com/"],
filter_rapidapi_endpoint_urls=["[https://yelp-reviews.p.rapidapi.com/business-search](https://yelp-reviews.p.rapidapi.com/business-search)"],
)
])
auth_ctx = auth.AuthContext(rapidapi=auth.RapidApiAuthContext(rapidapi_key=<YOUR_KEY>))
hub_ = openai_assistant_hub.OpenAIAssistantHub(
registry_=registry_,
auth_ctx=auth_ctx,
)
Simply initialize the registry with the tools to be used and pass to the hub. Random is the only one supported today.
from toolhub.lib import auth
from toolhub.lib import registry
from toolhub.openai import openai_assistant_hub
from toolhub.standard_providers import random_provider
hub_ = openai_assistant_hub.OpenAIAssistantHub(
registry_=registry.Registry([random_provider.Provider()]),
auth_ctx=auth.AuthContext(),
)
For now, our OSS implementation only supports crunchbase as an example. To use crunchbase,
- Generate a Crunchbase API key at https://data.crunchbase.com/docs/crunchbase-basic-getting-started#generating-a-basic-api-key.
- Specify the Crunchbase key in the
AuthContext
used to initializeHub
:
from toolhub.lib import auth
from toolhub.lib import registry
from toolhub.integrations.openapi import provider as openapi_provider
from toolhub.openai import openai_assistant_hub
auth_ctx = auth.AuthContext(
openapi=auth.OpenApiAuthContext(
api_to_headers={"crunchbase": {"X-cb-user-key": <crunchbase_key>}}
),
)
hub_ = openai_assistant_hub.OpenAIAssistantHub(
registry_=registry.Registry(
[openapi_provider.Provider.standard()],
filter_collections='crunchbase',
),
auth_ctx=auth_ctx,
)
Note: the OpenAPI AuthContext above is provided by default in toolhub/demo/utils.py
.
From the command line:
export DYNACONF_OPENAI__API_KEY=<your OpenAI key> # or specify in toolhub/settings.yml:openai.api_key
# Make sure to specify your RapidAPI key (from initializing the hub above) in demo/utils.py
# Convert currencies using RapidAPI; use OpenAI assistant.
python3 toolhub/demo/openai_assistant.py \
--task="Can you convert 100 MXN to USD?" \
--rapidapi_api_hostnames="https://currency-converter18.p.rapidapi.com"
# Generate a random string using Python; use OpenAI chat completions.
# NOTE: this demonstration module supports only a single tool call.
python3 toolhub/demo/openai_chat.py \
--task="Could you generate a random password for me of 15 characters?" \
--collections="random"
As a Python library (refer to toolhub/demo/openai_assistant.py):
# Initialize the auth_ctx as explained above, e.g.
auth_ctx = auth.StandardAuthContext(
openapi=auth.OpenApiAuthContext(
api_to_headers={"crunchbase": {"X-cb-user-key": <crunchbase_key>}}
),
rapidapi=auth.RapidApiAuthContext(
rapidapi_key=<rapidapi_key>,
host_to_headers=None,
),
)
# Initialize a registry as explained above, e.g. with multiple providers.
registry_ = registry.Registry(
[
random_provider.Provider(),
openapi_provider.Provider.standard(),
rapidapi_provider.Provider.standard(
filter_rapidapi_api_hostnames=["https://alpha-vantage12.p.rapidapi.com/"],
filter_rapidapi_endpoint_urls=["[https://yelp-reviews.p.rapidapi.com/business-search](https://yelp-reviews.p.rapidapi.com/business-search)"],
),
],
filter_collections=[
"random",
"crunchbase",
"Financial.currency_converter_v2"
],
)
# Set up the Hub
hub = openai_assistant_hub.OpenAIAssistantHub(
registry_=registry_, auth_ctx=auth_ctx
)
# Initialize the Open API client and run
client = openai.OpenAI(api_key=settings.openai.api_key)
assistant = client.beta.assistants.create(
name="ToolHub assistant",
instructions="You are a helpful assistant who uses tools to perform tasks.",
tools=hub.tools_spec(), # OpenAI currently supports upto 128 tools
model="gpt-4-1106-preview",
)
run = client.beta.threads.runs.create(
thread_id=thread.id,
assistant_id=assistant.id,
)
...
# Set up any application logic here
...
if (
run.status == _REQUIRES_ACTION_RUN_STATUS
and (tool_calls := run.required_action.submit_tool_outputs.tool_calls)
):
tool_outputs = []
for tool_call_id, result in hub.call_tools({
tc.id: tc for tc in tool_calls
}).items():
if isinstance(result, hub.ToolCallErrors):
errors_fmt = "\\n".join(str(e) for e in result.errors)
tool_outputs.append(
openai_assistant_hub.ToolOutput(
tool_call_id=tool_call_id,
output=f"Failure:\\n{errors_fmt}",
)
)
else:
tool_outputs.append(result)
client.beta.threads.runs.submit_tool_outputs(
thread_id=thread.id,
run_id=run_id,
tool_outputs=tool_outputs,
)
- Try to restrict the number of functions that you’re providing - adding too many can overwhelm the LLM’s context, and lead to poor results.
- Filter to only the collections/functions you’d like to use.
- Tools which provide a very large amount of output do not work well, since the model is unable to comprehend the results.
- Calling sequences of events can often fail - the model is sometime not able to perform planning without careful prompting.
- Currently, only RapidAPI GETs are supported - we are actively working on a framework for POSTs as well
Want to overcome these limitations? Check out our premium offering at https://www.bloomerai.com/
Add an OpenAPI API
We support adding any API defined in the OpenAPI format. Follow these instructions, and please do contribute back! Refer to demo/staockbot.py for an example
-
Identify and download the OpenAPI JSON schema to build against.
-
Define a Python loader for this API - you'll need the OpenAPI schema and the base URL. The standard API loader in toolhub.integrations.openapi should work in most cases.
- [if you need customization] In the Python loader module, implement a custom
MyParser(parser.Parser)
withfilter_endpoint
(to filter out certain functions) and/ormap_parameter
(to change certain parameters, e.g. updating descriptions).
- [if you need customization] In the Python loader module, implement a custom
-
Construct a registry with this loader, and any other loaders you need. You may optionally filter the functions to a subset you'd like to support
SCHEMA_PATH = pathlib.Path( os.path.join(os.path.dirname(__file__), "alpaca.yaml") ) BASE_URL = 'https://paper-api.alpaca.markets' alpaca_api_loader = openapi_provider.standard_api_loader( api="alpaca", schema_path=SCHEMA_PATH, request_body_descriptions_path=None, base_url=BASE_URL, ) registry_ = registry.Registry( [ openapi_provider.Provider( api_loaders=[alpaca_api_loader], filter_function_names=["alpaca_v2_orders_post"] ), rapidapi_provider.Provider.standard( filter_rapidapi_endpoint_urls=[ # Any other RapidAPI endpoints you want to support "https://alpha-vantage12.p.rapidapi.com/", "https://yelp-reviews.p.rapidapi.com/business-search", ],
-
Construct an authcontext with the correct credentials
auth_ctx = auth.StandardAuthContext( rapidapi=auth.RapidApiAuthContext( rapidapi_key=_RAPIDAPI_KEY, ), openapi=auth.OpenApiAuthContext( api_to_headers={ "alpaca": { "APCA-API-KEY-ID": _ALPACA_KEY_ID, "APCA-API-SECRET-KEY": _ALPACA_SECRET_KEY, } }, ), )
-
Use this auth context with an agent and pass in any queries
agent = openai_assistant.Agent( registry_=registry_, openai_client=openai.OpenAI(api_key=_OPENAI_KEY) ) agent(auth_ctx, query)
This will enable your LLM to access any APIs, including internal ones.
Customize auth
You can pass custom additional authentication headers via the AuthContext
@dataclasses.dataclass
class OpenApiAuthContext:
api_to_headers: dict[str, dict[str, str]] | None
@dataclasses.dataclass
class RapidApiAuthContext:
rapidapi_key: str
host_to_headers: dict[str, dict[str, str]] | None
@dataclasses.dataclass
class AuthContext:
openapi: OpenApiAuthContext | None
rapidapi: RapidApiAuthContext | None
Note: the AuthContext
can be also configured without code via dynaconf (e.g. with toolhub/settings.yml
) - see AuthContext.from_settings()
.
We welcome contributions of new functions! Our goal is to help everyone accelerate their AI development timeline. If you found this project helpful for an API that was not covered, pleas feel free to submit a PR and give back!
Questions? Concerns? Join our Discord at https://discord.gg/jN5ePfvV