/toolhub

A collection of tools that can be used for LLM function calling

Primary LanguagePythonMIT LicenseMIT

Welcome to ToolHub!

ToolHub is simple library to help your LLMs take API-based actions. This project includes an initial set of tools we provide on RapidAPI, which has over 10,000 APIs that can be accessed, and an interface to support any OpenAPI spec, as well as arbitrary Python functions that can be implemented and called locally.

The core value prop is:

  • Ability to call a large number of APIs using RapidAPI; and easily integrate other APIs that have OpenAPI specs - with robust error handling / communication that allows the LLM to retry.
  • Easily integrate your own Python or HTTP backends (future support for GraphQL, gRPC, Thrift, etc.).
  • Easily switch LLM backends, from OpenAI chat to assistant (future support for Google Gemini, Anthropic Claude, etc.).
  • Customizable Auth

Why did we build this instead of just using one of the following?

  • Langchain: requires installation of Python clients (supports OpenAPI but not RapidAPI), not very easy to use.
  • LlamaHub: fairly broad support for document readers; does not extend well to reading dynamic content, or writes.
  • ToolBlench: RapidAPI-only, not very easy to use, not regularly updated or maintained.
  • ChatGPT plugins: plugins are OpenAI-specific, and furthermore are ChatGPT-specific - they are not available to all users via the chat completions or assistant APIs.

Example Use Cases

For Fintechs

For Marketing and Sales

For All Chat bots

Installation!

Toolhub can be installed via pip:

pip install toolhub==0.1.0

How to use

The core product of ToolHub is toolhub.lib.hub::Hub, which manages a set of available tools.

The key functionalities of a hub are:

  1. Prepare a list of the available tools - in a format the LLM API can consume directly.
  2. Call one or more tools with arguments, and return the output - both input and output in a format the LLM API can feed/consume directly.

A hub can be constructed with a standard or curated set of tools using a toolhub.lib.registry::Registry. At the moment, the standard tools include 10k+ RapidAPI endpoints, OpenAPI endpoints for select APIs, and select Python functions. A registry can also be constructed simply with custom tools.

At the moment, ToolHub implements hubs for the OpenAI chat completions and assistant APIs (support for other LLM APIs/formats coming soon). Examples of how to a use hub are in toolhub/demo/openai_chat.py and toolhub/demo/openai_assistant.py.

Initializing the Hub

RapidAPI

  1. Signup for an account at rapidapi.com, create an application (make sure to create a new App, use RapidAPI type), and get a key for it
  2. For any APIs that you wish to use, from the API’s pricing page in RapidAPI, click Subscribe
  3. Create a registry object with the set of tools that you want to support. This can be filtered on the URL of specific endpoints, or APIs in RapidAPI. Specify the RapidAPI key in the AuthContext used to initialize Hub: AuthContext(rapidapi=RapidApiAuthContext(rapidapi_key=<your key>))
from toolhub.lib import auth
from toolhub.lib import registry
from toolhub.integrations.rapidapi import provider as rapidapi_provider
from toolhub.openai import openai_assistant_hub

# A collection of the tools that we want the registry to support.
# For example, this registry supports:
# 1. All Alpha Vantage endpoints, and
# 2. The Yelp business search endpoint.
registry_ = registry.Registry([
    rapidapi_provider.Provider.standard(
        filter_rapidapi_api_hostnames=["https://alpha-vantage12.p.rapidapi.com/"],
        filter_rapidapi_endpoint_urls=["[https://yelp-reviews.p.rapidapi.com/business-search](https://yelp-reviews.p.rapidapi.com/business-search)"],
    )
])

auth_ctx = auth.AuthContext(rapidapi=auth.RapidApiAuthContext(rapidapi_key=<YOUR_KEY>))

hub_ = openai_assistant_hub.OpenAIAssistantHub(
    registry_=registry_,
    auth_ctx=auth_ctx,
)

Local Tools

Simply initialize the registry with the tools to be used and pass to the hub. Random is the only one supported today.

from toolhub.lib import auth
from toolhub.lib import registry
from toolhub.openai import openai_assistant_hub
from toolhub.standard_providers import random_provider

hub_ = openai_assistant_hub.OpenAIAssistantHub(
    registry_=registry.Registry([random_provider.Provider()]),
    auth_ctx=auth.AuthContext(),
)

OpenAPI

For now, our OSS implementation only supports crunchbase as an example. To use crunchbase,

  1. Generate a Crunchbase API key at https://data.crunchbase.com/docs/crunchbase-basic-getting-started#generating-a-basic-api-key.
  2. Specify the Crunchbase key in the AuthContext used to initialize Hub:
from toolhub.lib import auth
from toolhub.lib import registry
from toolhub.integrations.openapi import provider as openapi_provider
from toolhub.openai import openai_assistant_hub

auth_ctx = auth.AuthContext(
    openapi=auth.OpenApiAuthContext(
        api_to_headers={"crunchbase": {"X-cb-user-key": <crunchbase_key>}}
    ),
)

hub_ = openai_assistant_hub.OpenAIAssistantHub(
    registry_=registry.Registry(
        [openapi_provider.Provider.standard()],
        filter_collections='crunchbase',
    ),
    auth_ctx=auth_ctx,
)

Note: the OpenAPI AuthContext above is provided by default in toolhub/demo/utils.py.

Using the hub

From the command line:

export DYNACONF_OPENAI__API_KEY=<your OpenAI key> # or specify in toolhub/settings.yml:openai.api_key

# Make sure to specify your RapidAPI key (from initializing the hub above) in demo/utils.py

# Convert currencies using RapidAPI; use OpenAI assistant.
python3 toolhub/demo/openai_assistant.py \
--task="Can you convert 100 MXN to USD?" \
--rapidapi_api_hostnames="https://currency-converter18.p.rapidapi.com"

# Generate a random string using Python; use OpenAI chat completions.
# NOTE: this demonstration module supports only a single tool call.
python3 toolhub/demo/openai_chat.py \
--task="Could you generate a random password for me of 15 characters?" \
--collections="random"

As a Python library (refer to toolhub/demo/openai_assistant.py):

# Initialize the auth_ctx as explained above, e.g.
auth_ctx = auth.StandardAuthContext(
    openapi=auth.OpenApiAuthContext(
        api_to_headers={"crunchbase": {"X-cb-user-key": <crunchbase_key>}}
    ),
    rapidapi=auth.RapidApiAuthContext(
        rapidapi_key=<rapidapi_key>,
        host_to_headers=None,
    ),
)

# Initialize a registry as explained above, e.g. with multiple providers.
registry_ = registry.Registry(
    [
        random_provider.Provider(),
        openapi_provider.Provider.standard(),
        rapidapi_provider.Provider.standard(
            filter_rapidapi_api_hostnames=["https://alpha-vantage12.p.rapidapi.com/"],
            filter_rapidapi_endpoint_urls=["[https://yelp-reviews.p.rapidapi.com/business-search](https://yelp-reviews.p.rapidapi.com/business-search)"],
        ),
    ],
    filter_collections=[
        "random",
        "crunchbase",
        "Financial.currency_converter_v2"
    ],
)

# Set up the Hub
hub = openai_assistant_hub.OpenAIAssistantHub(
    registry_=registry_, auth_ctx=auth_ctx
)

# Initialize the Open API client and run
client = openai.OpenAI(api_key=settings.openai.api_key)
assistant = client.beta.assistants.create(
	name="ToolHub assistant",
	instructions="You are a helpful assistant who uses tools to perform tasks.",
	tools=hub.tools_spec(), # OpenAI currently supports upto 128 tools
	model="gpt-4-1106-preview",
)
run = client.beta.threads.runs.create(
	thread_id=thread.id,
	assistant_id=assistant.id,
)

...
# Set up any application logic here
...
if (
    run.status == _REQUIRES_ACTION_RUN_STATUS
    and (tool_calls := run.required_action.submit_tool_outputs.tool_calls)
):
    tool_outputs = []
    for tool_call_id, result in hub.call_tools({
          tc.id: tc for tc in tool_calls
	}).items():
    if isinstance(result, hub.ToolCallErrors):
        errors_fmt = "\\n".join(str(e) for e in result.errors)
        tool_outputs.append(
            openai_assistant_hub.ToolOutput(
                tool_call_id=tool_call_id,
                output=f"Failure:\\n{errors_fmt}",
            )
        )
    else:
        tool_outputs.append(result)
        client.beta.threads.runs.submit_tool_outputs(
            thread_id=thread.id,
            run_id=run_id,
            tool_outputs=tool_outputs,
        )

Limitations

  • Try to restrict the number of functions that you’re providing - adding too many can overwhelm the LLM’s context, and lead to poor results.
    • Filter to only the collections/functions you’d like to use.
  • Tools which provide a very large amount of output do not work well, since the model is unable to comprehend the results.
  • Calling sequences of events can often fail - the model is sometime not able to perform planning without careful prompting.
  • Currently, only RapidAPI GETs are supported - we are actively working on a framework for POSTs as well

Want to overcome these limitations? Check out our premium offering at https://www.bloomerai.com/

Extend and customize

Add an OpenAPI API

We support adding any API defined in the OpenAPI format. Follow these instructions, and please do contribute back! Refer to demo/staockbot.py for an example

  1. Identify and download the OpenAPI JSON schema to build against.

    1. e.g. https://github.com/alpacahq/alpaca-docs/blob/master/oas/trading/openapi.yaml
  2. Define a Python loader for this API - you'll need the OpenAPI schema and the base URL. The standard API loader in toolhub.integrations.openapi should work in most cases.

    1. [if you need customization] In the Python loader module, implement a custom MyParser(parser.Parser) with filter_endpoint (to filter out certain functions) and/or map_parameter (to change certain parameters, e.g. updating descriptions).
  3. Construct a registry with this loader, and any other loaders you need. You may optionally filter the functions to a subset you'd like to support

    SCHEMA_PATH = pathlib.Path(
        os.path.join(os.path.dirname(__file__), "alpaca.yaml")
    )
    BASE_URL = 'https://paper-api.alpaca.markets'
    alpaca_api_loader = openapi_provider.standard_api_loader(
        api="alpaca",
        schema_path=SCHEMA_PATH,
        request_body_descriptions_path=None,
        base_url=BASE_URL,
    )
    registry_ = registry.Registry(
        [
            openapi_provider.Provider(
                api_loaders=[alpaca_api_loader],
                filter_function_names=["alpaca_v2_orders_post"]
            ),
            rapidapi_provider.Provider.standard(
                filter_rapidapi_endpoint_urls=[ # Any other RapidAPI endpoints you want to support
                    "https://alpha-vantage12.p.rapidapi.com/",
                    "https://yelp-reviews.p.rapidapi.com/business-search",
                ],
  4. Construct an authcontext with the correct credentials

       auth_ctx = auth.StandardAuthContext(
        rapidapi=auth.RapidApiAuthContext(
            rapidapi_key=_RAPIDAPI_KEY,
        ),
        openapi=auth.OpenApiAuthContext(
            api_to_headers={
                "alpaca": {
                    "APCA-API-KEY-ID": _ALPACA_KEY_ID,
                    "APCA-API-SECRET-KEY": _ALPACA_SECRET_KEY,
                }
            },
        ),
    )
  5. Use this auth context with an agent and pass in any queries

    agent = openai_assistant.Agent(
        registry_=registry_, openai_client=openai.OpenAI(api_key=_OPENAI_KEY)
    )
    agent(auth_ctx, query)

This will enable your LLM to access any APIs, including internal ones.

Customize auth

You can pass custom additional authentication headers via the AuthContext

@dataclasses.dataclass
class OpenApiAuthContext:
    api_to_headers: dict[str, dict[str, str]] | None

@dataclasses.dataclass
class RapidApiAuthContext:
    rapidapi_key: str
    host_to_headers: dict[str, dict[str, str]] | None

@dataclasses.dataclass
class AuthContext:
    openapi: OpenApiAuthContext | None
    rapidapi: RapidApiAuthContext | None

Note: the AuthContext can be also configured without code via dynaconf (e.g. with toolhub/settings.yml) - see AuthContext.from_settings().

Contributing

We welcome contributions of new functions! Our goal is to help everyone accelerate their AI development timeline. If you found this project helpful for an API that was not covered, pleas feel free to submit a PR and give back!

Discord

Questions? Concerns? Join our Discord at https://discord.gg/jN5ePfvV