Skip to content

API Reference

Learn more about the API reference for embedded function calling.

runWithTools

This wrapper method enables you to do embedded function calling. You pass it the AI binding, model, inputs (messages array and tools array), and optional configurations.

  • AI BindingAi
    • The AI binding, such as env.AI.
  • modelBaseAiTextGenerationModels
    • The ID of the model that supports function calling. For example, @hf/nousresearch/hermes-2-pro-mistral-7b.
  • inputObject
    • messagesRoleScopedChatInput[]
    • toolsAiTextGenerationToolInputWithFunction[]
  • configObject
    • streamFinalResponseboolean optional
    • maxRecursiveToolRunsnumber optional
    • strictValidationboolean optional
    • verboseboolean optional
    • trimFunctionboolean optional - For the trimFunction, you can pass it autoTrimTools, which is another helper method we’ve devised to automatically choose the correct tools (using an LLM) before sending it off for inference. This means that your final inference call will have fewer input tokens.

createToolsFromOpenAPISpec

This method lets you automatically create tool schemas based on OpenAPI specs, so you don’t have to manually write or hardcode the tool schemas. You can pass the OpenAPI spec for any API in JSON or YAML format.

createToolsFromOpenAPISpec has a config input that allows you to perform overrides if you need to provide headers like Authentication or User-Agent.

  • specstring
    • The OpenAPI specification in either JSON or YAML format, or a URL to a remote OpenAPI specification.
  • configConfig optional - Configuration options for the createToolsFromOpenAPISpec function
    • overridesConfigRule[] optional
    • matchPatternsRegExp[] optional
    • options Object optional { verbose boolean optional }