Tool calls are a mechanism that enables a language model to function as a decision-making engine. The LLM suggests when a tool should be used based on the input it receives. Once the tool is invoked by the client, the output is passed back to the LLM for further processing. This setup allows the LLM to extend its capabilities by integrating external tools, such as internal or external APIs or databases.
LLMs may choose different tools or methods for the same prompt, which can make their behavior unpredictable and complicate debugging in complex environments.
LLMs lack inherent contextual awareness about the tools they invoke. They cannot anticipate specific results unless explicitly guided, leading to potential mismatches between expectations and outcomes. For example, if the LLM calls a weather API but receives an error or unexpected format, it may not handle the situation effectively without clear instructions.
Tool calls can be used to augment LLM's capabilities. By leveraging external tools, it's possible to extend the model's functionality beyond text-based responses. For instance:
Structured outputs is a way to ask LLM to adhere to a schema in it's responses. This is very useful in context of building logic on top of LLMs and is strongly recommended to do in almost every programmatic case possible.
In context of Javascript model schemas are done with Zod. OpenAi, for instance, provides necessary helpers in their NPM package.
Structured outputs can serve as guard rails between an LLM and the application logic. When invoking tools, one can manage the thread of outputs consistently and compose the final rendered content with greater confidence.
Structured outputs help bridge the gap between the freeform nature of LLM-generated text and the strict requirements of tool-based integrations.