Skip to content

LLM Support

Koog Integration

Trailblaze makes API calls leverage the koog.ai library.

Trailblaze takes an instance of a LLMClient from https://github.com/JetBrains/koog.

Koog currently has libraries for the following LLM providers

  • Google
  • OpenAI
  • Anthropic
  • OpenRouter
  • Ollama

Any of these can be passed as the llmClient argument to the TrailblazeAgent constructor.

Example usage with OpenAI

Gradle Dependency: ai.koog:prompt-executor-openai-client:VERSION

import ai.koog.prompt.executor.clients.LLMClient
import ai.koog.prompt.executor.clients.openai.OpenAILLMClient
import ai.koog.prompt.executor.clients.openai.OpenAIModels
import xyz.block.trailblaze.agent.TrailblazeRunner

private val llmModel: LLModel = OpenAIModels.Chat.GPT4_1
private val llmClient: LLMClient = OpenAILLMClient("API_KEY_HERE")
TrailblazeRunner(
    llmClient = llmClient,
    llmModel = llmModel,
    //...
)