industry-windowsProductionizing Kiln

How to take ideas from the lab (Kiln) into your product

Kiln is a great place to rapidly experiment with many models, providers, fine-tunes, and prompts.

Once you've found the ideal way to run your AI workload, you are ready to create (or iterate) on your product. This guide walks through the options for creating products from Kiln projects.

Productionizing Kiln Tasks

Option 1: Use the Kiln Python Library

The MIT open-source Kiln python library powers all of the requests in the Kiln app. You can use it in your app or server, pointing it to the Kiln project of your choice.

Our python library docs have instructions for exporting your project and running it via pythonarrow-up-right.

Option 2: Exporting Kiln Search Tools / RAG

We have multiple options for deploying search tools you build inside of Kiln. See the docs for options for deploying search tools / RAG.

Option 3: [Simple Tasks Only] Replicate Kiln Prompts and Requests in the Framework of your Choice

circle-exclamation

Every model request sent from Kiln is logged to ~/.kiln_ai/logs/model_calls.log . This includes the messages and any parameters sent to the API (model, temperature, reasoning, etc). It even includes a curl command to replicate the call Kiln made from the command line.

You can use the data in this log to exactly replicate any Kiln request in the language/framework of your choice. You can choose which tech-stack makes sense for your product; anything from simple HTTP requests, to integrating provider libraries (pip install openai), to integrating llama.cpparrow-up-right into your app/process.

Last updated