Our client needed to parse thousands of messy, freetext documents into structured JSON every day. The data was too sensitive to send to a third-party API. So we fine-tuned an open-source 8B parameter model, deployed it on our own infrastructure, and built a Python pipeline around it that now runs in production with automated monthly retraining.
This talk is the full blueprint. I'll walk through every stage: a Django + Pydantic QA portal for curating high-quality training data, synthetic data generation at scale via BigQuery, fine-tuning with Axolotl and LoRA adapters (yes, on consumer GPUs), a DuckDB-powered benchmark framework that gates every model release, and a FastAPI + Docker deployment that keeps the whole thing running without anyone babysitting it. No API keys. No vendor lock-in. Just Python, open-source models, and a lot of iteration. If you've ever wanted to move beyond prompt engineering and actually own your model end to end, this talk will show you how.
PyOhio 2026