Schroddy

OCR Magic in Go: Build AI-Powered Text Extraction with Ollama in Minutes

Back to Schedule

Abstract

Are you tired of people saying Go doesn't work well with AI? Or that Python and JavaScript are better suited for AI? If you love Go (or you're an aspirant gopher), join my talk and unleash the potential of Go in the AI world!

In this talk, we will focus on how OCR technologies work with Go. To play with OCR (Optical Character Recognition), we'll use the Ollama platform to run large language models locally in Go. Ollama allows us, among other features, to run models both locally and in the cloud.

The models we're going to use are vision models, such as qwen3-vl, granite3.2-vision, llava, etc.
We'll explore the approaches - the generate and the chat - we can use to talk to LLMs.

Understanding these differences is key to building effective AI applications in Go. This hands-on session blends theory with live coding to demystify integrating vision models into Go applications.

Leave with the skills to build real-world OCR tools using Go and open-source vision models.