How a simple Ollama front-end grew into a professional-grade, fully offline translation environment
Not long ago, working with a large language model for translation meant opening a terminal, typing a prompt, waiting, and copying the output by hand. The arrival of Ollama made it straightforward to run powerful open-source LLMs on a personal computer, but interacting with them remained a largely technical affair. The Local AI Translator, developed by Terence Lewis with support from Philip Staiger and available at LocalAI.World, set out to change that — and the distance it has travelled since its first release tells a compelling story about what offline AI tools can become when they are built with the practising translator firmly in mind.
The earliest version of the application was, in essence, a graphical wrapper around Ollama — a way of sending text to a locally running LLM and receiving a translation back, without needing to touch a command line. That was already genuinely useful. A translator could type or paste a passage, choose a model, and get a rendered result in a clean desktop window. No API key, no subscription, no data sent to a remote server.
The privacy dimension was central from the outset. Sensitive corporate documents, legal texts, medical records — content that organisations are rightly reluctant to upload to cloud services — could now be handled entirely on the user’s own machine. The application runs on any Windows 11 PC with 16 GB of RAM, with no GPU required, making it accessible to professionals on standard hardware.
As the application matured, a row of feature buttons began to define its identity. Model Manager, Prompt Manager, RAG Manager, Back-translate, Batch Translation, and Tools — each button represents a layer of capability added in response to the real demands of translation work.
Model Manager and Model Evaluator
The Model Manager allows users to download, switch, and delete LLMs directly from the GUI, with direct access to the Ollama model library. Alongside it sits a Model Evaluator — a genuinely important feature given the proliferation of models claiming multilingual capability. The evaluator lets users benchmark a series of LLMs against each other for a specific language pair using standard metrics, so they can make an informed choice rather than guessing which model will perform best for their use case. Each evaluated translation can be saved with its model name for later comparison.
Prompt Manager and RAG Manager
The Prompt Manager gives translators something that cloud tools rarely offer: the ability to write a proper brief for the model. Custom prompts can range from a handful of terminological preferences to comprehensive style guides or extensive domain-specific glossaries. The RAG Manager goes further, enabling Retrieval Augmented Generation in three modes: a flat glossary import, a knowledge graph built from a source text and an existing translation memory, and import of an existing knowledge graph in JSON, TTL, or GML format. A simple checkbox toggles RAG on and off, so the user can compare results with and without augmentation in real time.
File Translation and Batch Processing
Beyond screen input, the application can translate text files at sentence, paragraph, or document level — a meaningful distinction, since paragraph and document-level processing gives the model the broader context it needs for consistent terminology and register. MS Office files, LibreOffice files, and PDFs are all supported. A Batch Translation module handles multiple files simultaneously, across any mix of formats and language pairs.
The Tools Suite
A Tools menu provides ten additional utilities: a Model Supported Language Checker, an AI Chat box for ad-hoc research and glossary generation, an Image Analyser capable of intelligent OCR and image reasoning, a Bitext Creator, an XLIFF Generator, a Transcriber for audio and video, a Rewriter, a Proofreading and Correction tool, a Terminology Extractor, and a Summarizer. These utilities are designed to orbit the central translation task, supporting it from multiple angles rather than functioning as standalone products.
The most significant step in the application’s evolution has been the introduction of Interactive Mode — the point at which it stopped being a translation generator and became a translation environment. Activating Interactive Mode brings up a segmented grid that will feel immediately familiar to anyone who has worked in Trados, MemoQ, or any other CAT (Computer-Assisted Translation) tool.
The grid displays source and target segments side by side, at sentence or paragraph level as the user chooses. The LLM generates a proposal for each segment, but the translator remains firmly in control: any proposal can be edited directly in the target cell, or discarded entirely and replaced with a human translation. The model can even be changed mid-session, allowing the user to call on a different LLM for a particularly difficult segment without abandoning the current project.
One of the most distinctive features of the Interactive Mode is the ability to present the translator with multiple proposals for each segment. In one configuration, a single model generates three alternative translations, giving the user a choice of renderings and a basis for selecting or synthesising the best option. In another, three different LLMs each contribute their own proposal for the same source segment.
This multi-model approach is a genuine innovation. Different LLMs have different strengths — in fluency, in terminological precision, in handling idiomatic language — and presenting their outputs side by side turns model variation from a problem into a resource. The translator can pick the proposal that best fits the context, or use the spread of suggestions as a prompt for their own formulation. It is, in effect, a form of AI-assisted quality control built into the translation workflow itself.
A translation produced in Interactive Mode does not have to stay within the application. Once the translator is satisfied with the segmented bilingual content, it can be saved in several formats designed for onward use in professional workflows.
Output formats include:
• XLIFF — the XML Localisation Interchange File Format, the lingua franca of the translation industry, importable directly into Trados, MemoQ, Wordfast, and most other professional CAT environments
• TMX — Translation Memory eXchange format, allowing the completed bilingual content to be added to an existing translation memory in any compatible tool
• DOCX — as a standard or bilingual Word document
• ODT — for LibreOffice compatibility
• TSV and plain text for lightweight downstream processing
The XLIFF and TMX outputs are particularly significant. They mean that a translator can use the Local AI Translator to produce a first-pass bilingual draft entirely offline, then import it into a commercial CAT tool for terminology checking, quality assurance, or client delivery in a familiar format. The offline and professional worlds, often seen as separate, become directly interoperable.
Addendum:
The Local AI Translator also supports translating subtitle files in SRT and VTT formats. Furthermore, it is able to produce subtitle files from transcription, starting with audio files such as mp3 or wav, and the audio track from video files (mp4, wmv), which can subsequently be translated.
The application’s developers have been explicit about one of its core motivations: making professional-grade AI translation tools available beyond the markets where cloud services are cheap and connectivity is reliable. For researchers, educators, and translators in countries where API costs denominated in US dollars represent a significant burden, where internet access is intermittent, or where data sovereignty concerns make cloud processing inadvisable, a one-time purchase of €149 that runs entirely locally is a materially different proposition from a monthly subscription to a remote service.
Once a model has been downloaded, the application requires no connection at all. That single characteristic — complete offline operation — is not a minor technical detail. For a significant portion of the world’s translators, it is the difference between access and exclusion.
The Local AI Translator has travelled a long way from its origins as a graphical front-end for Ollama. It now offers a translation environment that competes meaningfully with commercial CAT tools in its core workflow, while adding capabilities — multi-model proposal comparison, knowledge graph RAG, integrated back-translation, offline audio and video transcription — that most commercial tools do not provide at any price.
Its development reflects a broader truth about offline AI tools: the constraints of local deployment, rather than limiting what is possible, can focus development on what actually matters to the people doing the work. By keeping the translator at the centre of the process — in control of every segment, able to choose between proposals, able to bring their own judgement to bear at every point — the Local AI Translator has arrived at something more interesting than a translation machine. It is a translation workbench.
The Local AI Translator is available at localai.world. A trial version (10€) and full licence (149€) are offered. The application runs on Windows 11 with a minimum of 16 GB RAM; bespoke Linux versions are available on request.