<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>AI on Tom Burkert</title><link>https://blog.burkert.me/tags/ai/</link><description>Recent content in AI on Tom Burkert</description><image><title>Tom Burkert</title><url>https://blog.burkert.me/assets/</url><link>https://blog.burkert.me/assets/</link></image><generator>Hugo -- 0.148.0</generator><language>en-us</language><lastBuildDate>Mon, 30 Mar 2026 21:39:05 +0200</lastBuildDate><atom:link href="https://blog.burkert.me/tags/ai/index.xml" rel="self" type="application/rss+xml"/><item><title>Multilingual support in LLMs is not a nice-to-have</title><link>https://blog.burkert.me/posts/ai_multilingual_llm_study/</link><pubDate>Mon, 30 Mar 2026 21:39:05 +0200</pubDate><guid>https://blog.burkert.me/posts/ai_multilingual_llm_study/</guid><description>&lt;p>I remember back in November 2022, when ChatGPT was released and everyone was going crazy about how smart and humanlike it was. I sat down and started talking to it in Czech. After reading so many raving reviews and hyperbolic predictions, I was shocked by how quickly I ran into the model&amp;rsquo;s limits in Czech. Now don&amp;rsquo;t get me wrong: I know my native tongue is not a top-tier language, given it only has about 10 million native speakers. I don&amp;rsquo;t expect AI models to be as proficient in Czech as in English or other major languages. But it still made me think: if I am getting a much worse experience from large language models than English speakers are, what is the experience like for people who speak languages with even less digital support?&lt;/p></description><content:encoded><![CDATA[<p>I remember back in November 2022, when ChatGPT was released and everyone was going crazy about how smart and humanlike it was. I sat down and started talking to it in Czech. After reading so many raving reviews and hyperbolic predictions, I was shocked by how quickly I ran into the model&rsquo;s limits in Czech. Now don&rsquo;t get me wrong: I know my native tongue is not a top-tier language, given it only has about 10 million native speakers. I don&rsquo;t expect AI models to be as proficient in Czech as in English or other major languages. But it still made me think: if I am getting a much worse experience from large language models than English speakers are, what is the experience like for people who speak languages with even less digital support?</p>
<p>It set me on a path toward working on multilingual LLM capabilities and the challenge (and fun) of measuring them. We were already helping some major labs train AI models in various languages, and <em>some</em> of our clients used to give even longer-tail languages a lot of attention; some still do. But I feel like the &ldquo;race to AGI&rdquo; has completely overshadowed the importance of truly inclusive multilingual systems.</p>
<p>Let&rsquo;s start with OpenAI and its latest model, GPT-5.4: <a href="https://openai.com/index/introducing-gpt-5-4/" target="_blank" rel="noopener">the press release</a>, which is almost a 4,000-word document, mentions the words &ldquo;language&rdquo;, &ldquo;linguistic&rdquo; or &ldquo;multilingual&rdquo; exactly zero times. OpenAI used to give some attention to multilingual performance of its models: it translated the Massive Multitask Language Understanding (MMLU) benchmark into <a href="https://github.com/openai/simple-evals/blob/main/multilingual_mmlu_benchmark_results.md" target="_blank" rel="noopener">14 languages</a> and used to report the scores in its press releases and models&rsquo; system cards. For the newest models, the scores are not mentioned in the press release, and even worse, not even in the <a href="https://deploymentsafety.openai.com/gpt-5-4-thinking" target="_blank" rel="noopener">system card</a>. The only acknowledgment that the model can handle non-English text appears in a footnote. Neither the note nor the linked help article makes clear whether &ldquo;support&rdquo; refers to UI localization or actual model capability, and, if the latter, whether that applies to text, voice, or both.</p>
<p>OpenAI is not alone in this. Anthropic also makes no mention of multilingual capabilities in its <a href="https://www.anthropic.com/news/claude-opus-4-6" target="_blank" rel="noopener">press release for Claude Opus 4.6</a>, unless you count &ldquo;multilingual coding&rdquo;, that is, the knowledge of different programming languages. Anthropic does produce a 213-page <a href="https://www-cdn.anthropic.com/6a5fa276ac68b9aeb0c8b6af5fa36326e0e166dd.pdf" target="_blank" rel="noopener">system card</a>, and it actually does comment on the multilingual MMLU score! Sadly, the commentary spans two sentences and does not even provide per-language statistics.</p>
<p>When it comes to documenting which specific languages are supported, Anthropic is even worse than OpenAI. Anthropic has a <a href="https://platform.claude.com/docs/en/build-with-claude/multilingual-support" target="_blank" rel="noopener">Multilingual support</a> page, which seems like a good start, until you realize that&rsquo;s just a place to put results for the multilingual MMLU benchmark. At the time of writing, that page does not contain results for Claude Opus 4.5, Opus 4.6, or Sonnet 4.6. Anthropic also does not provide the full list or even a number. Instead, it just says &ldquo;Note that Claude is capable in many languages beyond those benchmarked below.&rdquo; Will it work well in Czech? Or Basque? Who knows! The most we get from Anthropic is that &ldquo;Claude processes input and generates output in most world languages that use standard Unicode characters.&rdquo;</p>
<p>There are AI labs that do better than the two behemoths. Mistral, for example, touts its <a href="https://mistral.ai/news/mistral-3" target="_blank" rel="noopener">Mistral Large 3</a> as &ldquo;The next generation of open multimodal and multilingual AI&rdquo; and claims its model supports more than 40 languages. The <a href="https://docs.mistral.ai/models/mistral-large-3-25-12" target="_blank" rel="noopener">model page</a> and the <a href="https://legal.cms.mistral.ai/assets/1e37fffd-7ea5-469b-822f-05dcfbb43623" target="_blank" rel="noopener">technical documentation</a> provide no additional information (in fact, they do not even repeat the figure), but at least you can get a <em>partial</em> list at <a href="https://huggingface.co/mistralai/Mistral-Large-3-675B-Instruct-2512#:~:text=English%2C%20French%2C%20Spanish%2C%20German%2C%20Italian%2C%20Portuguese%2C%20Dutch%2C%20Chinese%2C%20Japanese%2C%20Korean%2C%20Arabic" target="_blank" rel="noopener">Hugging Face</a>. Not from Mistral&rsquo;s own docs, mind you, but from Hugging Face.</p>
<p>Not everyone does this badly, though. Google publishes <a href="https://docs.cloud.google.com/vertex-ai/generative-ai/docs/models#expandable-1" target="_blank" rel="noopener">a full list of supported languages</a>, both generally and by model family. Similarly, Alibaba publishes a <a href="https://qwen.ai/blog?id=qwen3#:~:text=Multilingual%20Support" target="_blank" rel="noopener">full table of supported languages</a> in its model announcements. Both also report multilingual MMLU scores in their model cards; aggregated rather than per-language, but at least present.</p>
<p>Why does this matter? It is easy to argue that English is the lingua franca, that many knowledge workers already use it, and that this is especially true in higher-income markets. But that view is short-sighted. If frontier AI labs expect their products to become truly universal, multilingual support cannot remain an afterthought. Sooner or later, it becomes both a product-quality issue and a growth issue. Once the English-speaking market is saturated, the next wave of users will come from other languages and cultural contexts. And if labs are serious about serving them, they need to start treating multilingual capability as something to measure, document and improve, not something users are left to discover on their own.</p>
<p>To be fair, multilingual performance has genuinely improved. Models handle Czech far better today than they did in 2022, and the gap between English and non-English scores on standardized benchmarks has narrowed. I have some reservations about those benchmarks that I&rsquo;ll address in a separate post, but the trend is clear. The question is how much of that improvement is concentrated effort and how much is just a side effect of more linguistically diverse pre-training data or better model architectures. And I&rsquo;m not claiming labs do no multilingual work internally; I&rsquo;m saying the fact that they don&rsquo;t consider it worth reporting tells you where it sits among their priorities. What isn&rsquo;t publicly reported isn&rsquo;t being held to a standard. More importantly, multilingualism in LLMs is far from a solved problem, despite the lack of attention it receives.</p>
<p>I have spent much of the last several years working on this problem directly. My team and I built a multilingual benchmark that deliberately moves away from automated public benchmarks such as multilingual MMLU. Those benchmarks are useful, but they mostly test multiple-choice comprehension rather than real conversational ability. Our study focuses instead on open-ended language generation and manipulation, evaluated at scale by language professionals. I do not usually write about my day job on this blog, but this is one of the rare cases where it feels directly relevant. It is about as close to a passion project as work gets for me. We call it the &ldquo;Multilingual LLM synthetic data generation study,&rdquo; and the full 80-page report is available <a href="https://www.rws.com/artificial-intelligence/train-ai-data-services/multilingual-llm-synthetic-data-gen-study-2/" target="_blank" rel="noopener">here</a>.</p>
<p>Of course, releasing a report once a year is not enough. We are already working on something that should shed more light on models&rsquo; multilingual performance on a more continuous basis. Perhaps the most important thing, though, is to get people talking about this. While I do not believe LLMs are the road to AGI, and I have been vocal on this very blog about their negative aspects, I also believe there is a great deal of potential in them in their current form. There are underserved communities across the world that could benefit from easier access to language models that can communicate with them well.</p>
<p>Czech has gotten much better since 2022. But if models still trip up in a language with 10 million speakers and a strong online presence, imagine what it&rsquo;s like in Kinyarwanda, Fijian, or Kyrgyz.</p>
]]></content:encoded></item><item><title>An AI taught me to listen to birds</title><link>https://blog.burkert.me/posts/birding/</link><pubDate>Sun, 01 Mar 2026 16:46:25 +0100</pubDate><guid>https://blog.burkert.me/posts/birding/</guid><description>&lt;p>A few days ago, I opened the window and heard a familiar sound I have not heard in a while: it was &lt;a href="https://youtu.be/19iP35PwYIU?t=7" target="_blank" rel="noopener">a fieldfare&amp;rsquo;s call&lt;/a>, probably one of the first who migrated back to my town this year. And just today, I saw my first chaffinch of the year perched on top of a branch. These birds are very common where I live, so this is not too surprising, but what struck me is how much I have grown in terms of knowledge of their calls, appearance and behaviour. A few years ago, I wouldn&amp;rsquo;t have been able to name a single bird by call or name too many of them by looking at them. If you&amp;rsquo;re curious how I got here or want to build some knowledge and appreciation of birds, read on.&lt;/p></description><content:encoded><![CDATA[<p>A few days ago, I opened the window and heard a familiar sound I have not heard in a while: it was <a href="https://youtu.be/19iP35PwYIU?t=7" target="_blank" rel="noopener">a fieldfare&rsquo;s call</a>, probably one of the first who migrated back to my town this year. And just today, I saw my first chaffinch of the year perched on top of a branch. These birds are very common where I live, so this is not too surprising, but what struck me is how much I have grown in terms of knowledge of their calls, appearance and behaviour. A few years ago, I wouldn&rsquo;t have been able to name a single bird by call or name too many of them by looking at them. If you&rsquo;re curious how I got here or want to build some knowledge and appreciation of birds, read on.</p>
<p>Before I go any further, I have to stress that I am not a bird expert and I am not even a super serious bird watcher / birder. But I have built an immense admiration for birds and whenever I go out and especially when I travel, I am on the lookout for them. I have always been fascinated by the outdoors, but my interest in birds actually started when I kept hearing a strange call at a playground in front of our house and kept wondering what that was. It turned out it was a fieldfare, but I only knew after I downloaded the <a href="https://merlin.allaboutbirds.org/" target="_blank" rel="noopener">Merlin Bird ID app</a> and managed to capture the call. Here it was, clearly matching the already captured calls, with a description and plenty of additional information. Bingo!</p>
<p><img alt="Screenshot of the Merlin Bird ID app" loading="lazy" src="/images/merlin.jpg#center"></p>
<p>The magic of the Merlin Bird ID app is being able to identify a bird in multiple ways: either going through a decision tree about your location, the bird&rsquo;s appearance and behaviour, or through recording its calls, or from a photo. I&rsquo;d argue these are some of the most brilliant ways artificial intelligence has been used. As Ian Campbell put it in his <a href="https://www.inverse.com/tech/merlin-bird-id-app-cornell-university-ai" target="_blank" rel="noopener">Inverse article</a> about the app, it&rsquo;s &ldquo;an AI that helps push you away from your phone and appreciate the world around you,&rdquo; which stands in stark contrast to algorithmic feeds and the personalized hell of social media that try to suck you into your phone&rsquo;s screen.</p>
<p>Most importantly, the Merlin app set me on a path to grow my interest in birds, because it made identifying birds easy and frictionless; it gave me a great head start and the motivation to continue. Soon, I didn&rsquo;t even need to pull out my phone to identify the more common birds in my area just by their call. It may not be the most useful skill in the world, but I was so proud of myself, and felt happiness and warmth for reconnecting with the natural world, something we seem to do less and less.</p>
<p>The app also encourages you to collect &ldquo;lifers&rdquo; (birds you were able to identify), very much like you&rsquo;d collect Pokemon. Gotta catch them all! And I do try to catch them all, whether it&rsquo;s blue jays and great-tailed grackles in Texas, kestrels in the UK or common magpies in Spain. It is genuinely fun to collect them, and it invites you to learn more about the ecosystems of your neighborhood, as well as places you visit. The app is created by the Cornell Lab of Ornithology, is fully free and has so far avoided any enshittification, which I am very glad for.</p>
<p>Perhaps the most interesting bit is that this little app has changed me in a meaningful way: I acquired quite a few books on birds and learned more than what the app itself can provide. I now jump with joy when I hear the first common swifts fly over our house in droves, screaming &ldquo;sreeee&rdquo; over each other. We rescued quite a few birds in distress, and we watched with delight and anticipation when a pair of blackbirds made a nest on our balcony; doubly so when we noticed that it&rsquo;s the same pair we see each year at our bird feeder. Identifying a single bird&rsquo;s call led me on a path that I could not have foreseen. So I urge you: give it a try. You might thank yourself in a year or two!</p>
]]></content:encoded></item><item><title>How many AI PhDs does it take to change a hinge?</title><link>https://blog.burkert.me/posts/ai_phds_changing_a_hinge/</link><pubDate>Mon, 08 Dec 2025 21:35:09 +0100</pubDate><guid>https://blog.burkert.me/posts/ai_phds_changing_a_hinge/</guid><description>&lt;p>It started with a broken hinge in our kitchen cabinet. I am no handyman, but I like to do as much as I can around the house and I dare say I&amp;rsquo;ve gotten pretty okay at it. The one thing that can set me back, though, is not knowing the correct terminology. What type of hinge do I need? Concealed? Inset? Overlay? Clip-On? Soft-close? Sprung? An avalanche of terms that I only barely understand paralyzes me. Luckily, there&amp;rsquo;s a new helper in town: AI!&lt;/p></description><content:encoded><![CDATA[<p>It started with a broken hinge in our kitchen cabinet. I am no handyman, but I like to do as much as I can around the house and I dare say I&rsquo;ve gotten pretty okay at it. The one thing that can set me back, though, is not knowing the correct terminology. What type of hinge do I need? Concealed? Inset? Overlay? Clip-On? Soft-close? Sprung? An avalanche of terms that I only barely understand paralyzes me. Luckily, there&rsquo;s a new helper in town: AI!</p>
<p>I&rsquo;ve used vision-enabled LLMs before to identify types of tools, bike parts, or PC components with mixed results, but the net result was usually positive. At the least, I became armed with some of the terminology and knew what to look up on Google. The recent releases of Gemini 3 Pro (&quot;<a href="https://blog.google/technology/developers/gemini-3-pro-vision/" target="_blank" rel="noopener">The frontier of vision AI</a>&quot;), GPT-5 (&quot;<a href="https://openai.com/index/introducing-gpt-5/" target="_blank" rel="noopener">can reason more accurately over images and other non-text inputs</a>&quot;) and Claude Opus 4.5 (&quot;<a href="https://www.anthropic.com/news/claude-opus-4-5" target="_blank" rel="noopener">better vision, reasoning, [&hellip;] and it is state-of-the-art in many domains</a>&quot;) have made me curious to test the much hyped visual reasoning capabilities. Plus, they are touted to have PhD- or expert-level intelligence. Can an AI PhD help me change a hinge?</p>
<h2 id="mission-impossible">Mission (im)possible</h2>
<p>I set out with a simple request: &ldquo;Identify this type of hinge and find out where it can be bought in Czechia,&rdquo; accompanied by the following picture:<br>
<img alt="Hinge" loading="lazy" src="/images/hinge.jpg#center"></p>
<p>Now, a small aside about hinges — the little I know. I promise it won&rsquo;t be long, it can be explained in 30 seconds.</p>
<p>This is a hinge with a cup:<br>
<img alt="Hinge with a cup" loading="lazy" src="/images/hinge-cup.jpg#center"></p>
<p>And this, my friends, is a type of hinge that does not use a cup, but is mounted directly to the face frame of the cabinet:<br>
<img alt="Hinge without a cup" loading="lazy" src="/images/hinge-face-frame.jpg#center"></p>
<p>You can tell them apart very easily - I actually think 99% of you could now tell them apart without any issues. You can probably also tell that the hinge from my original picture is very clearly the latter one, as there is no place to mount the cup to in the cabinet face frame.</p>
<p>Our PhD-level expert LLM friends? Not so much.</p>
<h2 id="getting-unhinged">Getting unhinged</h2>
<p>Let&rsquo;s start with Gemini. Even though I prompted it in English, it started thinking and responding in Czech, so I can&rsquo;t quote it verbatim, but here&rsquo;s a faithful translation from Google Translate:</p>
<blockquote>
<p>The specific type in the photo appears to be: A cup hinge with a cup diameter of 35mm (this is the standard size, but it&rsquo;s a good idea to check).</p></blockquote>
<p>OK, that&rsquo;s clearly wrong.</p>
<p>Let&rsquo;s move on to Claude Opus 4.5, with excellent visual reasoning:</p>
<blockquote>
<p>This is a Hettich concealed cabinet hinge (also called a European-style or cup hinge).</p></blockquote>
<p>Not really! But wait, Claude is searching the web, maybe it can recover:</p>
<blockquote>
<p>Based on my search, this is a Hettich Intermat concealed cabinet hinge (also known as a &ldquo;miskový závěs&rdquo; or cup hinge in Czech). It features a 35mm cup diameter and includes a clip-on mounting plate for easy installation.</p></blockquote>
<p>Not quite either.</p>
<p>Maybe ChatGPT with GPT-5 will save the day?</p>
<blockquote>
<p>Based on your photo, this appears to be a “European-style concealed cabinet hinge” — but specifically a variant that is face-frame / surface-mounted rather than the hidden “cup-in-door” type often used in frameless cabinets.</p></blockquote>
<p>Alright, that&rsquo;s much better! It seems that GPT-5 can reason more accurately over images. So what is its recommendation?</p>
<blockquote>
<p>My Recommendation<br>
Given your photo and situation, I’d start by ordering a clip-on concealed hinge for face-frame cabinets with a 35 mm cup — either basic or soft-close depending on whether you care about silent closing. That has the highest chance of matching the old hinge and mounting method.</p></blockquote>
<p>Somehow, in the stream of tips and examples, GPT-5 convinced itself that this was after all a hinge with a cup, and ended up giving me the wrong recommendation.</p>
<p>I honestly do not know why all the models faceplanted so badly. It does seem that my type of hinges is relatively rare, but at least GPT-5 identified it correctly at first.</p>
<h2 id="just-ask-a-real-expert">Just ask a real expert</h2>
<p>Turns out, the AI labs aren&rsquo;t optimizing for mundane DIY realities, and so you&rsquo;d be better off asking literally anyone at your local hardware store or DIY shop. Even the SOTA LLMs are not there in terms of giving you reliable information for this category of problems. They fell short of even identifying my hinge, much less finding a site where I could buy a replacement.</p>
<h2 id="the-smallest-of-wins">The smallest of wins</h2>
<p>There is a silver lining. When I confronted the LLMs with the fact that the hinge in the picture clearly does not have a cup, all three recovered and provided some genuinely useful information. They helped me nail down the terminology, and pointed me toward several e-shops I could browse to look for what I needed.</p>
<p>However, none of them were able to identify a specific product I could actually order. And perhaps more importantly, they only became helpful after I called them out on their mistakes. If I had trusted the initial responses (or if this had been a topic I knew even less about), I could easily have ended up ordering the wrong part.</p>
<p>An assistant you have to constantly fact-check and correct is of questionable value. For now, these expensive statistical behemoths remain more useful as a second opinion than a first resort for this type of problems.</p>
]]></content:encoded></item></channel></rss>