Sr. Content Developer at Microsoft, working remotely in PA, TechBash conference organizer, former Microsoft MVP, Husband, Dad and Geek.
152523 stories
·
33 followers

Beyond Python: Why LLMs Need More Stable, Open Source Code

1 Share
python

In the future dreams of Large Language Models, there will be no need for human intervention and they will write code in a highly efficient form unreadable to us. Perhaps they will write directly in binary. But for now, we need our LLM assistants to write readable code, in an established programming language.

So what languages do LLMs prefer to use today? In terms of the programming language most popular with human developers, where we assume models get their training material from, the choices should be between JavaScript/TypeScript, Java and Python. But in fact, we don’t quite see that.

The Current Python Bias in LLM Code Generation

What we see right now is a massive Python bias, as an academic study points out. The conclusion is straightforward: “Our findings demonstrate that LLMs exhibit a strong bias towards Python, preferring it for both benchmark tasks and project initialisation tasks, even when there are more suitable choices.”

However, the same study makes a more important claim: “LLMs display a limited range when choosing coding libraries, favouring well-established options over high-quality alternatives.” Python is getting more popular, but one suspects that the LLM creators have favoured Python training sets.

This is not particularly worrying. In many cases LLM tools start working on legacy code, where a language has already been chosen. Or the language chosen is part of the identity of the main library or platform that isn’t available in Python.

But the study also found that when a Chain of Thought prompting is used for GPT-4o (e.g. “think step by step”), the programming language used for project initialisation tasks has much less of a Python preference:

(The languages used are given, with the percentage of responses that used the language, and the rank assigned to the language by the LLM.)

But as LLM use grows within industry professionals, one assumes that Javascript and Java will assert themselves.

The Growing Influence of Open Source Models

The better question is: what languages should an LLM choose? The answer to this will probably be guided by two things; the increase in open source models, and the growing influence of open source components.

I asked Zach Lloyd, CEO of Warp (the excellent terminal tool), what AI technology Warp is likely to exploit most in 2026? He made it clear that it was open source models. “As they continue improving, we’ll use them alongside proprietary options — giving us more optionality and resilience,” he said. “Competition at this layer is also great for the developer ecosystem because it drives quality up and prices down.”

Open source models have no corporate projects to favour. So you might expect OpenAI to gently push Microsoft’s C#, and maybe Gemini will have greater access to Golang. But open source models will just tend to train with the code legitimately available to the developers.

Why Maintainability Matters in AI-Generated Code

The strongest signal from the ‘survival of the fittest’ tools will simply be the need for generated code to be less ‘vibey’ and more maintainable. This means a preference away from currently popular languages and frameworks, towards those with proven pedigree and more trusted examples.

For instance, we can see the prominence of Web Components for the same reason. Web components are a standard that is finally achieving mass appeal. Yes, they have always offered encapsulation, reusability and framework independence, but only recently have some of the rough edges been smoothed off.

Engineers, especially senior ones, read and review more code than they write; and that is likely to increase with more LLM generated code. So cool new patterns are actually a friction if used too often.

Reducing Nondeterminism in LLM Computing

The other reason to stick to tried and trusted code is to reduce the nondeterministic nature of LLM computing — that is, their tendency to choose different options depending on what day it is. While the temperature for tools like coding assistants will always be set low, the nature of LLM token-by-token generation is that they don’t know what they will write until they are done generating.

The fuzzy ‘mind’ of an LLM may produce one answer at one point, and another entirely different answer another time. The answers it builds at any one time rely on statistical reasoning, but these use sets of probabilities, not the binary methods we usually associate with computing.

So for these reasons, I can see training biases moving towards more stable projects, more open projects, and projects with a longer history of openly available examples. As LLMs move towards commoditisation, or to the right of a Wardley Diagram, stability will become the dominant factor.

The Case for a ‘Seed Bank’ for Code

We are told that all the world’s vital plants appear in seed banks, so that we can repopulate after a disaster. A seed bank is a repository that stores seeds from diverse plant species (wild and cultivated) under suitably stable conditions. So it is a “Noah’s Ark,” but for plants. I’m writing this within sight of Kew Gardens, which manages the Millennium Seed Bank.

Millennium Seed Bank; image via Kew Gardens.

Every time we say “training data,” we wave loosely at the forums and pages available on the internet. This is why we have to assume that training is based on what is on the internet right now. What we really need is a seed bank for code. This should be straightforward for a trusted organization to set up, so that a growing set of examples can be maintained without the risk of vendor taint or third-party poisoning. While the averaging of vast quantities of internet text will provide a solid average, clearly, a tighter set would be a better place for a new model to start training from.

We don’t like to talk about the internet suffering severe damage, as that might imply some catastrophic event. And we know the military design in its heritage makes this unlikely. What we really mean is that there should exist some “other” place, where we know a safe pool of data exists so that training isn’t always dependent on the current — and very dynamic — state of the web.

The Future of Programming for LLMs

We are still near the start of the LLM journey; and right now they will use the code and projects that appear most often in their training data when generating greenfield example code. For now, that means the plusher bits of the internet with an added Python bias.

The next step will be to use code from projects that are least likely to alter over time, in order to beat down the nondeterminism of LLMs. We are probably on the edges of that now.

Only in the far future will AIs communicate with one another and develop their own intermediate language, in which human accessibility is not a priority.

The post Beyond Python: Why LLMs Need More Stable, Open Source Code appeared first on The New Stack.

Read the whole story
alvinashcraft
53 minutes ago
reply
Pennsylvania, USA
Share this story
Delete

Microsoft confirms it’s killing offline phone-based activation method for Windows 11 after 20+ years

1 Share

In a statement to Windows Latest, Microsoft confirmed that it retired the traditional Windows “telephone-based” activation method, which truly worked offline. With a telephone-based approach, you could call Microsoft’s activation phone number, and the automated process would activate your Windows or Office license. Now, this method has been retired.

I’m told the company retired the Windows or Office phone-based activation method as part of the efforts to modernize the ‘activation experience.’ If you want to activate Windows 11 or Office today, the easiest way is to link your license to a Microsoft account, which will automatically verify the product.

Windows activation

You can find out how your Windows is activated from Settings > System > Activation, and then tap on ‘Activation state.’ In my case and in most cases, it’s activated using a digital license linked to a Microsoft account.”

Windows Activation State

Phone-based activation no longer works

Microsoft retired the phone-based activation method for Windows and Office products on December 3, 2025. When you call one of the listed telephone numbers for offline activation, you will be asked to use a Microsoft account instead. But that does not mean MSA is the only way to activate Windows.

For perpetual licenses, Microsoft tells me that advanced customers can use the Product Activation Portal.

Microsoft product activation

Unlike phone-based activation, which did not require an internet connection at any point of the process, this new online Product Activation Portal wants you to log into a portal, which is “secure, reliable, and user-friendly,” according to Microsoft. Once you’re logged in and have entered the details, you can still activate Windows offline.

“Customers who rely on traditional offline activation can continue using it without changes to their environment,” Microsoft argues. The company says the process has changed, and phone call-based activation is no longer supported, but that does not mean you cannot activate Windows offline.

“While the process has been updated, offline activation capabilities remain supported,” the company said.

What do you need to activate Windows using the new Product Activation Portal flow, after “activate by telephone” has been retired?

First, you need to reach the “Activate by Telephone” screen inside Windows (or the product you’re activating). You’ll be shown activation information and a phone number, but instead of calling, you can note down the info from that screen and use it in the Product Activation Portal.

However, unlike phone-based activation, the Product Activation Portal requires a browser and internet connection, but the target PC can be offline. Now, the portal cannot be used unless you’ve a Microsoft account.

Microsoft product activation portal

The portal specifically supports a Personal Microsoft account (MSA), a work or school account, a Microsoft Entra ID account, or an Azure Government tenant account.

Did Microsoft retire “activate by telephone” to push MSA?

Some of you might argue that Microsoft retired phone-based activation because it allowed customers to avoid creating a Microsoft account. While it’s possible, I’m going to play devil’s advocate here.

I believe Microsoft retired the telephone-based approach due to low usage. Moreover, Product Activation Portal is a better option because it covers all products, not just Windows and Office, which were the two supported products of phone-based activation.

The post Microsoft confirms it’s killing offline phone-based activation method for Windows 11 after 20+ years appeared first on Windows Latest

Read the whole story
alvinashcraft
54 minutes ago
reply
Pennsylvania, USA
Share this story
Delete

GeekWire Podcast: Alexa’s next act, Microsoft’s retail play, Google’s AI Inbox, and a smart bird feeder fail

1 Share

This week on the GeekWire Podcast: Amazon and Microsoft are racing to define the next era of consumer AI, on multiple fronts. We discuss Amazon’s attempt to upgrade Alexa into a true generative AI home chatbot — complete with a new web portal and updated Alexa app — while Microsoft tries to win over retailers with a new Copilot Checkout feature.

Plus, we explore Google’s upcoming “AI Inbox” for Gmail, which promises to act like an executive assistant for your email. We talk about our smart bird feeder experiment that resulted in “fuzzy birds,” due to improper focal length. And we share our initial experience with AI automation on the Windows PC desktop using Vy from Seattle startup Vercept.

Finally, we offer a Netflix recommendation, Cover-Up, the documentary about legendary investigative journalist Seymour Hersh. We couldn’t help but wonder: what would uncover if he could digitize all those notes and put them through an AI model?

And on that theme, we lament the loss of a major American newspaper, the Pittsburgh Post-Gazette, and reminisce about the time GeekWire made an appearance on its editorial page.

Subscribe to GeekWire in Apple Podcasts, Spotify, or wherever you listen.

With GeekWire co-founders Todd Bishop and John Cook. Edited by Curt Milton.

Read the whole story
alvinashcraft
54 minutes ago
reply
Pennsylvania, USA
Share this story
Delete

125: Donut Labs Could Kill Combustion With The Solid-State Battery in This Bike

1 Share
In this episode:
• The solid-state battery that could kill combustion cars
• Volvo teasing the EX60 with big range promises
• Racing 2,000 miles non-stop in a Lucid Gravity




Download audio: https://dts.podtrac.com/redirect.mp3/audioboom.com/posts/8828416.mp3?modified=1768060927&sid=5141110&source=rss
Read the whole story
alvinashcraft
54 minutes ago
reply
Pennsylvania, USA
Share this story
Delete

Mike McQuaid on 16 years of maintaining Homebrew | Episode 5 | The GitHub Podcast

1 Share
From: GitHub
Duration: 34:06
Views: 244

Homebrew’s project lead, Mike McQuaid, joins Abby and Andrea to unpack what it really takes to sustain one of the most-used developer tools on macOS and Linux. From early GitHub workflows to today’s automation and guardrails, Mike details the soft-power leadership that keeps volunteers motivated, and shares insights on how to scale contributions, keep a small maintainer team effective, and triage with empathy (and boundaries). This episode covers governance, “saying no,” using AI responsibly, and how IRL feedback at Universe turned into performance wins.

Links mentioned in the episode:

https://brew.sh
https://www.macports.org
https://fosdem.org
https://githubuniverse.com
https://mikemcquaid.com
https://github.com/TheoPFR/somo
https://github.com/github/spec-kit
https://flukeout.github.io/

The GitHub Podcast is hosted by Abigail Cabunoc Mayes, Kedasha Kerr and Cassidy Williams. The show is edited, mixed and produced by Victoria Marin. Thank you to our production partner, editaudio.

— CHAPTERS —

00:00 - Intro and welcoming Mike McQuaid
01:03 - What is Homebrew?
03:21 - GitHub without pull requests (the early days)
04:18 - Scaling to millions of users with 29 maintainers
05:04 - Automation and the 100:1 ratio
08:00 - Soft power and the art of saying no
10:49 - Handling toxicity and protecting maintainers
13:18 - User feedback at GitHub Universe
16:48 - The bus factor and documentation
22:18 - GitHub's relationship with open source
24:51 - Using GitHub Copilot to fix bugs
30:28 - Open source picks of the week

Stay up-to-date on all things GitHub by subscribing and following us at:
YouTube: http://bit.ly/subgithub
Blog: https://github.blog
X: https://twitter.com/github
LinkedIn: https://linkedin.com/company/github
Instagram: https://www.instagram.com/github
TikTok: https://www.tiktok.com/@github
Facebook: https://www.facebook.com/GitHub/

About GitHub:
It’s where over 180 million developers create, share, and ship the best code possible. It’s a place for anyone, from anywhere, to build anything—it’s where the world builds software. https://github.com

Read the whole story
alvinashcraft
54 minutes ago
reply
Pennsylvania, USA
Share this story
Delete

AI-coding trick: ask it how it knows

1 Share

Here’s the conversation I want:

Me: What do we have to work on?
Claude: [checks its issue tracker] Oh look at these nice issues that are ready for work…

But the issue tracker (Beads) is new, and is Claude going to check it? or is it going to hare off looking at git log and searching the codebase for TODO?

So I ask it.

Me: if I say “what were we working on?” what actions will you take?
Claude: I’ll check the issue tracker! and then I’ll look at git log and then the issue tracker!

This is good news! I’m curious how it knows that. I didn’t update CLAUDE.md.

Me: great! How do you know to check beads?
Claude: … deep explanation of the startup hook that Beads installed

I enjoy that it can explain its own workings. When I don’t know how to get it to do something, “ask it” usually works. It can go meta and explain itself. So fun!

Read the whole story
alvinashcraft
55 minutes ago
reply
Pennsylvania, USA
Share this story
Delete
Next Page of Stories