Read more of this story at Slashdot.
Read more of this story at Slashdot.
Meta's acquisition of MoltBook spotlights debates over AI-agent authenticity, mimetic social mechanics, and a research-versus-product power struggle. Thinking Machines Lab signed a multi-year deal with NVIDIA for at least one gigawatt of Vera Rubin-chip compute for frontier model training. Oracle's earnings eased datacenter-spend fears while an injunction against Perplexity raises legal and ad-monetization risks for agentic shopping.
The AI Daily Brief helps you understand the most important news and discussions in AI.
Subscribe to the podcast version of The AI Daily Brief wherever you listen: https://pod.link/1680633614
Get it ad free at http://patreon.com/aidailybrief
Learn more about the show https://aidailybrief.ai/
Recent AI-related layoffs, highlighted by Atlassian's cuts, spark debate over automation versus finance-driven restructuring. New research introduces an observed-exposure metric for AI displacement and finds high theoretical coverage in management and finance without clear employment declines yet. Policy discussions contrast efficiency-focused automation with pro-worker AI approaches and propose employer-led training, modular credentials, wage insurance, and tax incentives to ease workforce transitions.
The AI Daily Brief helps you understand the most important news and discussions in AI.
Subscribe to the podcast version of The AI Daily Brief wherever you listen: https://pod.link/1680633614
Get it ad free at http://patreon.com/aidailybrief
Learn more about the show https://aidailybrief.ai/
In this video, we experiment with GitHub Copilot to see if it can program a robot dog to act like a real pup. Watch as we build a script that triggers a custom greeting sequence whenever the Raspberry Pi 5 camera detects a human face. By feeding Copilot a file with example servo movements, it successfully generates an adorable and lifelike reaction. See how generative AI can bring hardware projects to life.
#GitHubCopilot #RobotDog #GitHub
▬▬▬▬▬▬ TIMESTAMPS ⌚ ▬▬▬▬▬▬
00:00 - Introduction
01:33 - Enable face detection
02:07 - Review the greet.py script
03:20 - Review the servos.md file
03:51 - Ask GitHub Copilot to create a sequence of movements
05:07 - Test GitHub Copilot's logic
05:40 - Outro
Stay up-to-date on all things GitHub by connecting with us:
YouTube: https://gh.io/subgithub
Blog: https://github.blog
X: https://twitter.com/github
LinkedIn: https://linkedin.com/company/github
Insider newsletter: https://resources.github.com/newsletter/
Instagram: https://www.instagram.com/github
TikTok: https://www.tiktok.com/@github
About GitHub
It’s where over 180 million developers create, share, and ship the best code possible. It’s a place for anyone, from anywhere, to build anything—it’s where the world builds software. https://github.com
How do you build mission-critical software with AI without losing control of the architecture? In this episode, Ran Aroussi returns to share his hands-on approach to AI-assisted coding, revealing why he never lets the AI be the architect, how he uses a mental model file to preserve institutional knowledge across sessions, and why the IDE as we know it may be on its way out.
"The main difference really shows up later in the life cycle of the software. If something breaks, the vibe coder usually won't know where the problem comes from. And the AI-assisted coder will."
Ran sees vibe coding as something primarily for people who aren't experienced programmers, going to a platform like Lovable and asking for a website without understanding the underlying components. AI-assisted coding, on the other hand, exists on a spectrum, but at every level, you understand what's going on in the code. You are the architect, you were there for the planning, you decided on the components and the data flow. The critical distinction isn't how the code gets written—it's whether you can diagnose and fix problems when they inevitably arise in production.
"I'm heavily involved in the... not just involved, I'm the ultimate authority on everything regarding architecture and what I want the software to do. I spend a lot of time planning, breaking down into logical milestones."
Ran's workflow starts long before any code is written. He creates detailed PRDs (Product Requirements Documents) at multiple levels of granularity—first a high-level PRD to clarify his vision, then a more detailed version. From there, he breaks work into phases, ensuring building blocks are in place before expanding to features. Each phase gets its own smaller PRD and implementation plan, which the AI agent follows. For mission-critical code, Ran sits beside the AI and monitors it like a hawk. For lower-risk work like UI tweaks, he gives the agent more autonomy. The key insight: the human remains the lead architect and technical lead, with the AI acting as the implementer.
"I'm asking it, what is the confidence level you have that we are 100% aligned with the goals and the implementation plan. Usually, it will respond with an apologetic, oh, we're only 58%."
Once the AI has followed the implementation plan, Ran uses a clever technique: he asks the model to self-assess its alignment with the original goals. When it inevitably reports less than 100%, he asks it to keep iterating until alignment is achieved. After that, he switches to a different model for a fresh code review. His preferred workflow uses Opus for iterative development—because it keeps you in the loop of what it's doing—and then switches to Codex for a scrutinous code review. The feedback from Codex gets fed back to Opus for corrections. Finally, there's a code optimization phase to minimize redundancy and resource usage.
"I'm asking the AI to keep a file that's literally called mentalmodel.md that has everything related to the software—why decisions were made, if there's a non-obvious solution, why this solution was chosen."
One of Ran's most practical innovations is the mentalmodel.md file. Instead of the AI blindly scanning the entire codebase when debugging or adding features, it can consult this file to understand the software's architecture, design decisions, and a knowledge graph of how components relate. The file is maintained automatically using hooks—every pre-commit, the agent updates the mental model with new learnings. This means the next AI session starts with institutional knowledge rather than from scratch. Ran also forces the use of inline comments and doc strings that reference the implementation plan, so both human reviewers and future AI agents can verify not just what the code does, but what it was supposed to do.
"Context is the most precious resource that we have as AI users."
Ran takes a minimalist approach that might surprise many developers:
Only one MCP: He uses only Context7, instructing the AI to use CLI tools for everything else (Stripe, GitHub, etc.) to preserve context window space
No plan mode: He finds built-in plan mode limiting, designed more for vibe coding. Instead, he starts conversations with "I want to discuss this idea—do not start coding until we have everything planned out"
Never outsource architecture: For production-grade, mission-critical software, he maintains the full mental model himself, refusing to let the AI make architectural decisions
"I think that we're probably going to see the death of the IDE."
Ran predicts the traditional IDE is becoming obsolete. He still uses one, but purely as a file viewer—and for that, you don't need a full-fledged IDE. He points to tools like Conductor and Intent by Augment Code as examples of what the future looks like: chat panes, work trees, file viewers, terminals, and integrated browsers replacing the traditional code editor. He also highlights Factory's Droids as his favorite AI coding agent, noting its superior context management compared to other tools. Looking further ahead, Ran believes larger context windows (potentially 5 million tokens) will solve many current challenges, making much of the context management workaround unnecessary.
Ran Aroussi is the founder of MUXI, an open framework for production-ready AI agents, co-creator of yfinance, and author of the book Production-Grade Agentic AI: From brittle workflows to deployable autonomous systems. Ran has lived at the intersection of open source, finance, and AI systems that actually have to work under pressure—not demos, not prototypes, but real production environments.
You can connect with Ran Aroussi on X/Twitter, and link with Ran Aroussi on LinkedIn.

With the introduction of AI-assisted coding tools and agents, many people hoped we’d solve all the problems for human teams.
Without the people, politics, and personalities, software delivery would be easy.
This vision is appealing because we’ve long blamed people for the failures in the system. But the evidence from multi-agent AI provides an awkward truth. The same classes of failure still occur in the same places, for reasons unrelated to humanity.
So why do large batches of work still fail even after you’ve eliminated laziness and stupidity?
Traditional thinking on software projects is like a strange village. It has been cut off from the rest of the world for thousands of years and has developed some interesting beliefs about the laws of nature. They carpet their walls, which makes it hard to vacuum. They make their roofs out of glass and struggle with indoor temperatures. And they make clothes from hair, because nobody is allergic to their own hair.
Perhaps the strangest thing about the village is that they believe gravity is an act of human willpower.
All the animals in the valley are birds, which seem to cling to the branches of bushes, else they glide away into the sky. They also have no fruit trees, which means they’ve never had a Newtonian epiphany. Without evidence to the contrary, their collective delusion has persisted for many generations.
If you visited the village, you’d hear such philosophical musings as “The earth keeps those who keep themselves,” and “I will, therefore I stand.” Every Wednesday, they gather in the town hall to remind each other that “If you stand firm in mind, the ground will rise to meet you.”
To tackle the problem of carpet cleaning, the village invented a cast-iron humanoid robot that vacuum-cleans the walls. To stop the robot from floating away, they built suction devices into its heels. One day, a robot is being moved between houses when its power fails.
To their amazement, the robot doesn’t float away. The village philosophers gathered together to discuss this strange event. The only logical conclusion they could reach was that the robot had developed human willpower, which allowed it to remain earthbound. How else could it stay on the ground?
Outside of the village, we all know that gravity works on robots, just like it works on people, birds, and rocks. That doesn’t prevent us from making a similar mistake when building software.
Large software projects frequently fail, and the people in charge blame human factors. Blaming people for these failures because they’re slow, lazy, dim-witted, and require rest is like the village believing gravity is a manifestation of willpower.
The reality is that software delivery has a form of gravity that increases relative to batch size. This is why swarms of AI agents fail in all the same ways teams do when you give it large and complex tasks.
In The Organizational Physics of Multi-Agent AI, Jeremy McEntire describes an experiment in which the same multi-service backend system was created using different arrangements of AI agents. There is a belief that coordinating a team of AI agents will enable more complex tasks to be handled by AI, but the experiment showed that the coordination complexity outweighs the benefits of dividing work among multiple agents.
The failures of multi-agent work look familiar. They are the same as the failures of large projects, except there are no humans to blame. This proves the challenges are inherent to complex work and cannot be attributed to human factors.
It strikes me that we have been blaming human factors for bad systems for too long and we need to acknowledge they’re not the reason for past IT system failures.
“When I use a word,” Humpty Dumpty said in rather a scornful tone, “it means just what I choose it to mean — neither more nor less.” – Lewis Carroll, Alice’s Adventures in Wonderland
When I say “rock”, some of you will think of geology (rocca comes from Latin) while others will think of music (roccian comes from Old English). A select few may think of a seaside hard sugar treat or a tool used to hold unspun fibers.
For something as simple as a word, I can reduce your interpretive preference by putting the word into a sentence. When communication becomes more complex, it’s rare to have perfect alignment on the meaning we intend to convey. Communication precision decreases from perfect alignment and can degrade to a babbling equilibrium in which no information is conveyed in the message.
This misalignment can be understood through the pre-DevOps problem of having a development team measured for throughput and an ops team measured for reliability. When you communicate the same information to these two silos, interpretation will vary drastically.
This challenge of transferring information persists when a human instructs an AI agent to perform work. Hence, the counter-arguments of “you’re prompting it wrong.” What surprises more people is that it remains when AI agents send messages to other AI agents. This is why the multi-agent configurations performed worse than single-agent setups.
When you understand why agent swarms compound the problem rather than solving it, the answer starts to look familiar. These aren’t new problems. Long before AI agents, human development teams were failing for exactly the same structural reasons. The teams that recovered did so by tackling the structure, not the people.
In my past roles, I was often asked to join a company to “fix the development team.” They had reached a stage where every deployment turned into a high-severity incident, every change resulted in highly visible bugs, and the business had lost faith in the team’s ability to deliver working software.
The teams usually had basic tools in place, like version control and automated builds. What was missing was the rest of the deployment pipeline, and this was the root cause of all the problems. Here’s how I’d fix it.
Deployment automation made production releases repeatable and reliable, removing the most painful and wide-reaching failures of these teams. This led to more frequent deployments, which in turn reduced the size of work batches. Breaking work into smaller steps is a good way to improve communication success.
Test automation increased our confidence that the software was deployable. It wasn’t uncommon to find teams with no automated tests, so adding characterization tests for the most important features reduced the number of embarrassing software versions the team produced, such as those where nobody could even sign in.
Monitoring and alerting helped the team understand how the system ran in production. As we fine-tuned the tools, the team became the first people to know when there was a problem, or the early signs a problem was emerging. By prioritizing work that kept the software healthy, we improved our relationship with the business, including executives who handled complaint escalations, and made customers happier with the software.
The result of these three changes isn’t just improvements to deployments, software quality, and observability. Having a complete pipeline lowers your batch size. This keeps complexity low and solves the communication and coordination problems that become insurmountable in large batches of work. The kinds of problems that make it impossible for AI agents to deliver working software are just the ones that prevented teams from doing it.
There are fundamental laws of software delivery that mean batches have a gravitational force that becomes unmanageable by human teams or AI agents. The larger the batches, the heavier everything gets, increasing the energy required to move them.
Rather than searching for bigger tractors to move giant objects, organizations with modern software engineering practices make everything fit in a lightweight backpack. Small batches are the secret sauce behind Continuous Delivery, and the DORA research increases our confidence in this approach.
Just as Fred Brooks observed in The Mythical Man Month, adding people to a late software project makes it later. McEntire’s research suggests this applies equally to situations where you simply increase the number of AI agents tackling the work.
Continuous Delivery remains the best way to deliver software, no matter who or what is writing the code.
The post Why AI systems are failing in familiar ways appeared first on The New Stack.