Sr. Content Developer at Microsoft, working remotely in PA, TechBash conference organizer, former Microsoft MVP, Husband, Dad and Geek.
147522 stories
·
32 followers

The new Xbox is not an Xbox

1 Share

Your TV is an Xbox. Your computer is an Xbox. Your phone, your crummy Android tablet, your car's infotainment display, oops all Xboxes. If Microsoft is right about the future of gaming, your game console might soon be an Xbox even if it's not an Xbox. It's an interesting idea, and raises some fascinating questions about the gaming industry. Here's one of those questions: shouldn't an Xbox play Xbox games? Here's another: are they ever going to be any good?

On this episode of The Vergecast, The Verge's Sean Hollister begins the show by telling us about his experience with two new Xboxes: the ROG Xbox Ally and Ally X. In a sea of Windows-powe …

Read the full story at The Verge.

Read the whole story
alvinashcraft
1 minute ago
reply
Pennsylvania, USA
Share this story
Delete

Amazon hopes to replace 600,000 US workers with robots, according to leaked documents

1 Share
Bipedal robots in testing phase move containers during a mobile-manipulation demonstration at Amazon’s “Delivering the Future” event at the company’s BFI1 Fulfillment Center, Robotics Research and Development Hub in Sumner, Washington on October 18, 2023.
Amazon has deployed over a million robots to its facilities, and is testing bipedal bots like Agility Robotics' “Digit” (pictured). | Image: Jason Redmond / Getty Images

Amazon is reportedly leaning into automation plans that will enable the company to avoid hiring more than half a million US workers. Citing interviews and internal strategy documents, The New York Times reports that Amazon is hoping its robots can replace more than 600,000 jobs it would otherwise have to hire in the United States by 2033, despite estimating it’ll sell about twice as many products over the period.

Documents reportedly show that Amazon’s robotics team is working towards automating 75 percent of the company’s entire operations, and expects to ditch 160,000 US roles that would otherwise be needed by 2027. This would save about 30 cents on every item that Amazon warehouses and delivers to customers, with automation efforts expected to save the company $12.6 billion from 2025 to 2027.

Amazon has considered steps to improve its image as a “good corporate citizen” in preparation for the anticipated backlash around job losses, according to The NYT, reporting that the company considered participating in community projects and avoiding terms like “automation” and “AI.” More vague terms like “advanced technology” were explored instead, and using the term “cobot” for robots that work alongside humans.

In a statement to The NYT, Amazon said the leaked documents were incomplete and did not represent the company’s overall hiring strategy, and that executives are not being instructed to avoid using certain terms when referring to robotics. We have also reached out to Amazon for comment. 

“Nobody else has the same incentive as Amazon to find the way to automate. Once they work out how to do this profitably, it will spread to others, too,” Daron Acemoglu, winner of the Nobel Prize in economic science last year, told The NYT. Adding that if Amazon achieves its automation goal, “one of the biggest employers in the United States will become a net job destroyer, not a net job creator.”

Read the whole story
alvinashcraft
1 minute ago
reply
Pennsylvania, USA
Share this story
Delete

Tell me when: Building agents that can wait, monitor, and act

1 Share
Workflow icons showing tasks, thinking, and time, linked to a person symbol on a gradient background.

Modern LLM Agents can debug code, analyze spreadsheets, and book complex travel. Given those capabilities, it’s reasonable to assume that they could handle something simpler: waiting. Ask an agent to monitor your email for a colleague’s response or watch for a price drop over several days, and it will fail. Not because it can’t check email or scrape prices. It can do both. It fails because it doesn’t know when to check. Agents either give up after a few attempts or burn through their context window, checking obsessively. Neither work. 

This matters because monitoring tasks are everywhere. We track emails for specific information, watch news feeds for updates, and monitor prices for sales. Automating these tasks would save hours, but current agents aren’t built for patience.

To address this, we are introducing SentinelStep (opens in new tab), a mechanism that enables agents to complete long-running monitoring tasks. The approach is simple. SentinelStep wraps the agent in a workflow with dynamic polling and careful context management. This enables the agent to monitor conditions for hours or days without getting sidetracked. We’ve implemented SentinelStep in Magentic-UI, our research prototype agentic system, to enable users to build agents for long-running tasks, whether they involve web browsing, coding, or external tools. 

Spotlight: AI-POWERED EXPERIENCE

Microsoft research copilot experience

Discover more about research at Microsoft through our AI-powered experience

Opens in a new tab

How it works

The core challenge is polling frequency. Poll too often, and tokens get wasted. Poll too infrequently, and the user’s notification gets delayed. SentinelStep makes an educated guess at the polling interval based on the task at hand—checking email gets different treatment than monitoring quarterly earnings—then dynamically adjusts based on observed behavior. 

There’s a second challenge: context overflow. Because monitoring tasks can run for days, context overflow becomes inevitable. SentinelStep handles this by saving the agent state after the first check, then using that state for each subsequent check.

These demonstrations capture Magentic-UI with SentinelStep at work, completing a range of tasks in a timelapse sequence. 

Core components

As the name suggests, SentinelStep consists of individual steps taken as part of an agent’s broader workflow. As illustrated in Figure 1, there are three main components: the actions necessary to collect information, the condition that determines when the task is complete, and the polling interval that determines timing. Once these components are identified, the system’s behavior is simple: every [polling interval] do [actions] until [condition] is satisfied. 

Figure 1. SentinelSteps’s three main components in Magentic-UI’s co-planning interface. 
Figure 1. SentinelSteps’s three main components in Magentic-UI’s co-planning interface. 

These three components are defined and exposed in the co-planning interface of Magentic-UI. Given a user prompt, Magentic-UI proposes a complete multi-step plan, including pre-filled parameters for any monitoring steps. Users can accept the plan or adjust as needed.

Processing

Once a run starts, Magentic-UI assigns the most appropriate agent from a team of agents to perform each action. This team includes agents capable of web surfing, code execution, and calling arbitrary MCP servers.

When the workflow reaches a monitoring step, the flow is straightforward. The assigned agent collects the necessary information through the actions described in the plan. The Magentic-UI orchestrator then checks whether the condition is satisfied. If it is, the SentinelStep is complete, and the orchestrator moves to the next step. If not, the orchestrator determines the timestamp for the next check and resets the agent’s state to prevent context overflow.

Evaluation

Evaluating monitoring tasks in real-world settings is nearly impossible. Consider a simple example: monitoring the Magentic-UI repository on GitHub until it reaches 10,000 stars (a measure of how many people have bookmarked it). That event occurs only once and can’t be repeated. Most real-world monitoring tasks share this limitation, making systematic bench marking very challenging.

In response, we are developing SentinelBench, a suite of synthetic web environments for evaluating monitoring tasks. These environments make experiments repeatable. SentinelBench currently supports 28 configurable scenarios, each allowing the user to schedule exactly when a target event should occur. It includes setups like GitHub Watcher, which simulates a repository accumulating stars over time; Teams Monitor, which models incoming messages, some urgent; and Flight Monitor, which replicates evolving flight-availability dynamics. 

Initial tests show clear benefits. As shown in Figure 2, success rates remain high for short tasks (30 sec and 1 min) regardless of whether SentinelStep is used. For longer tasks, SentinelStep markedly improves reliability: at 1 hour, task reliability rises from 5.6% without SentinelStep to 33.3% with it; and at 2 hours, it rises from 5.6% to 38.9%. These gains demonstrate that SentinelStep effectively addresses the challenge of maintaining performance over extended durations.

Figure 2. SentinelStep improves success rates on longer running tasks (1–2 hours) while maintaining comparable performance on shorter tasks.
Figure 2. SentinelStep improves success rates on longer running tasks (1–2 hours) while maintaining comparable performance on shorter tasks.  

Impact and availability

SentinelStep is a first step toward practical, proactive, longer‑running agents. By embedding patience into plans, agents can responsibly monitor conditions and act when it matters—staying proactive without wasting resources. This lays the groundwork for always‑on assistants that stay efficient, respectful of limits, and aligned with user intent.

We’ve open-sourced SentinelStep as part of Magentic-UI, available on GitHub (opens in new tab) or via pip install magnetic-ui. As with any new technique, production deployment should be preceded through testing and validation for the specific use case. For guidance on intended use, privacy considerations, and safety guidelines, see the Magentic-UI Transparency Note. (opens in new tab) 

Our goal is to make it easier to implement agents that can handle long-running monitoring tasks and lay the groundwork for systems that anticipate, adapt, and evolve to meet real-world needs. 

Opens in a new tab

The post Tell me when: Building agents that can wait, monitor, and act appeared first on Microsoft Research.

Read the whole story
alvinashcraft
2 minutes ago
reply
Pennsylvania, USA
Share this story
Delete

How to update community health files with AI

1 Share

Maintaining your project’s community health files shouldn’t get in the way of writing code. GitHub Copilot can help you update and enhance your documentation, so you can stay focused on what really matters: working on the projects that excite you most.

In this blog, we’ll touch on some of the most common community health files (focusing on README, contributor guides, and licenses) and why they’re so important for maintainers, along with actionable steps you can take to add them to your projects. ✨

What are community health files and why are they so important?

Community health files are standardized documents that help maintain a welcoming, organized, and collaborative environment in open source projects. These files communicate expectations, guide contributors, and support the overall health of a repository. They do not include technical documentation or code itself, but rather the scaffolding that supports healthy collaboration. You can typically find them in a repository’s root directory or in a special .github folder (if they need to be applied across multiple repositories).

Keeping these files up-to-date should be considered a practical investment into your project’s future and reputation, as they’re often the first touchpoint for new contributors, and their existence signals project maturity and maintainability. They not only improve transparency, consistency, and collaboration, but also help set the tone for how contributors and maintainers interact and engage productively. 

If crucial community health files are missing or outdated, everyone feels the effects. Picture this: Your open source project starts gaining traction with new contributors. They want to help, but your repository doesn’t have the right files, which leads to contributors unintentionally formatting pull requests incorrectly, opening vague issues, and even introducing security vulnerabilities—all because they didn’t know the proper procedures from the start. Now, your maintainers are overwhelmed and faced with answering the same questions over and over, while also trying to retroactively enforce standards.

It’s clear that the presence of these files helps promote efficiency and clearly communicates best practices, which in turn, creates a better environment for contributors and makes life easier for maintainers—and thanks to AI, the process doesn’t have to be manual. AI tools like GitHub Copilot, for example, can automatically detect missing or stale files, suggest updates, and even generate drafts—saving time and reducing human error. 

Here are three common types of community health files and why they’re so important for building a welcoming community (and don’t worry, we’ll tell you exactly how you can generate your own with Copilot later in this blog!): 

README
Often one of the first things a visitor sees when viewing a repository, a README.MD introduces the project and explains its purpose, along with how to get started. Intended to help remove barriers, this document gives your users crucial information they need to quickly get up and running—like what the project is, information on its features, and how to install or use it. 

CONTRIBUTOR GUIDE
A contributor guide provides guidelines on how contributors can and should participate—things like coding standards and pull request instructions. This guide tells users how they can efficiently contribute and what to expect. For instance, does the project even accept contributions? Contributor guides help set standards and expectations.

LICENSE
A license specifies the legal terms under which the project can be used, modified, and distributed. In short, it tells people how they can use your software. A common example of this type of file is the MIT License. 

Here are some other popular community health files

ISSUE/PULL REQUEST TEMPLATESStandardizes the format and information required when submitting issues or pull requests.
SECURITYProvides instructions for reporting vulnerabilities and outlines the project’s security policy.
GOVERNANCEExplains how the project is managed, including roles, responsibilities, and decision-making processes.
CODE OF CONDUCTDefines standards for how to engage in a community.
SUPPORTShares specific guidance on how others can get help with your project.
FUNDINGDisplays a sponsor button in your repository to increase the visibility of funding options for your open source project.

And while it’s not exactly considered a community health file, we wanted to give an honorable mention to… the Copilot instructions file, which is an AI configuration that complements health docs. It uses the other community health files as context and tells GitHub Copilot exactly how to interact with the codebase, including what to prioritize or avoid. This file helps ground the LLM—whether you’re using GitHub Copilot or another LLM in VS Code, on github.com, or Copilot coding agent—giving it an understanding of what your project is and how it’s structured, allowing for consistency across your codebase. 

Having these kinds of files in your project is so important, especially when it comes to scaling open source projects where maintainers probably don’t have time to personally help every contributor.

That’s where time-saving tools like GitHub Copilot come in handy. Keep on reading for actionable next steps, tips, and tutorials on the most efficient ways to add these files to your repositories. ✨

Starter kit: How to update community health files using GitHub Copilot

We created a starter kit for you that explains how you can use AI to add these valuable files to your projects, complete with prompting best practices, a checklist full of things to consider, and step-by-step tutorials on how to add three common files to your repository using Copilot. Let’s dive in. 

Part one: Prompting

Whether you’re starting from scratch or refining existing documentation, GitHub Copilot can help you write clearer, more consistent community health files with just a few prompts. 

One thing to note: The LLMs powering GitHub Copilot are nondeterministic, which means that you can receive different outputs each time you prompt the model. Prompt engineering can drastically improve the quality and relevance of the outputs you get from an LLM, but you’ll still want to verify the accuracy of these outputs, especially when using Copilot to generate more sensitive files like licenses that have legal weight.  

Part two: Checklist

This checklist helps ensure that Copilot-generated content is accurate, inclusive, secure, and aligned with your project’s goals.

🔍 Before you start

  • Have you reviewed existing community health files in similar or related repositories?
  • Do you have clear goals for what each file should communicate (e.g., onboarding, behavior expectations, security reporting)?
  • Are you familiar with your organization’s GitHub usage policies and branding guidelines?

🧠 Prompting Copilot effectively

  • Are your prompts specific and contextual? (e.g., “Generate a CONTRIBUTING.md for a Python-based open source project with a code style guide.”)
  • Have you included examples or tone preferences in your prompt? (e.g., “Use inclusive language and a welcoming tone.”)

🛡️ Security & privacy

  • Are you avoiding prompts that include sensitive or proprietary information (e.g., internal credentials, private URLs, confidential project names)?
  • Have you reviewed your repository’s visibility settings (public vs. private) and ensured that community health files are appropriate for that audience?
  • Are you familiar with GitHub Copilot’s privacy settings and how your prompts and suggestions are handled?
  • Will your SECURITY.md include:
    • A clear contact method for reporting vulnerabilities?
    • A brief explanation of how security issues are triaged?
    • Any relevant links to your organization’s responsible disclosure policy?

🧾 Reviewing Copilot output

  • Does the generated content reflect your project’s values and community standards?
  • Have you checked for hallucinated links, names, or policies that don’t exist?
  • Are all references to external resources accurate and up-to-date?

🧪 Testing & feedback

  • Have you asked a teammate or contributor to review the generated files?
  • Have you tested any instructions (e.g., setup steps in README or CONTRIBUTING) to ensure they work?
  • Are you open to iterating based on community feedback?

Part three: Tutorial

In this tutorial, we’ll walk through how you can use Copilot to quickly and easily update README.md, a LICENSE file, and CONTRIBUTING.md.

📝 Create a README 

Why make a README? Adding a README provides a clear overview of your project, helping users and contributors quickly understand its purpose, setup, and usage. Without it, potential users could abandon your repository due to confusion or lack of context.

Here’s how to make one: 

  1. Open GitHub Copilot Chat in your IDE (e.g., VS Code).
  2. Switch to agent mode to enable project-aware assistance.
  3. Select your preferred model (e.g., Claude for strong writing and coding support).
  4. Ensure your project is open in the IDE so Copilot can read its context (e.g., package.json, app.tsx).
  5. In the chat window, type: “Help me write a README.md for my project. Ensure it includes installation instructions, a project overview, and follows standard README practices.”
  6. Review the generated README.md. Copilot will analyze your project files and generate a structured README.md.
  7. Validate the installation instructions manually to ensure accuracy (LLMs may hallucinate).
  8. If satisfied, click “Keep” to save the README.md file.
  9. Commit the README.md to your repository.

📄 Add a license

Why make a license? A license defines how others can legally use, modify, and distribute your code, protecting both your rights and theirs. It removes ambiguity and prevents misuse, making your project safer to adopt and contribute to.

Here’s how to add one: 

  1. Open GitHub Copilot Chat in your IDE.
  2. Decide what kind of license you want to add.
  3. Type the following prompt: “Can you add [the license you want] to my project?”
  4. Copilot will generate a LICENSE file with the license of your choice. 
  5. Review the license to ensure it’s accurate (especially any copyright owner names and statements).
  6. If correct, click “Keep” to save the file.
  7. Commit the LICENSE file to your repository.

🤝  Create a contributor guide

Why make a contributor guide? A contributor guide streamlines collaboration by outlining contribution standards, workflows, and expectations. This makes it easier for others to get involved with your project. The goal is to reduce friction and errors while also encouraging consistent, scalable contributions.

Here’s how to create one: 

  1. Open GitHub Copilot Chat in your IDE.
  2. Click the “+” icon to start a new chat.
  3. Type this prompt: “Create a contributing guide file that follows best practices and link it in the README.”
  4. Copilot will generate a CONTRIBUTING.md file with:
    • Contribution guidelines
    • Code standards
    • Pull request instructions
    • Issue reporting process
  5. Review and edit the guide to match your team’s workflow.
  6. Save and commit the CONTRIBUTING.md file.
  7. Update your README to include a link to the contributor guide:
## Contributing
See CONTRIBUTING.md for guidelines.

Take this with you

GitHub Copilot isn’t just for writing code—it can be your documentation sidekick, too. Helping you write smarter, faster, and with less friction, Copilot sharpens your community health files, scales best practices, and turns good intentions into great documentation. 

The result? Better docs, stronger communities, and happier maintainers.

Read the Docs to learn more about GitHub Copilot features or get started today.

The post How to update community health files with AI appeared first on The GitHub Blog.

Read the whole story
alvinashcraft
2 minutes ago
reply
Pennsylvania, USA
Share this story
Delete

Inside the Trump Adminstration's Bluesky Invasion

1 Share
On Friday, after months of internal discussions, federal agencies began posting on the left-friendly social network. Within days, they dominated a list of the most-blocked accounts.
Read the whole story
alvinashcraft
2 minutes ago
reply
Pennsylvania, USA
Share this story
Delete

Start learning all things AI on the new Google Skills

1 Share
Google Skills is a new home for building skills in AI, and learning about other topics like data analytics and security.
Read the whole story
alvinashcraft
2 minutes ago
reply
Pennsylvania, USA
Share this story
Delete
Next Page of Stories