Orply.

GitHub Agentic Workflows Turn Actions Into AI-Run Development Processes

Yash LaraPeli HalleuxMicrosoft ResearchThursday, May 14, 20265 min read

Microsoft Research’s Peli Halleux and Yash Lara present GitHub Agentic Workflows as a move from AI-assisted coding to repository-level process automation. Their argument is that agents should be embedded inside GitHub Actions to research, plan, assign, and open pull requests under human review, rather than operate as unconstrained swarms. The system’s promised scale depends on orchestration, sandboxing, limited permissions, and Microsoft-hosted models on Azure.

The claim is not faster coding; it is process automation at repository scale

Yash Lara introduced GitHub Agentic Workflows as an attempt to make a repository “run itself”: handle issues, automation, and workflows end to end without brittle scripts or manual glue. The system brings AI agents into GitHub Actions, uses built-in guardrails, and runs Microsoft-hosted models on Azure.

Peli Halleux described the project as a collaboration among Microsoft Research, GitHub, GitHub Next, and Azure Core. Its ambition is broad: use agents to automate “the entire spectrum of the software development lifecycle,” including testing, code generation, documentation, and other repository work.

The important distinction is between individual productivity and process multiplication. A slide contrasted three levels: completion at “1.4x,” a solo developer at “10x,” and automation at “100-1000x.” Microsoft Research’s case is that automation, not assisted coding alone, is “the true accelerator of agentic transformation.”

100-1000x
multiplication attributed to automation in the presentation’s framing

That scale is not supposed to come from unleashing many independent agents into a repository. Halleux explicitly rejected “swarms and chaos” as the model. The intended model is “order and orchestration,” built around processes.

The repository becomes an orchestrated process, not just a place to answer prompts

GitHub Agentic Workflows combines three layers: automation through GitHub Actions, safety through a sandbox, and reasoning through GitHub Copilot CLI. Together, those layers are meant to support “agentic human processes powered by GitHub.”

The phrase “agentic human processes” matters. The system is not presented as a fully autonomous replacement for development teams. The example includes multiple agents and multiple human decision points. Agents research, transform reports into work items, assign tasks, and open pull requests. Humans inspect findings, decide whether to continue, comment, review, and merge.

The concrete example was a “Research - Plan - Assign” workflow. It begins with Deep Research running on a schedule inside the repository. In Halleux’s example, the agent tries to solve a problem such as finding duplicate code and produces a report.

A developer reviews the report. If the findings look useful, the developer starts another agent to turn the report into work items. An automated agent called Issue Monster then picks up issues tagged with “cookie” — Halleux noted that it “loves to eat cookies” — and assigns them to Copilot. Copilot converts those tasks into code through the usual review loop: the developer can add comments and reviews; the work eventually becomes a pull request and may be merged.

StageActorOutput or action
ResearchDeep ResearchResearches in the repository and creates an artifact
PlanDeveloper and another agentDeveloper inspects the artifact; if useful, another agent turns the report into work items
AssignIssue Monster, Copilot, and developerIssue Monster picks up tagged issues and assigns them to Copilot; Copilot works on the task, opens a pull request, and the developer reviews and merges
The Research-Plan-Assign workflow shown for GitHub Agentic Workflows

The example shows the intended unit of automation. The agent is not just answering a prompt. It is embedded in a repository process with triggers, artifacts, issue labels, assignments, pull requests, and human checkpoints. GitHub Actions supplies the event-driven automation layer; the agent supplies reasoning over the task; and the sandbox constrains what the agent can access, do, and affect.

Safety is treated as a first-order design constraint

Peli Halleux emphasized that agents are dangerous when left unattended. In an agentic repository process, untrusted content can enter through pull requests, issues, web pages, web queries, or other inputs. Those inputs can contain adversarial strings intended to take over an agent.

Agents are dangerous if they are left unattended.

Peli Halleux

The security diagram named several threat categories: prompt injection, permissions abuse, data exfiltration, and malware injection. It also identified the kinds of incoming material that can carry risk: pull requests, issues, web pages, and other inputs.

The proposed response is a sandbox that enforces what the agent can “access, do, and affect.” The diagram described a set of constraints and controls: read-only permissions, zero secrets, containers and firewalls, and integrity filtering. Safe outputs included creating an issue, creating a pull request, and sending a message.

AreaRisk, control, or boundary shown
ThreatsPrompt injection, permissions abuse, data exfiltration, and malware injection
Untrusted inputsPull requests, issues, web pages, and other inputs can carry adversarial strings
Sandbox enforcementConstrains what the agent can access, do, and affect
Permissions and secretsRead-only permissions and zero secrets
Execution and filteringContainers, firewalls, and integrity filtering
Allowed outputsCreate issue, create pull request, and send message
The sandbox model shown for constraining unattended or semi-attended agents

Halleux said Microsoft Research had invested in “a lot of layers of security” and described safety as a central aspect of agentic workflows.

Authoring moves from YAML to Markdown prompts

To support agents, the project changes the workflow authoring format. GitHub Actions YAML has been converted into Markdown, “which is very popular for agents.”

The file format shown has front matter at the top containing the older Actions-style configuration plus “agentic stuff,” followed by a prompt. The screenshot used an issue-clarifier.md file under .github/workflows. Its visible content showed a trigger for opened issues, read-only permissions by default, an allowed write operation to add a comment, and a prompt instructing the agent to analyze the current issue and ask for additional details if the issue is unclear.

The same screenshot showed a mobile chat interface asking: “Create a workflow that asks for clarification on unclear issues.” Users are not expected to edit these files manually. The workflow itself can be authored through an agent, using natural language.

Halleux also suggested that interaction with agents will not require a laptop. It could happen from a phone or through voice. The development artifact remains a repository workflow, but the interface for creating and modifying it becomes conversational.

The larger target is information work, not only software development

The repository is the first setting for the model rather than the boundary of the ambition. Peli Halleux widened the scope beyond software development, arguing that agentic human processes will change “any process that uses information and uses reasoning.”

That includes non-developers in marketing, sales, and operations. The claim is that anyone using a Copilot today for reasoning may eventually want to automate that Copilot. The visual connected agents to organizational functions such as marketing, sales, and operations, alongside platform logos including GitHub, Salesforce, SAP, and Dynamics 365. It also included the constraint: “Agents can only reason about what’s accessible.”

GitHub Agentic Workflows is presented as open source, with the project available at github.github.com/gh-aw.

The frontier, in your inbox tomorrow at 08:00.

Sign up free. Pick the industry Briefs you want. Tomorrow morning, they land. No credit card.

Sign up free