Sadaf Mozaffarian, WW Public Sector Solutions Lead, Microsoft, says Governments do not need more AI experiments but decisions they can explain in public, act on safely, and defend in an audit.

Across government, enthusiasm for artificial intelligence (AI) is no longer the constraint. Capability is advancing quickly. Pilots are everywhere. Yet many public leaders are reaching the same conclusion: clever demonstrations are not the same as operational outcomes
Governments do not need more AI experiments. They need decisions they can explain in public, act on safely, and defend in an audit.
That requires a shift – from isolated AI pilots to decision-ready intelligence: insight that is connected to action, grounded in evidence, and governed by design. The differentiator is not bigger models or more complex demos. It is whether a city can connect the intelligence it already has – operational data, institutional knowledge, and real work context. When those signals come together, government leaders and frontline teams can make holistic decisions with tangible outcomes for residents, backed by a single, accountable decision chain.
When essential services fail – water systems, roads, permitting, inspections, benefits – governments do not get a quiet postmortem. They get audit scrutiny, media attention, parliamentary questions, and frustrated residents.
In that environment, an AI system that produces an answer without context, traceability, or controls does not reduce risk. It adds a new one. Public leaders are therefore making a pragmatic pivot. The question is no longer “Can AI generate insight?” It is “Can this insight be operationalised, governed, explained, and scaled?”
Decision-ready intelligence connects inputs, outputs, and controls into a coherent operating model – enabled by a platform that can unify data, apply permissions, and carry governance from insight through execution:
This model recognises a simple truth: in government, a response is not the same as a resolution. Leaders need intelligence that moves seamlessly from signal to decision to action – without breaking accountability.
Public sector teams aren’t short on ideas. They face persistent constraints: fragmented data, uneven governance, and limited operational visibility across departments and partners. Those constraints show up in many ways – from slower decision-making and duplication of effort to inconsistent service and avoidable risk.
At the centre is an information problem: government knowledge is distributed across systems, teams, and formats, with different definitions, owners, and access rules. Even when data exists, it is often hard to find, hard to trust, and hard to use in the flow of work. Two common layers illustrate the challenge:
When these layers – and the silos between departments – remain disconnected, AI can produce answers in isolation but cannot reliably support defensible decisions or safe execution at scale. Meaning is not shared. Workflows remain fragmented. And trust questions – privacy, security, bias, records obligations – arrive late, when they are hardest to fix. Closing the intelligence gap reduces the risk of getting stuck between a promising proof point and an operational capability leaders can stand behind.
Cities provide a powerful lens because services are tangible, time-critical, and cross-departmental by nature. When intelligence is decision-ready, it shows up in faster response, clearer accountability, and better resident outcomes.
Mission 1: Asset reliability and infrastructure health
What leaders want: Fewer outages and disruptions, safer and more efficient operations (including inspection and maintenance), and longer-lived assets across roads, water, facilities, and fleets.
A decision-ready moment: a water main break triggers alarms, field reports, and a spike in resident contacts. Within minutes, leaders need a single operating picture: what is affected, what is at risk, what is already in motion, and what decision is required now.
Operators, meanwhile, need clear work orders, safe isolation steps, traffic coordination, approved public messaging, and context that makes response safer and faster (for example, the last inspection notes, prior repairs, known failure modes, and what crews learned the last time this asset was worked on).
What gets logged: The source signals used, the recommended plan, the human approval, the actions taken, and the communications issued.
This is intelligence designed not just to inform, but to manage physical infrastructure under scrutiny.
Mission 2: Citizen service and experience
What leaders want: Fast, clear, and inclusive service across 311, permitting, housing, and other frontline interactions.
A decision-ready moment: a resident seeks emergency rent assistance after a job loss and contacts the city through 311, a walk-in centre, or the website – in a language that is not the default for the service team. Leaders need confidence that eligibility guidance is accurate and consistent, that handoffs to housing partners are tracked, and that the experience reduces repeat contacts and missed deadlines – not creating new backlogs. Operators need address history, route status, policy criteria, and a clear escalation path when judgment is required.
What gets logged: The policy basis for the response, the service data used, any escalation, and the final outcome.
Here, decision-ready intelligence improves trust by making service both faster and fairer – without losing accountability.
Mission 3: Planning and growth management
What leaders want: Transparent planning decisions that balance equity, safety, and economic growth.
A decision-ready moment: City leaders must choose between competing infrastructure investments – for example, a transit corridor upgrade, flood mitigation works, or a housing-enabling streets package – each with different costs, delivery risks, and community impacts. Leaders need a defensible narrative: what evidence was used, which policies and plans apply, what trade-offs exist (economic development, safety, livability, equity), and what risks must be managed. Operators need a decision packed with precedents, statutory guidance, and consultation themes – linked back to source documents for transparency.
What gets logged: To support explainability and defensibility, the system records relevant signals – such as evidence used, consultation inputs, applicable requirements, and decision rationale – based on how the environment is configured. This information is captured in audit logs and retained in line with the organisation’s configured records and retention policies. Access is managed through defined roles and permissions, and where enabled, authorised teams can review this information through Microsoft Purview.
This is how planning decisions move faster while standing up to scrutiny.
Once leaders align outcomes and audit requirements, the operating model becomes clearer: connect context to decisions to actions, with controls built in.
In Microsoft’s approach, this is supported by three complementary layers that ground AI in organisational reality:
Together, these layers help close the loop from context to decision to action, while preserving traceability across people, data, and AI-driven workflows. Productivity copilots support staff in the flow of work, data platforms unify operational signals, and Agent365 helps orchestrate governance, monitoring, and administrative controls across AI agents built in Copilot Studio, extended through custom development, or connected to third-party systems. The result is not a product catalogue, but a governed decision chain designed to help agencies apply consistent security practices, maintain audit logging, and support compliance requirements as they scale from individual copilots to enterprise-wide agent ecosystems.
Practically, adopting this model is a transition, not a switch. Cities typically start by naming a mission outcome, selecting a small number of “moments of truth,” and then connecting the minimum viable set of work context, operational data, and policy knowledge needed to make those moments decision-ready. From there, teams harden governance (permissions, logging, retention) and scale by reusing the same patterns – data products, knowledge bases, and agent workflows – across departments.
Public trust is hard won. Leaders should expect – and design for – healthy scepticism:
Microsoft’s responsible AI principles – fairness, reliability and safety, privacy and security, inclusiveness, transparency, and accountability – are most valuable when they translate into operational answers, not just values statements.
Governments retain control of their data, and systems are designed to respect defined boundaries and sovereignty requirements.
To move from interest to impact:
AI in government will be judged the same way every public investment is judged: by outcomes, fairness, and public confidence. Decision-ready intelligence is how cities close the intelligence gap – turning AI from experimentation into accountable, operational impact.