Microsoftโs move to end Windows 10 support on 14 October 2025 has created an inflection point for CIOs. Three million PCs in active use in Australia canโt upgrade to Windows 11, according to Microsoft, which leaves many organisations facing security risk and a widening gap in frontline productivity.
Most corporate AI adoption so far has been cloud-led, but concerns about privacy, rising AI service costs and the movement of sensitive client data offshore are prompting a rethink. Local inferencing on modern hardware now offers a practical way to accelerate AI use without increasing exposure.
For enterprise IT leaders, the risk is not simply running out-of-support devices and the security risk that entails. The deeper issue is falling behind in the shift to AI-assisted work. If outdated hardware limits access to new capabilities in Windows 11, productivity losses accumulate quietly but consistently across frontline teams.
Why legacy approaches fall short
Relying solely on cloud AI services appears convenient, but it comes with constraints for regulated sectors. Sensitive information still travels outside the organisation, model performance depends on connectivity, and costs can climb quickly with usage.
A like-for-like hardware refreshโswapping out PCs to new minimum spec models that can run Windows 11โalso offers little uplift. It wonโt enable AI acceleration, intelligent automation, or the performance for larger applications that include local large language models (LLMs).
As organisations become more dependent on real-time summarisation, content creation, translation and threat detection, traditional processors will struggle to keep pace. These limitations make a hardware-based step change more compelling. Organisations are now looking for devices designed for AI workloads to avoid incremental fixes that will push cost and risk into later years.
How AI PCs address the problem
Modern AI PCs pair Windows 11 with AI-specific processors, including neural processing units (NPUs) capable of more than 40 trillion operations per second. They support local generation, summarisation and automation tasks without sending data to offshore services, reducing compliance exposure and cutting response times.
On-device AI helps streamline everyday work. Tasks such as meeting follow-ups, document clean-ups and content drafting can happen in the background, freeing staff to focus on higher-value decisions. These small increments add up when multiplied across thousands of employees.
AI PCs also support accessibility capabilities, such as real-time translation, on-device captioning and voice-driven interfaces, helping more employees engage effectively with digital tools.
Analyst firm Canalys forecasts that 60 per cent of PCs shipped in 2027 will be AI-capable, more than triple the volume shipped in 2024. This shift reflects a broader recognition that frontline productivity depends on pairing AI software with suitable hardware.
Lenovoโs approach to AI-driven computing
Lenovoโs Aura Edition Copilot+ AI PC portfolio, powered by Intel Core Ultra processors, is designed around this need for secure, local AI acceleration.
Silke Barlow, Lenovoโs Australian country manager, says the aim is to make AI immediately useful. โIf a device can quietly draft documents, take notes and tidy data while the user focuses on making decisions that really take advantage of their skill set, thatโs real productivity,โ he explains.
These processors separate AI tasks from traditional system operations, improving responsiveness even when multiple applications are active. For creative and analytical teams, the hardware also supports faster rendering, predictive modelling and advanced editing tools.
Beyond the device itself, Lenovo positions AI PCs as part of a broader digital workplace strategy. Its Smart Care uses predictive analytics to anticipate hardware issues before they become outages, reducing support overhead and improving uptime. Lenovo Smart Modes adjusts system behaviour automatically as employees move between tasks, optimising performance without manual tuning.
For organisations scaling AI workflows, the Lenovo AI Digital Workplace Solution integrates device-level capabilities with broader collaboration, security and management tools, allowing IT teams to operationalise new ways of working more smoothly.
Barlow says Lenovo sees AI PCs as a foundation for long-term capability building. โWindows 11 and AI-optimised hardware are reshaping how people work. Weโve invested in a new ecosystem that helps organisations transition at a pace that strengthens security and productivity together.โ
Making the upgrade decision
Past operating system transitions show that delaying replacements increases cost and complexity. Unsupported devices require more IT effort, are harder to secure and limit access to new features that drive workplace innovation.
A structured upgrade path helps. Many organisations begin by identifying devices that cannot move to Windows 11 and prioritising business units where AI-enabled automation can remove repeatable work. This reduces risk while improving margins and freeing teams to focus on more specialised tasks.
The end of Windows 10 support is prompting a broader reassessment of how PCs contribute to productivity, compliance and resilience. AI PCs extend beyond solving an immediate problem: they provide a platform for sustained improvement in how work gets done.
Planning your fleet refresh? Learn more about how Lenovo Aura Edition Copilot+ PCs offer personalised, productive, and protected AI with the latest Intel Core Ultra processors.
Deadline to Submit:ย January 14, 2026 |ย Nominate Now
The annual US CIO 100 Awards,ย entering its 28th year, celebrates 100 organizations and the IT teams within them that use technology in innovative ways to deliver business value, whether by creating competitive advantage, optimizing business processes, enabling growth, or improving relationships with customers. The award is an acknowledged mark of enterprise excellence.
Winning a CIO 100 Award signals to the industryโand to your organizationโthat your team is delivering true business value through innovation. It elevates your companyโs brand, strengthens talent attraction and retention, and showcases your leadershipโs commitment to transformation. And because the award is given to companies rather than individuals, itโs an honor that entire teams may enjoy.
Winners will be recognized at the CIO 100 Symposium & Awards at the Omni PGA Frisco Resort & Spa in Frisco, TX, from August 17-19, 2026.
At the conference, finalists and winners take the spotlight among a national community of CIOs and technology leaders. Itโs a rare opportunity to share your story, learn from the yearโs most impactful initiatives, and connect with peers who are redefining whatโs possible in enterprise IT.
The CIO 100 Awards are an acknowledged mark of enterprise excellence in business technology. The awards are given annually to 100 IT organizations in companies from a range of industries, from financial services to manufacturing, health care, higher ed, and more.
2.
What are the benefits of winning the CIO 100?
Winning a CIO 100 awards generates positive PR and provides tangible, companywide recognition of the IT organizationโs hard work and accomplishments.
The CIO 100 award is given to companies rather than individuals, so it is an honor everyone on your staff can take pride in receiving. Executives from the winning companies will be recognized among their peers and colleagues at the annual CIO 100 Symposium & Awards in August 2025. Winning organizations will have access to a Press Release Guide from CIO, including sample copy and quotes from CIOโs Editors. The awards also generate coverage on CIO.com, which lists the winning companies and judges.
3.
What is the CIO 100 Symposium?
CIO 100 Symposium is a conference produced by the award winning media brand, CIO.com, this exclusive event is your gateway to three days of transformative insights, unparalleled networking, and recognition of groundbreaking achievements. The conference concludes with an awards ceremony celebrating the CIO 100 winners and Hall of Fame inductees.
4.
Who are some past winners?
Aflac, Deloitte, Johnson & Johnson, TIAA, Ulta Beauty, UPS, Verizon, UC San Diego, and Nationwide are some of the well-known companies that won a CIO 100 Award in 2025. View the full list of 2025 winners.
5.
Is there a fee for entry?
Yes, the cost of entry is $50.ย If you plan to submit multiple projects, you may purchase up to 10 applications at a time. Note that your organization can win only one award in a given year.
6.
What are the eligibility requirements?
Any US based projects that have produced an internally beneficial technology or service is eligible to be nominated. Projects must be at least in a pilot stage; and have already delivered early findings/results.
Technology vendors can apply for the award, but this is not an award for IT vendorsโ products. If the outcome of your submitted IT project (i.e., the business benefit) is to produce a technology product or service sold directly to IT buyers, it will be disqualified from consideration. However, if your innovative project produces an internally beneficial technology or service (ie, one not designed be sold to customers), you are welcome to apply for a CIO 100 Award.
7.
Who can nominate?
Technology leaders, directors and executives, other qualified team members, as well as internal or external PR representatives may nominate a company or organization.
Technology vendors and their PR representatives are also invited to submit nominations, which can be a great way to recognize your clients and their customers. The nomination must be on behalf of the customerโs project/initiative. Technology vendors should not nominate any technology product or service they sell to external customers or clients.
Last quarter, during a board review, one of our directors asked a question I did not have a ready answer for. She said, โIf an AI-driven system takes an action that impacts compliance or revenue, who is accountable: the engineer, the vendor or you?โ
The room went quiet for a few seconds. Then all eyes turned toward me.
I have managed budgets, outages and transformation programs for years, but this question felt different. It was not about uptime or cost. It was about authority. The systems we deploy today can identify issues, propose fixes and sometimes execute them automatically. What the board was really asking was simple: When software acts on its own, whose decision is it?
That moment stayed with me because it exposed something many technology leaders are now feeling. Automation has matured beyond efficiency. It now touches governance, trust and ethics. Our tools can resolve incidents faster than we can hold a meeting about them, yet our accountability models have not kept pace.
I have come to believe that this is redefining the CIOโs role. We are becoming, in practice if not in title, the chief autonomy officer, responsible for how human and machine judgment operate together inside the enterprise.
Autonomy rarely begins as a strategy. It arrives quietly, disguised as optimization.
A script closes routine tickets. A workflow restarts a service after three failed checks. A monitoring rule rebalances traffic without asking. Each improvement looks harmless on its own. Together, they form systems that act independently.
When I review automation proposals, few ever use the word autonomy. Engineers frame them as reliability or efficiency upgrades. The goal is to reduce manual effort. The assumption is that oversight can be added later if needed. It rarely is. Once a process runs smoothly, human review fades.
Many organizations underestimate how quickly these optimizations evolve into independent systems. As McKinsey recently observed, CIOs often find themselves caught between experimentation and scale, where early automation pilots quietly mature into self-operating processes without clear governance in place.
This pattern is common across industries. Colleagues in banking, health care and manufacturing describe the same evolution: small gains turning into independent behavior. One CIO told me their compliance team discovered that a classification bot had modified thousands of access controls without review. The bot had performed as designed, but the policy language around it had never been updated.
The issue is not capability. It is governance. Traditional IT models separate who requests, who approves, who executes and who audits. Autonomy compresses those layers. The engineer who writes the logic effectively embeds policy inside code. When the system learns from outcomes, its behavior can drift beyond human visibility.
To keep control visible, my team began documenting every automated workflow as if it were an employee. We record what it can do, under what conditions and who is accountable for results. It sounds simple, but it forces clarity. When engineers know they will be listed as the manager of a workflow, they think carefully about boundaries.
Autonomy grows quietly, but once it takes root, leadership must decide whether to formalize it or be surprised by it.
Where accountability gaps appear
When silence replaces ownership
The first signs of weak autonomy are subtle. A system closes a ticket and no one knows who approved it. A change propagates successfully, yet no one remembers writing the rule. Everything works, but the explanation disappears.
When logs replace memory
I saw this during an internal review. A configuration adjustment improved performance across environments, but the log entry said only executed by system. No author, no context, no intent. Technically correct, operationally hollow.
Those moments taught me that accountability is about preserving meaning, not just preventing error. Automation shortens the gap between design and action. The person who creates the workflow defines behavior that may persist for years. Once deployed, the logic acts as a living policy.
When policy no longer fits reality
Most IT policies still assume human checkpoints. Requests, approvals, hand-offs. Autonomy removes those pauses. The verbs in our procedures no longer match how work gets done. Teams adapt informally, creating human-AI collaboration without naming it and responsibility drifts.
There is also a people cost. When systems begin acting autonomously, teams want to know whether they are being replaced or whether they remain accountable for results they did not personally touch. If you do not answer that early, you get quiet resistance. When you clarify that authority remains shared and that the system extends human judgment rather than replaces it โ adoption improves instead of stalling.
Making collaboration explicit
To regain visibility, we began labeling every critical workflow by mode of operation:
Human-led โ people decide, AI assists.
AI-led โ AI acts, people audit.
Co-managed โ both learn and adjust together.
This small taxonomy changed how we thought about accountability. It moved the discussion from โwho pressed the button?โ to โhow we decided together.โ Autonomy becomes safer when human participation is defined by design, not restored after the fact.
How to build guardrails before scale
Designing shared control between humans and AI needs more than caution. It requires architecture. The objective is not to slow automation, but to protect its license to operate.
Define levels of interaction
We classify every autonomous workflow by the degree of human participation it requires:
Level 1 โ Observation: AI provides insights, humans act.
Level 2 โ Collaboration: AI suggests actions, humans confirm.
Level 3 โ Delegation: AI executes within defined boundaries, humans review outcomes.
These levels form our trust ladder. As a system proves consistency, it can move upward. The framework replaces intuition with measurable progression and prevents legal or audit reviews from halting rollouts later.
Create a review council for accountability
We established a small council drawn from engineering, risk and compliance. Its role is to approve accountability before deployment, not technology itself. For every level 2 or level 3 workflow, the group confirms three things: who owns the outcome, what rollback exists and how explainability will be achieved. This step protects our ability to move fast without being frozen by oversight after launch.
Build explainability into the system
Each autonomous workflow must record what triggered its action, what rule it followed and what threshold it crossed. This is not just good engineering hygiene. In regulated environments, someone will eventually ask why a system acted at a specific time. If you cannot answer in plain language, that autonomy will be paused. Traceability is what keeps autonomy allowed.
Over time, these practices have reshaped how our teams think. We treat autonomy as a partnership, not a replacement. Humans provide context and ethics. AI provides speed and precision. Both are accountable to each other.
In our organization we call this a human plus AI model. Every workflow declares whether it is human-led, AI-led or co-managed. That single line of ownership removes hesitation and confusion.
Autonomy is no longer a technical milestone. It is an organizational maturity test. It shows how clearly an enterprise can define trust.
The CIOโs new mandate
I believe this is what the CIOโs job is turning into. We are no longer just guardians of infrastructure. We are architects of shared intelligence defining how human reasoning and artificial reasoning coexist responsibly.
Autonomy is not about removing humans from the loop. It is about designing the loop on how humans and AI systems trust, verify and learn from each other. That design responsibility now sits squarely with the CIO.
That is what it means to become the chief autonomy officer.
This article is published as part of the Foundry Expert Contributor Network. Want to join?
If 2024 was the year of experimentation and 2025 the year of the proof of concept, then 2026 is shaping up to be the year of scale or fail.
Across industries, boards and CEOs are increasingly questioning whether incumbent technology leaders can lead them to the AI promised land. That uncertainty persists even as many CIOs have made heroic efforts to move the agenda forward, often with little reciprocation from the business. The result is a growing imbalance between expectation and execution.
So what do you do when AI pilots arenโt converting into enterprise outcomes, when your copilot rollout hasnโt delivered the spontaneous innovation you hoped for and when the conveyor belt of new use cases continues to outpace the limited capacity of your central AI team? For many CIOs, this imbalance has created an environment where business units are inevitably branching off on their own, often in ways that amplify risk and inefficiency.
Leading CIOs are breaking this cycle by tackling the 2026 agenda on two fronts, beginning with turning IT into a productivity engine and extending outward by federating AI delivery across the enterprise. Together, these two approaches define the blueprint for taking back the AI narrative and scaling AI responsibly and sustainably.
Inside out: Turning IT into a productivity engine
Every CEO is asking the same question right now: Whereโs the productivity? Many have read the same reports promising double-digit efficiency gains through AI and automation. For CIOs, this is the moment to show what good looks like, to use IT as the proving ground for measurable, repeatable productivity improvements that the rest of the enterprise can emulate.
The journey starts by reimagining what your technology organization looks like when itโs operating at peak productivity with AI. Begin with a job family analysis that includes everyone: Architects, data engineers, infrastructure specialists, people managers and more. Catalog how many resources sit in each group and examine where their time is going across key activities such as development, support, analytics, technical design and project management. The focus should be on repeatable work, the kind of activities that occur within a standard quarterly cycle.
For one Fortune 500 client, this analysis revealed that nearly half of all IT time was being spent across five recurring activities: development, support, analytics, technical design and project delivery. With that data in hand, the CIO and their team began mapping where AI could deliver measurable improvements in each job familyโs workload.
Consider the software engineering group. Analysis showed that 45% of their time was spent on development work, with the rest spread across peer review, refactoring and environment setup, debugging and other miscellaneous tasks. Introducing a generative AI solution, such as GitHub Copilot enabled the team to auto-generate and optimize code, reducing development effort by an estimated 34%. Translated into hard numbers, that equates to roughly six hours saved per engineer each week. Multiply that by 48 working weeks and 100 developers and the result is close to 29,000 hours, or about a million dollars in potential annual savings based on a blended hourly rate of $35. Over five years, when considering costs and a phased adoption curve, the ROI for this single use case reached roughly $2.4 million
Repeating this kind of analysis across all job families and activities produces a data-backed productivity roadmap: a list of AI use cases ranked by both impact and feasibility. In the case of the same Fortune 500 client, more than 100 potential use cases were identified, but focusing on the top five delivered between 50% and 70% of the total productivity potential. With this approach, CIOs donโt just have a target; they have a method. They can show exactly how to achieve 30% productivity gains in IT and provide a playbook that the rest of the organization can follow.
Outside in: Federating for scale
If the inside-out effort builds credibility, the outside-in effort lays the foundation to attack the supply-demand imbalance for AI and ultimately, build scale.
No previous technology has generated as much demand pull from the business as AI. Business units and functions want to move quickly and they will, with or without ITโs involvement. But few organizations have the centralized resources or funding needed to meet this demand directly. To close that gap, many are now designing a hub-and-spoke operating model that will federate AI delivery across the enterprise while maintaining a consistent foundation of platforms, standards and governance.
In this model, the central AI center of excellence serves as the hub for strategy, enablement and governance rather than as a gatekeeper for approvals. It provides infrastructure, reusable assets, training and guardrails, while the business units take ownership of delivery, funding and outcomes. The power of this model lies in the collaboration between the hubโs AI engineers and the business teams in the spokes. Together, they combine enterprise-grade standards and tools with deep domain context to drive adoption and accountability where it matters most.
One Fortune 500 client, for example, is in the process of implementing its vision for a federated AI operating model. Recognizing the limits of a centralized structure, the CIO and leadership team defined both an interim state and an end-state vision to guide the journey over the next several years. The interim state would establish domain-based AI centers of excellence within each major business area. These domain hubs would be staffed with platform experts, responsible AI advisors and data engineers to accelerate local delivery while maintaining alignment with enterprise standards and governance principles.
The longer-term end state would see these domain centers evolve into smaller, AI-empowered teams that can operate independently while leveraging enterprise platforms and policies. The organization has also mapped out how costs and productivity would shift along the way, anticipating a J-curve effect as investments ramp up in the early phases before productivity accelerates as the enterprise โlearns to fishโ on its own.
The value of this approach lies not in immediate execution but in intentional design. By clearly defining how the transition will unfold and by setting expectations for how the cost curve will behave, the CIO is positioning the organization to scale AI responsibly, in a timeframe that is realistic for the organization.
2026: The year of execution
After two years of experimentation and pilots, 2026 will be the year that separates organizations that can scale AI responsibly from those that cannot. For CIOs, the playbook is now clear. The path forward begins with proving the impact of AI on productivity within IT itself and then extends outward by federating AI capability to the rest of the enterprise in a controlled and scalable way.
Those who can execute on both fronts will win the confidence of their boards and the commitment of their businesses. Those who canโt may find themselves on the wrong side of the J-curve, investing heavily without ever realizing the return.
This article is published as part of the Foundry Expert Contributor Network. Want to join?
CIOs and business leaders everywhere are striving to upgrade legacy technology to meet burgeoning demand for cloud services and artificial intelligence (AI).
But many are stymied by aging IT infrastructure and technical debt. CIOs spend an average of 70% of their IT budgets maintaining legacy systems, according to IBM research, leaving them little room to invest in the innovative solutions they need.
To ease the transition, IBM and Amazon Web Services (AWS) are together helping governments and industry IT leaders modernize infrastructure, applications, and business processes with AI-powered transformation. โIBMโs proprietary agentic AI framework for application migration and modernization embeds agentic AI into the way that IBM drives large-scale migrations to reduce risk and improve efficiency,โ says Dan Kusel, global managing partner & general manager responsible for IBM Consultingโs Global AWS Practice.
โIBM works with AWS to leverage their agentic AI tools, bringing the best capabilities to our clients,โ says Kusel. โThe partnership brings the fastest path to impactful ROI for our clients. This combination is delivering results, including lower costs, faster time-to-market, and happier customers.โ
This article illustrates a few examples of how together IBM and AWS are transforming organizations across a range of industries.
Sports & entertainment: Elevating fan experience
IBM has been working with some of the worldโs most iconic sports organizations. Scuderia Ferrari HP, the renowned Formula 1 racing team, has a fan base of nearly 400 million people who receive news and updates through an app. But new tech-savvy fans wanted more interactivity and personalization.
Ferrari HP partnered with IBM Consulting to redesign the appโs architecture and interface. After studying usersโ habits and engagement patterns, IBM created an intuitive platform that delivers fans just the right mix of racing insights, interactive features, and personalized content.
Results were immediate and impressive. Within a few short months of the new appโs launch, active daily users doubled, and average time spent on the app rose by 35%. The hybrid-cloud infrastructure IBM built on AWS also enabled Ferrari HP to launch AI automations that have already sped up development cycles and improved uptime and reliability. A built-in IBM watsonx.dataยฎ data store ensures the app can expand to reach an even larger fan base as its popularity continues to grow.
Energy & resources: Delivering scale, security, and savings ย
In addition to extending their geographic reach, IBM and AWS are jointly pursuing ventures in new industries. โWorking with AWS, we have seen a starburst of growth in an array of industries: energy and utilities, telecommunications, healthcare, life sciences, financial services, travel and transportation, and manufacturing,โ says Kusel. ย
Southwest Gas โ a natural gas distributor for over 2 million customers in Arizona, Nevada, and California โ also needed the cloud to realize its potential. Like many of its industry peers, the company used data-heavy SAP applications to manage enterprise resources on premises. Technology leaders wanted to improve the performance, resilience, and scalability of these core applications.
Working with IBM Consulting experts, the company migrated the applications to RISE with SAP, an AWS service helping businesses transition to a cloud-based enterprise resource planning (ERP) system.
The big move, completed in less than five months, lowered operating costs and improved SAP application performance by 35%. That means Southwest Gas can process 80 million SAP transactions in less than 10 milliseconds โ an achievement that would have been unthinkable with its legacy systems. The company is now exploring agentic AI as a transformative opportunity to redefine the customer experience.
Travel & transportation: Achieving flexibility, speed, and resiliency
IBM and AWS have continued to transform the travel industry, especially airlines. From Japan Airlines to Finnair to Delta Air Lines, IBM Consulting has partnered with major airlines around the world.
To stay ahead in the hypercompetitive travel industry, Delta Air Lines sought to improve its customer experience. The airline needed to increase agility and responsiveness for 100,000 front-line employees. IBM experts worked closely with Deltaโs IT leaders to plan and execute a combination of migration, containerization, and modernization of over 500 applications to AWS.
Moving to AWS allowed Delta to quickly launch free in-flight Wi-Fi on 1,000 planes and provide more personalized in-flight service. With its new hybrid cloud, Delta can deploy consistent, secure workloads from anywhere, paving the way for exceptional customer service at scale. Leaders also expect the project to continually improve metrics for cost, time-to-market, productivity, and employee engagement.
Automotive: Solving supply chain challenges
Together, IBM and AWS work with global automotive companies, such as Toyota Motors, Daimler, and other industry leaders.
While the industry has undergone continuous disruption and transformation, and has been seriously impacted by supply chain disruption, companies are leveraging technology to improve performance and customer experience. ย
IBM Consulting and Toyota Motors North America have partnered to transform Toyotaโs supply chain processes.ย Working with IBM, Toyota has moved towards an agentic AI experience with an Agent AI Assist built with Amazon Bedrock. This is driving instant supply chain visibility and proactive delay detection, with humans in the loop for decision-making.
IBM and AWS have been working with government agencies around the world.ย Managing ventures of this magnitude requires not only internal resources but also expert third-party help with planning, execution, and scaling.
For example, demand for cloud and AI services are expanding at unprecedented rates across the Middle East. Both governments and industries are making significant investments in modernization and AI to jumpstart productivity and launch new business models.
IBM Consultingโs new collaboration agreement with AWS combines industry expertise in cloud migration and modernization with AWS AI technologies and virtually unlimited scalability. The two companies aim to accelerate technology transformation throughout the region, starting with Saudi Arabia and the UAE.
The two partners โ who together previously built innovation hubs in India and Romania โ are now creating a new innovation hub in Riyadh. The center allows government and enterprise customers to gain hands-on experience with the latest cloud technologies and explore proof-of-concept projects tailored to their needs.
The hub will also expand regional job opportunities. โIt will be staffed domestically, focused on helping our clients deliver projects with local talent,โ says Kusel.
IBM + AWS: Partnership defined by scale
IBM Consulting brings deep domain and industry expertise and a broad range of services and solutions that can help organizations accelerate digital transformation, creating a virtuous cycle of agility, innovation, and resilience.
For large enterprises and governments alike, modernizing business in the AI era can be complex. Together, IBM and AWS offer unparalleled expertise in planning, launching, and scaling tailored initiatives that will deliver bottom-line benefits and real business value for years to come.
El origen del banco ING en Espaรฑa estรก intrรญnsecamente unido a una gran apuesta por la tecnologรญa, su razรณn de ser y clave de un รฉxito que le ha llevado a tener, solo en este paรญs, 4,6 millones de usuarios y ser el cuarto mercado del grupo segรบn este parรกmetro despuรฉs de Alemania, Paรญses Bajos y Turquรญa.
La entidad neerlandesa, que llegรณ al mercado nacional en los aรฑos 80 de mano de la banca corporativa de inversiรณn, realizรณ su gran desembarco empresarial en el paรญs a finales de los 90, cuando empezรณ a operar como el primer banco puramente telefรณnico. Desde entonces, ING ha ido evolucionado al calor de las innovaciones tecnolรณgicas de cada momento, como internet o la telefonรญa mรณvil hasta llegar al momento actual, con un claro protagonismo de la inteligencia artificial.
Como parte de su comitรฉ de direcciรณn y al frente de la estrategia de las tecnologรญas de la informaciรณn del banco en Iberia โy de un equipo de 500 profesionales, un tercio de la plantilla de la compaรฑรญaโ estรก la teleco Rocรญo Lรณpez Valladolid, su CIO desde septiembre de 2022. La ejecutiva, en la โcasaโ desde hace mรกs de 15 aรฑos y distinguida como CIO del aรฑo en los CIO 100 Awards en 2023, explica en entrevista con esta cabecera cรณmo trabaja ING para evolucionar sus sistemas, procesos y forma de trabajar en un contexto enormemente complejo y cambiante como el actual.
Asegura ser consciente, desde que se incorporรณ a ING, de la relevancia de las TI para el banco desde sus inicios, un rol que โno ha sido a menosโ en los tres aรฑos de Lรณpez Valladolid como CIO de la filial ibรฉrica. โMi estrategia y la estrategia de tecnologรญa del banco va ligada a la del banco en sรญ mismaโ, recalca, apostillando que desde su รกrea no perciben las TI โcomo una estrategia que reme solo en la direcciรณn tecnolรณgica, sino siempre como el mayor habilitador, el mayor motor de nuestra estrategia de negocioโ.
Una ambiciosa transformaciรณn tecnolรณgica
Los 26 aรฑos de operaciรณn de ING en Espaรฑa han derivado en un gran legado tecnolรณgico que la compaรฑรญa estรก renovando. โTenemos que seguir modernizando toda nuestra arquitectura tecnolรณgica para asegurar que seguimos siendo escalables, eficientes en nuestros procesos y, sobre todo, para garantizar que estamos preparados para incorporar las disrupciones que, una vez mรกs, vienen de la mano de la tecnologรญa, en especial de la inteligencia artificialโ, asevera la CIO.
Fue hace tres aรฑos, cuenta, cuando Lรณpez Valladolid y su equipo hicieron un replanteamiento de la experiencia digital para modernizar la tecnologรญa que da servicio directo a sus clientes. โEmpezamos a ofrecer nuevos productos y servicios de la mano de nuestra app en el canal mรณvil, que ya se ha convertido en el principal canal de acceso de nuestros clientesโ, seรฑala.
Mรกs tarde, continรบa, su equipo siguiรณ trabajando en modular los sistemas del banco. โAquรญ uno de nuestros grandes hitos tecnolรณgicos fue la migraciรณn de todos nuestros activos a la nube privada del grupoโ โsubrayaโ. Un hito que cumplimos el aรฑo pasado, siendo el primer banco en afrontar este movimiento ambicioso, que nos ha proporcionado mucha escalabilidad tecnolรณgica y eficiencia en nuestros sistemas y procesos, ademรกs de unirnos como equipoโ.
Un proyecto, el de migraciรณn a cloud, clave en su carrera profesional. โNo todo el mundo tiene la oportunidad de llevar un banco a la nubeโ, afirma. โY he de decir que todos y cada uno de los profesionales del รกrea de tecnologรญa hemos trabajado codo con codo para conseguir ese gran hito que nos ha posicionado como un referente en innovaciรณn y escalabilidadโ.
En la actualidad, agrega, su equipo estรก trabajando en evolucionar el core bancario de ING. โLlegar a transformar las capas mรกs profundas de nuestros sistemas es uno de los grandes hitos que muchos bancos ambicionanโ, relata. ยฟEl objetivo? Ser mรกs escalables en los procesos y estar mejor preparados para incorporar las ventajas que vienen de mano de la inteligencia artificial.
Gran parte de las inversiones de TI del banco โla CIO no desvela el presupuesto especรญfico anual de su รกrea en Iberiaโ estรกn enfocadas a la citada transformaciรณn tecnolรณgica y al desarrollo de los productos y servicios que demandan los clientes.
Muestra de la confianza en las capacidades locales del grupo es el establecimiento en las oficinas del banco en Madrid de un centro de innovaciรณn y tecnologรญa global que persigue impulsar la transformaciรณn digital del banco en todo el mundo. El proyecto, una iniciativa de la corporaciรณn, espera generar mรกs de mil puestos de trabajo especializados en tecnologรญa, datos, operaciones y riesgos hasta el aรฑo 2029. Aunque Lรณpez no lidera este proyecto corporativo โKonstantin Gordievitch, en la casa desde hace casi dos dรฉcadas, estรก al frenteโ sรญ cree que โes un orgullo y pone de manifiesto el reconocimiento global del talento que tenemos en Espaรฑaโ. Gracias al nuevo centro, explica, โse va a dotar al resto de paรญses de ING de las capacidades tecnolรณgicas que necesitan para llevar a cabo sus estrategiasโ.
Garpress | Foundry
โNo todo el mundo tiene la oportunidad de llevar un banco a la nubeโ
Pilares de la estrategia de TI de ING en Iberia
La estrategia de ING, dice Lรณpez Valladolid, es โcustomer centricโ, es decir, centrada en el cliente y ese es uno de sus grandes pilares. โDe alguna manera, todos trabajamos y desarrollamos para nuestros clientes, asรญ que estos son uno de los pilares fundamentales tanto en nuestra estrategia como banco como en nuestra estrategia tecnolรณgicaโ.
La escalabilidad, continรบa la CIO, es el siguiente. โING estรก creciendo en negocio, productos, servicios y segmentos, asรญ que el รกrea de tecnologรญa debe dar respuesta de manera escalable y tambiรฉn sostenible, porque este incremento no puede conllevar que aumente el coste y la complejidadโ.
โPor supuesto โaรฑadeโ la seguridad desde el diseรฑo es un pilar fundamental en todos nuestros procesos y en el desarrollo de productosโ. Su equipo, afirma, trabaja con equipos multidisciplinares y, en concreto, sus equipos de producto y tecnologรญa trabajan conjuntamente con el de ciberseguridad para garantizar este enfoque.
La innovaciรณn es otro de los cimientos tecnolรณgicos del banco. โEstamos viviendo una revoluciรณn que va mรกs allรก de la tecnologรญa y va a afectar a todo lo que hacemos: a cรณmo trabajamos, cรณmo servimos a nuestros clientes, cรณmo operamosโฆ Asรญ que la innovaciรณn y cรณmo incorporamos las nuevas disrupciones para mejorar la relaciรณn con los clientes y nuestros procesos internos son aspectos claves en nuestra estrategia tecnolรณgicaโ.
Finalmente, afirma, โel รบltimo pilar y el mรกs importante son las personas, el equipo. Para nosotros, por supuesto para mรญ, es fundamental contar con un equipo diverso, muy conectado con el propรณsito del banco y que sienta que su trabajo redunda en algo positivo para la sociedadโ.
Impacto de los nuevos sabores de IA
Preguntada por la sobreexpectaciรณn que ha generado en la alta direcciรณn de negocio la apariciรณn de los sabores generativo y agentivo de la IA, Lรณpez Valladolid lo ve con buenos ojos: โQue los CEO tengan esas expectativas y ese empuje es bueno. Histรณricamente, a los tecnรณlogos nos ha costado explicar a los CEO la importancia de la tecnologรญa; que ahora ellos tiren de nosotros lo veo muy positivoโ.
ยฟCรณmo deben actuar los CIO en este escenario? โDiseรฑando las estrategias para que la IA genere el impacto positivo que sabemos que va a tenerโ, explica la CIO. โEn ING no vemos la IA generativa como un sustituto de las personas, sino como un amplificador de las capacidades de รฉstas. De hecho, tenemos ya planes para mejorar el dรญa a dรญa de los empleados y reinventar la relaciรณn que tenemos con los clientesโ.
ING, rememora, irrumpiรณ en el escenario de la banca en Espaรฑa hace 26 aรฑos con โun modelo de relaciรณn muy diferente, que no existรญa entonces. Primero fuimos un banco telefรณnico e inmediatamente despuรฉs un banco digital sin casi oficinas, un modelo de relaciรณn con el cliente entonces disruptivo y que se ha consolidado como el modelo de relaciรณn estรกndar de las personas con sus bancosโ. En la era actual, aรฑade, โtendremos que entender cuรกl va a ser el modelo de relaciรณn que las personas van a tener, gracias a la IA generativa, con sus bancos o sus propios dispositivos. Nosotros ya estamos trabajando para entender cรณmo quieren nuestros clientes que nos relacionemos con ellosโ. Una respuesta que vendrรก, dice, siempre de mano de la tecnologรญa.
Garpress | Foundry
โQueremos rediseรฑar nuestro modelo operativo para ser mucho mรกs eficientes internamente, asรญ que estamos trabajando para ver dรณnde [la IA generativa] nos puede aportar valorโ
De hecho, la compaรฑรญa ha lanzado un chatbot basado en IA generativa para dar respuesta de forma โmรกs natural y cercanaโ a las consultas del dรญa a dรญa de los clientes. โAsรญ podemos dejar a nuestros agentes [humanos] para atender otro tipo de cuestiones mรกs complejas que sรญ requieren la respuesta de una personaโ.
ING tambiรฉn aplicarรก la IA generativa a sus propios procesos empresariales. โQueremos rediseรฑar nuestro modelo operativo para ser mucho mรกs eficientes internamente, asรญ que estamos trabajando para ver dรณnde [la IA generativa] nos puede aportar valorโ.
La CIO es consciente de la responsabilidad que conlleva adoptar esta tecnologรญa. โTenemos que liderar el cambio y asegurarnos de que la inteligencia artificial generativa nos lleve donde queremos estar y que nosotros la llevemos donde tambiรฉn queremos que estรฉโ.
En lo que respecta a la aplicaciรณn de esta tecnologรญa al รกrea de TI en concreto, donde los analistas esperan un impacto grande, sobre todo en el desarrollo de software, la CIO cree que โpuede aportar muchรญsimoโ. La idea, cuenta, es emplearla para tareas de menos valor aรฑadido, mรกs tediosas, de modo que los profesionales de TI del banco puedan dedicarse a otro tipo de tareas dentro del desarrollo de software donde puedan aportar mรกs valor.
Garpress | Foundry
โHistรณricamente, a los tecnรณlogos nos ha costado explicar a los CEO la importancia de la tecnologรญa; que ahora ellos tiren de nosotros lo veo muy positivoโ
Desafรญos como CIO y futuro de la banca
Los lรญderes de TI afrontan todo un crisol de retos que engloban desde el liderazgo tecnolรณgico a desafรญos culturales o regulatorios, entre otros. โLos CIO nos enfrentamos a todo tipo de desafรญosโ, reflexiona Rocรญo Lรณpez. โPor un lado, soy colรญder de la estrategia del banco y del negocio; me preocupa y ocupa el crecimiento del banco y los servicios que damos a nuestros clientes, lo que conlleva un abanico de retos y disciplinas muy amplioโ.
Por otro, aรฑade, โlos lรญderes tecnolรณgicos marcamos el paso de la transformaciรณn y de la innovaciรณn, garantizando que la seguridad estรก en todo lo que hacemos desde el diseรฑo. En este sentido, siempre tenemos que reconciliar la innovaciรณn con la regulaciรณn, pues esta รบltima nos protege como sociedadโ. Por รบltimo, subraya, โlos CIO somos lรญderes de personas, asรญ que es muy importante dedicar tiempo y esfuerzo al desarrollo de nuestros equipos, de forma que estos crezcan y se desarrollen en una profesiรณn que me encantaโ.
Una de las iniciativas en la que la CIO participa activamente para impulsar la profesiรณn y potenciar que existan mรกs referentes femeninos en el mundo STEM (de ciencias, tecnologรญa, ingenierรญa y matemรกticas) es Leonas in Tech. โEs una comunidad formada por el equipo de mujeres del รกrea de tecnologรญa del banco con la que realizamos varias acciones, como talleres de robรณtica, entre otrosโ, explica. โNos preocupa que los perfiles tecnolรณgicos femeninos seamos una minorรญa en la sociedad. En un mundo donde ya todo es tecnologรญa, y en el futuro lo serรก mรกs aรบn, que las mujeres no tengamos una representaciรณn fuerte en este segmento nos pone en cierto riesgo como sociedad. Por eso trabajamos para fomentar que haya referentes y acercar la tecnologรญa a las edades mรกs tempranas; contar que la nuestra es una profesiรณn bonita caracterizada por la creatividad, la capacidad de resolver problemas, el ingenioโฆ y el pensamiento crรญticoโ, aรฑade la CIO.
De cara al futuro prรณximo, Lรณpez Valladolid estรก convencida de que โla inteligencia artificial va a cambiar la manera en la que en la que nos relacionamos. Es difรญcil anticipar lo que va a ocurrir a cinco aรฑos vista, pero sรญ sabemos que debemos seguir escuchando a nuestros clientes y saber quรฉ nos demandan. Esto siempre serรก una prioridad para nosotros. Y seguiremos estando donde los clientes nos pidan gracias a la tecnologรญaโ.
For all the buzz about AIโs potential to transform business, many organizations struggle to ascertain the extent to which their AI implementations are actually working.
Part of this is because AI doesnโt just replace a task or automate a process โ rather, it changes how work itself happens, often in ways that are hard to quantify. Measuring that impact means deciding what return really means, and how to connect new forms of digital labor to traditional business outcomes.
โLike everyone else in the world right now, weโre figuring it out as we go,โ says Agustina Branz, senior marketing manager at Source86.
That trial-and-error approach is what defines the current conversation about AI ROI.
To help shed light on measuring the value of AI, we spoke to several tech leaders about how their organizations are learning to gauge performance in this area โ from simple benchmarks against human work to complex frameworks that track cultural change, cost models, and the hard math of value realization.
The simplest benchmark: Can AI do better than you?
Thereโs a fundamental question all organizations are starting to ask, one that underlies nearly every AI metric in use today: How well does AI perform a task relative to a human? For Source86โs Branz, that means applying the same yardstick to AI that she uses for human output.
โAI can definitely make work faster, but faster doesnโt mean ROI,โ she says. โWe try to measure it the same way we do with human output: by whether it drives real results like traffic, qualified leads, and conversions. One KPI that has been useful for us has been cost per qualified outcome, which basically means how much less it costs to get a real result like the ones we were getting before.โ
The key is to compare against what humans delivered in the same context. โWe try to isolate the impact of AI by running A/B tests between content that uses AI and those that donโt,โ she says.
โFor instance, when testing AI-generated copy or keyword clusters, we track the same KPIs โ traffic, engagement, and conversions โ and compare the outcome to human-only outputs,โ Branz explains. โAlso, we treat AI performance as a directional metric rather than an absolute one. It is super useful for optimization, but definitely not the final judgment.โ
MarcโAurele Legoux, founder of an organic digital marketing agency, is even more blunt. โCan AI do this better than a human can? If yes, then good. If not, thereโs no point to waste money and effort on it,โ he says. โAs an example, we implemented an AI agent chatbot for one of my luxury travel clients, and it brought in an extra โฌ70,000 [$81,252] in revenue through a single booking.โ
The KPIs, he said, were simply these: โDid the lead come from the chatbot? Yes. Did this lead convert? Yes. Thank you, AI chatbot. We would compare AI-generated outcomes โ leads, conversions, booked calls โagainst human-handled equivalents over a fixed period. If the AI matches or outperforms human benchmarks, then itโs a success.โ
But this sort of benchmark, while straightforward in theory, becomes much harder in practice. Setting up valid comparisons, controlling for external factors, and attributing results solely to AI is easier said than done.
Hard money: Time, accuracy, and value
The most tangible form of AI ROI involves time and productivity. John Atalla, managing director at Transformativ, calls this โproductivity upliftโ: โtime saved and capacity released,โ measured by how long it takes to complete a process or task.
But even clear metrics can miss the full picture. โIn early projects, we found our initial KPIs were quite narrow,โ he says. โAs delivery progressed, we saw improvements in decision quality, customer experience, and even staff engagement that had measurable financial impact.โ
That realization led Atallaโs team to create a framework with three lenses: productivity, accuracy, and what he calls โvalue-realization speedโโ โhow quickly benefits show up in the business,โ whether measured by payback period or by the share of benefits captured in the first 90 days.
The same logic applies at Wolters Kluwer, where Aoife May, product management association director, says her teams help customers compare manual and AI-assisted work for concrete time and cost differences.
โWe attribute estimated times to doing tasks such as legal research manually and include an average attorney cost per hour to identify the costs of manual effort. We then estimate the same, but with the assistance of AI.โ Customers, she says, โreduce the time they spend on obligation research by up to 60%.โ
But time isnโt everything. Atallaโs second lens โ decision accuracy โ captures gains from fewer errors, rework, and exceptions, which translate directly into lower costs and better customer experiences.
Adrian Dunkley, CEO of StarApple AI, takes the financial view higher up the value chain. โThere are three categories of metrics that always matter: efficiency gains, customer spend, and overall ROI,โ he says, adding that he tracks โhow much money you were able to save using AI, and how much more you were able to get out of your business without spending more.โ
Dunkleyโs research lab, Section 9, also tackles a subtler question: how to trace AIโs specific contribution when multiple systems interact. He relies on a process known as โimpact chaining,โ which he โborrowed from my climate research days.โ Impact chaining maps each process to its downstream business value to create a โpre-AI expectation of ROI.โ
Tom Poutasse, content management director at Wolters Kluwer, also uses impact chaining, and describes it as โtracing how one change or output can influence a series of downstream effects.โ In practice, that means showing where automation accelerates value and where human judgment still adds essential accuracy.
Still, even the best metrics matter only if theyโre measured correctly. Establishing baselines, attributing results, and accounting for real costs are what turn numbers into ROI โ which is where the math starts to get tricky.
Getting the math right: Baselines, attribution, and cost
The math behind the metrics starts with setting clean baselines and ends with understanding how AI reshapes the cost of doing business.
Salome Mikadze, co-founder of Movadex, advises rethinking what youโre measuring: โI tell executives to stop asking โwhat is the modelโs accuracyโ and start with โwhat changed in the business once this shipped.โโ
Mitadzeโs team builds those comparisons into every rollout. โWe baseline the pre-AI process, then run controlled rollouts so every metric has a clean counterfactual,โ she says. Depending on the organization, that might mean tracking first-response and resolution times in customer support, lead time for code changes in engineering, or win rates and content cycle times in sales. But she says all these metrics include โtime-to-value, adoption by active users, and task completion without human rescue, because an unused model has zero ROI.โ
But baselines can blur when people and AI share the same workflow, something that spurred Poutasseโs team at Wolters Kluwer to rethink attribution entirely. โWe knew from the start that the AI and the human SMEs were both adding value, but in different ways โ so just saying โthe AI did thisโ or โthe humans did thatโ wasnโt accurate.โ
Their solution was a tagging framework that marks each stage as machine-generated, human-verified, or human-enhanced. That makes it easier to show where automation adds efficiency and where human judgment adds context, creating a truer picture of blended performance.
At a broader level, measuring ROI also means grappling with what AI actually costs. Michael Mansard, principal director at Zuoraโs Subscribed Institute, notes that AI upends the economic model that IT has taken for granted since the dawn of the SaaS era.
โTraditional SaaS is expensive to build but has near-zero marginal costs,โ Mansard says, โwhile AI is inexpensive to develop but incurs high, variable operational costs. These shifts challenge seat-based or feature-based models, since they fail when value is tied to what an AI agent accomplishes, not how many people log in.โ
Mansard sees some companies experimenting with outcome-based pricing โ paying for a percentage of savings or gains, or for specific deliverables such as Zendeskโs $1.50-per-case-resolution model. Itโs a moving target: โThere isnโt and wonโt be one โrightโ pricing model,โ he says. โMany are shifting toward usage-based or outcome-based pricing, where value is tied directly to impact.โ
As companies mature in their use of AI, theyโre facing a challenge that goes beyond defining ROI once: Theyโve got to keep those returns consistent as systems evolve and scale.
Scaling and sustaining ROI
For Movadexโs Mikadze, measurement doesnโt end when an AI system launches. Her framework treats ROI as an ongoing calculation rather than a one-time success metric. โOn the cost side we model total cost of ownership, not just inference,โ she says. That includes โintegration work, evaluation harnesses, data labeling, prompt and retrieval spend, infra and vendor fees, monitoring, and the people running change management.โ
Mikadze folds all that into a clear formula: โWe report risk-adjusted ROI: gross benefit minus TCO, discounted by safety and reliability signals like hallucination rate, guardrail intervention rate, override rate in human-in-the-loop reviews, data-leak incidents, and model drift that forces retraining.โ
Most companies, Mikadze adds, accept a simple benchmark: ROI = (ฮ revenue + ฮ gross margin + avoided cost) โ TCO, with a payback target of less than two quarters for operations use cases and under a year for developer-productivity platforms.
But even a perfect formula can fail in practice if the model isnโt built to scale. โA local, motivated pilot team can generate impressive early wins, but scaling often breaks things,โ Mikadze says. Data quality, workflow design, and team incentives rarely grow in sync, and โAI ROI almost never scales cleanly.โ
She says she sees the same mistake repeatedly: A tool built for one team gets rebranded as a company-wide initiative without revisiting its assumptions. โIf sales expects efficiency gains, product wants insights, and ops hopes for automation, but the model was only ever tuned for one of those, friction is inevitable.โ
Her advice is to treat AI as a living product, not a one-off rollout. โSuccessful teams set very tight success criteria at the experiment stage, then revalidate those goals before scaling,โ she says, defining ownership, retraining cadence, and evaluation loops early on to keep the system relevant as it expands.
That kind of long-term discipline depends on infrastructure for measurement itself. StarApple AIโs Dunkley warns that โmost companies arenโt even thinking about the cost of doing the actual measuring.โ Sustaining ROI, he says, โrequires people and systems to track outputs and how those outputs affect business performance. Without that layer, businesses are managing impressions, not measurable impact.โ
The soft side of ROI: Culture, adoption, and belief
Even the best metrics fall apart without buy-in. Once youโve built the spreadsheets and have the dashboards up and running, the long-term success of AI depends on the extent to which people adopt it, trust it, and see its value.
Michael Domanic, head of AI at UserTesting, draws a distinction between โhardโ and โsquishyโ ROI.
โHard ROI is what most executives are familiar with,โ he says. โIt refers to measurable business outcomes that can be directly traced back to specific AI deployments.โ Those might be improvements in conversion rates, revenue growth, customer retention, or faster feature delivery. โThese are tangible business results that can and should be measured with rigor.โ
But squishy ROI, Domanic says, is about the human side โ the cultural and behavioral shifts that make lasting impact possible. โIt reflects the cultural and behavioral shift that happens when employees begin experimenting, discovering new efficiencies, and developing an intuition for how AI can transform their work.โ Those outcomes are harder to quantify but, he adds, โthey are essential for companies to maintain a competitive edge.โ As AI becomes foundational infrastructure, โthe boundary between the two will blur. The squishy becomes measurable and the measurable becomes transformative.โ
John Pettit, CTO of Promevo, argues that self-reported KPIs that could be seen as falling into the โsquishyโ category โ things like employee sentiment and usage rates โ can be powerful leading indicators. โIn the initial stages of an AI rollout, self-reported data is one of the most important leading indicators of success,โ he says.
When 73% of employees say a new tool improves their productivity, as they did at one client company he worked with, that perception helps drive adoption, even if that productivity boost hasnโt been objectively measured. โWord of mouth based on perception creates a virtuous cycle of adoption,โ he says. โEffectiveness of any tool grows over time, mainly by people sharing their successes and others following suit.โ
Still, belief doesnโt come automatically. StarApple AI and Section 9โs Dunkley warn that employees often fear AI will erase their credit for success. At one of the companies where Section 9 has been conducting a long-term study, โstaff were hesitant to have their work partially attributed to AI; they felt they were being undermined.โ
Overcoming that resistance, he says, requires champions who โput in the work to get them comfortable and excited for the AI benefits.โ Measuring ROI, in other words, isnโt just about proving that AI works โ itโs about proving that people and AI can win together.
Consider the Turing test. Its challenge? Ask some average humans to tell whether theyโre interacting with a machine or another human.
The fact of the matter is, generative AI passed the Turing test a few years ago.
I suggested as much to acquaintances who are knowledgeable in the ways of artificial intelligence. Many gave me the old eyeball roll in response. In pitying tones, they let me know Iโm just not sophisticated enough to recognize that generative AI didnโt pass Turingโs challenge at all. Why not? I asked. Because the way generative AI works isnโt the same as how human intelligence works, they explained.
Now I could argue with my more AI-sophisticated colleagues but where would the fun be in that? Instead, Iโm willing to ignore what โImitation Gameโ means. If generative AI doesnโt pass the test, what we need isnโt better AI.
Itโs a better test.
What makes AI agentic
Which brings us to the New, Improved, AI Imitation Challenge (NIAIIC).
The NIAIIC still challenges human evaluators to determine whether theyโre dealing with a machine or a human. But NIAIICโs challenge is no longer about conversations.
Itโs about something more useful. Namely, dusting. I will personally pay a buck and a half to the first AI team able to deploy a dusting robot โ one that can determine which surfaces in an average testerโs home are dusty, and can remove the dust on all of them without breaking or damaging anything along the way.
Clearly, the task to be mastered is one a human could handle without needing detailed instructions (aka โprogrammingโ). Patience? Yes, dusting needs quite a bit of that. But instructions? No.
Itโs a task with the sorts of benefits claimed for AI by its most enthusiastic proponents: It takes over annoying, boring, and repetitive work from humans, freeing them up for more satisfying responsibilities.
(Yes, I freely admit that Iโm projecting my own predilections. If you, unlike me, love to dust and canโt get enough of it โฆ come on over! Iโll even make espresso for you!)
How does NIAIIC fit into the popular AI classification frameworks? It belongs to the class of technologies called โagentic AIโ โ who comes up with these names? Agentic AI is AI that figures out how to accomplish defined goals on its own. Itโs what self-driving vehicles do when they do what theyโre supposed to do โ pass the โtouring testโ (sorry).
Itโs also what makes agentic AI interesting when compared to earlier forms of AI โ those that depended on human experts encoding their skills into a collection of if/then rules, which are alternately known as โexpert systemsโ and โAI that reliably works.โ
Whatโs worrisome is how little distance separates agentic AI from the Worst AI Idea Yet, namely, volitional AI.
With agentic AI, humans define the goals, while the AI figures out how to achieve them. With volitional AI, the AI decides which goals it should try to achieve, then becomes agentic to achieve them.
Once upon a time I didnโt worry much about volitional AI turning into Skynet, on the grounds that, โExcept for electricity and semiconductors, itโs doubtful we and a volitional AI would find ourselves competing for resources intensely enough for the killer robot scenario to become a problem for us.โ
Itโs time to rethink this conclusion. Do some Googling and youโll discover that some AI chips arenโt even being brought online because there isnโt enough juice to power them.
It takes little imagination to envision a dystopian scenario in which volitional AIs compete with us humans to grab all the electrical generation they can get their virtual paws on. Their needs and ours will overlap, potentially more quickly than weโre able to even define the threat, let alone respond to it.
The tipping point
Speaking more broadly, anyone expending even a tiny amount of carbon-based brainpower regarding the risks of volitional AI will inevitably reach the same conclusion Microsoft Copilot does. I asked Copilot what the biggest risks of volitional AI are. It concluded that:
The biggest risks of volitional AI โ AI systems that act with self-directed goals or autonomy โ include existential threats, misuse in weaponization, erosion of human control, and amplification of bias and misinformation. These dangers stem from giving AI systems agency beyond narrow task execution, which could destabilize social, economic, and security structures if not carefully governed.
But itโs okay so long as we stay on the right side of the line that separates agentic from volitional AI, isnโt it?
In a word, โno.โ
When an agentic AI figures out how it can go about achieving a goal, what it must do is break down the goal assigned to it into smaller goal chunks, and then to break down these chunks into yet smaller chunks.
An agentic AI, that is, ends up setting goals for itself because thatโs how planning works. But once it starts to set goals for itself, it becomes volitional by definition.
Which gets us to AIโs IT risk management conundrum.
Traditional risk management identifies bad things that might happen, and crafts contingency plans that explain what the organization should do should the bad thing actually happen.
We can only wish that this framework would be sufficient when we poke and prod an AI implementation.
Agentic AI, and even more so volitional AI, stands this on its head, because when it comes to it, the biggest risk of volitional AI isnโt that an unplanned bad thing has happened. Itโs that the AI does what itโs supposed to do.
Volitional AI is, that is, dangerous. Agentic AI might not be as inherently risky, but itโs more than risky enough.
Sad to say, we humans are probably too shortsighted to bother mitigating agentic and volitional AIโs clear and present risks, even risks that could herald the end of human-dominated society.
The likely scenario? Weโll all collectively ignore the risks. Me too. I want my dusting robot and I want it now, the risks to human society be damned.
Googleโs AI research division, in collaboration with researchers from New York University and UC Santa Barbara, has launched a new tool and framework that they claim can rein in agent behavior by placing explicit limits on how much compute agents can consume and how freely they can invoke tools.
The tool, called Budget Tracker, is a plug-in module that injects continuous budget awareness into an agentโs reasoning loop, explicitly signaling how much token and tool-call budget remains so the agent can condition its actions on real-time resource availability, the researchers explained in a paper.
How autonomous agents choose to think, loop, and act when completing a task or user request is quietly inflating enterprise AI budgets, which is increasingly turning into a CIO nightmare.
A research report from IDC cited 92% of 318 decision makers surveyed in the US, UK, and Ireland saying that the cost of their deployed AI agents was higher than expected, with inference being the most common cause.
Another report โ Greyhound CIO Pulse 2025 โ found that 68% of digital leaders surveyed said that they hit major budget overruns during their first few deployments of agents, and nearly half of these leaders pointed to runaway tool loops and recursive logic being the driver behind the overruns.
That budget signal doesnโt stop at individual agents. Google researchers have also launched a framework, named Budget Aware Test-time Scaling (BATS), that builds on the tracker-driven awareness in an agent and tries to adapt the same to a larger multi-agent system.
This enables large, multi-agent systems to determine when itโs worth โdigging deeperโ into a promising line of reasoning versus when to pivot to alternative paths, thereby improving cost-performance trade-offs under tight budget constraints, the researchers wrote.
Practical path to TCO for agentic AI
Analysts say the tool and framework could offer CIOs a practical path to understand and control costs around agentic deployments.
โThe tool and the framework address one of the most pressing questions facing CFOs and CIOs: the true total cost of ownership (TCO) for agentic AI operations. And for enterprise adoption of any technology, TCO is very important,โ said Pareekh Jain, principal analyst at Pareekh Consulting.
โMost CIOs would rather accept a predictable agentic system that delivers 90% accuracy at a capped cost than a higher-accuracy system whose per-run cost can vary wildly from a few cents to several dollars,โ Jain added.
Seconding Jain, IT management consulting firm Avasantโs research director, Gaurav Dewan, pointed out that many enterprises deploy agents under the assumption that inference is โcheap enoughโ at scale.
In practice, however, Dewan said, costs grow non-linearly as agents are often deployed without hard budget caps, execution limits, or reasoning constraints.
He further noted that spending is amplified in multi-agent architectures as agents delegate tasks to other agents.
The tool and the framework mark a โbig shiftโ from approaches that similar tools take, according to Greyhound Research Chief Analyst Sanchit Vir Gogia.
โRight now, the tooling landscape is a mess. Plenty of platforms say they optimize agent efficiency, but few actually engage with the problem at runtime. Most solutions in the market operate on a delay; they aggregate usage data, send reports, or shut things down when itโs already too late. Thatโs not cost governance. Thatโs damage control,โ Gogia said.
The analyst was referring to tools such as LangSmith (LangChain) & Helicone, which offer observability by logging how much an agent spent.
In contrast, Googleโs Budget Tracker and BATS frameworks introduce budget awareness directly into the decision loop, giving agents real-time feedback about what they have spent, whatโs left, and whatโs worth doing next, that too without the need to babysit the agent, Gogia said.
In the absence of a framework like BATS, enterprises typically had to cobble together throttling rules, hard-coded prompt tweaks, and caching strategies to control costs around the agentic system, Gogia added.
Caveats to enterprise adoption
Despite the differentiated approach of the new tool and framework, analysts caution that they might not be a silver bullet in tackling cost overruns around agentic deployments on their own.
CIOs, according to Dewan, are likely to favor the tool or framework if Google integrates natively with existing agent frameworks and offers policy-driven controls spread granularly across all agents, workflows, and business units. โWhat CIOs need is enforcement, observability, and auditability in one place,โ Gogia said. โThey want to know what the agent spent, why it made that decision, what path it took, and whether it stayed within bounds. Especially in regulated sectors, where tool usage may touch PII or trigger downstream financial systems, audit trails are non-negotiable.โ