Keeping Up With the Pace of Change Within Microsoft AI Tools

Preview

In this blog post, we discuss the rapid pace of change in AI-building technologies, and share a few of the methods that Gravity Union uses to keep up to date and build future-proof AI solutions through:

  • Ongoing AI sharing practices

  • Keeping a log of architectural options

  • Creating an experiment log per project

  • Keeping and tracking output quality as a priority metric for success

 

The Gravity Union Team has been building AI solutions on the Microsoft platform for over five years. Microsoft’s Cognitive Services served as the foundation for our client chatbot agents. Power Virtual Agents supported our Power Platform solutions. When Microsoft introduced semantic capabilities in Azure Search Services in 2021, we built our first RAG index. Every one of our employees has been licensed for Copilot since its launch. If you were to look for any of these tools today, you’d find that they not only have a new name, but they have evolved significantly.

Recency is competing with experience

The AI arena is rapidly evolving. Within Microsoft's own tooling, product names keep changing, and the changes go well beyond rebranding. Layers of newly released AI capabilities are crowding admin dashboards and dominating licensing tables. It's a one-two (and three) punch: the tool you used yesterday not only has a new name and a new location, but has also been joined by another feature that sounds remarkably similar. There's a constant feeling that perhaps there's a better tool for the job.

With the pace of change, keeping on top of new features can feel like a full-time job.

 

So how do you keep up with the rapid pace of change and build solid, future-proof solutions?

 

Here's how we've adapted — along with a few of the tools we use along the journey:

We've elevated AI to a required competency

We want everyone in our organization to not only use AI, but to actively consider how to maximize productivity with it. To support this, we include AI in onboarding training and have updated our SDLC policies to keep developers current on new AI capabilities and safety concerns.

Our most impactful learning tool, though, has been our weekly "Thursday is AI Growth Day" sessions — a 30-minute drop-in where we share AI trends. Our monthly rhythm looks like this:

  • First Thursday – AI practical working tips

  • Second Thursday – Client enablement

  • Third Thursday – Innovation showcase

  • Fourth Thursday – Open mic

  • Fifth Thursday (~4 times/year) – AI safety and governance

Sessions are recorded and shared, keeping the team sharp and providing just-in-time resources when people need them.

We keep our own log of Microsoft AI capabilities

Microsoft promotes a number of AI building options and each one includes a number of configuration possibilities. For example, within Copilot Studio and Azure AI Foundry alone there are a number configuration choices— such as selecting between Microsoft's foundational models or choosing to use a bring-your-own-model approach. Common AI features such as temperature settings and RAG options may be surfaced uniquely or not available at all depending on these architecture choices. That inconsistency led us to start tracking what we could and couldn't do across our options.

Within our log, we track competing options for building AI solutions on the Microsoft platform, and create a detailed breakdown of what each solution is good at, bad at, and factors that might influence our use of that technology (such as licensing considerations).

Our log of agent options allows us to share insights with our team on which tool is best fit for the job.

We track our AI project experiments

In the past, choosing the best tool for the job wasn’t something that would changed frequently. But today, at the pace of change we’re experiencing with Microsoft, we now form a hypotheses about which tool to use and when. Architecture decisions have become weighted possibilities that require rapid testing.

Some of the questions we consider during AI experimentation:

  • Is this a single- or multi-agent problem?

  • Is there a model optimized for this type of work?

  • How do we optimize licensing — especially when the client is already paying for AI-building tools?

  • Where is data being processed, and how do we stay within required boundaries? (Note: this is shifting frequently as AI capabilities become more localized.)

We track all of this in an experimentation log. Each technical step in a workflow becomes a hypothesis with defined goals, assumptions, and performance factors. Like a scientist eliminating variables, the log moves us toward the best-fit answer.

A secondary benefit: we now have a historical record of what we've tried and why something did or didn't work.

Keeping track of experiments allows us to retain our learning on what we’ve tried and what did or didn’t work.

We continue to prioritize quality metrics

AI models have a dangerous ability to deliver confident wrong answers. Building solutions that produce reliable, accurate results should be a priority for every builder — and for us, this has become increasingly important as we test new features.

Factors that can influence AI quality include:

  • Model choice

  • Model settings (such as temperature and top-k)

  • Pipeline decisions, including orchestration and MCP choices

  • Document processing choices

  • Search and index factors (semantic and vector choices, chunking size)

  • RAG assembly and sizing

For each choice or setting adjustment, we evaluate:

  • Processing cost per transaction

  • Processing speed and response time

  • Accuracy of the response

  • Tone and character of the response

Benchmarking is one of the most important factors in ensuring AI is successfully adopted. Users — customers and employees alike — will quickly abandon tools like Q&A chatbots if the results aren't reliable.

Experimentation is a new AI competency

Building successful AI solutions requires more than understanding AI architectures. It requires ongoing engagement with the evolving tools and platforms those solutions are built on. Staying regularly informed, maintaining an architecture options log, and tracking experiments and quality are all ways to ensure your AI automation stays on track.

Brian Edwards

Brian Edwards is the Director of Artificial Intelligence at Gravity Union, where he drives innovation in delivery and operations while enhancing customer success with AI. With over 25 years of consulting experience, he pioneered a collaboration practice in SharePoint in 2001 and has served on multiple Microsoft client advisory boards. Passionate about exploring new technology frontiers, he thrives on bringing education and insights to future adopters—keeping both his audiences’ minds and his own ADHD brain engaged. 

https://www.allofushumans.com/
Previous
Previous

Getting Control of SharePoint Site Growth with Site Attestation

Next
Next

Winter 2026 Newsletter