I have started to rethink our motto.
After more than 2 decades, our motto: “The hallmark of service in the digital age is human interaction- and we employ great humans,” may be obsolete. At the time of its creation, we used it to highlight our empathetic and expert service delivery to our managed services clients. The motto was used to address a customer pain point: the increasingly universal use of the internet and the dramatic shift away from paper passed hand to hand or mailed physically left many customers feeling cold with the sterile interaction; we wanted our clients to know there were great people on the other side of their laptop screen and telephone call.
Today, we still employ great humans. But the digital age has given way to the Age of Artificial Intelligence, and that old promise is no longer enough. Placing a phone call, once the surest way to connect with a person, may now lead you to the trained voice of an AI agent. Business strategy, marketing content, and website chatbots are all increasingly driven by algorithms. The transformation is so profound that our motto feels less like a differentiator and more like a powerful signal of the disruption ahead.
We are no longer asking what AI can do; we are asking where humans fit into today’s service model.
AI is not coming; it is already here, embedded into the daily operations of service businesses. The real shift is not technological, but philosophical.
We are no longer asking, “Can AI do this?” We are asking, “Should it?” And perhaps more importantly, “What is left for humans to do?”
For service-based businesses, especially those built on trust, relationships, and expertise, this is not a trivial question. It is existential.
AI is quickly becoming part of your operational backbone. That means it belongs in your business continuity planning just as much as your internet connection or your CRM. In the relief of releasing difficult or tedious tasks to AI to accomplish, blind spots can be created: out of sight, out of mind.
Many organizations are quietly building processes that assume AI tools will always be available. But that assumption is fragile.
Consider the recent volatility in the AI market: major providers can change pricing or access overnight, experience outages, face regulatory shutdowns, or shift data policies in ways that violate your compliance needs or corporate ethics.
In regulated environments like defense contracting, this risk is magnified. Simply using the wrong AI tool with controlled unclassified information (CUI) can create a significant compliance violation, as that tool becomes part of your assessed security boundary under CMMC and NIST SP 800-171.
A fundamental strategy for building business resilience in the age of AI is to keep Humans In The Loop (HITL). It embeds human oversight and judgment directly into automated processes, creating a system that is not only efficient but also robust, adaptable, and trustworthy.
At its core, HITL ensures that while AI handles the bulk of the execution, a human is strategically placed to review, correct, and manage exceptions. This creates several layers of resilience that a fully automated, “black box” system cannot provide.
Ask yourself a simple question: If my primary AI tool disappeared tomorrow, could my team still deliver?
If the answer is anything but a confident “yes,” you haven’t built an AI strategy; you’ve inherited a critical vulnerability.
Here’s how HITL directly assists business resilience:
AI models make mistakes. They can “hallucinate” incorrect information, exhibit biases from their training data, or misinterpret nuance in a request. In a business context, these errors can lead to disastrous outcomes like incorrect financial reports, offensive marketing copy, or flawed compliance checks.
A human in the loop acts as a critical quality control checkpoint. They review the AI’s output before it becomes a final product or action. This oversight prevents costly errors, protects the company’s reputation, and mitigates legal and financial risk. It’s the circuit breaker that stops a small AI error from becoming a major business crisis.
Example: An AI system that generates complex sales quotes for custom projects might miss a critical client requirement. The human reviewer (the “loop”) catches the error before the quote is sent, preventing a low-margin deal or a disappointed client.
AI excels at tasks it has been trained on but struggles with “edge cases” like unique situations, or highly nuanced problems it has never experienced. A fully automated system facing an unexpected scenario will either fail, freeze, or provide a nonsensical response.
When the AI encounters a problem it can’t solve, the HITL model automatically escalates it to a human expert. This person can apply context, creativity, and strategic judgment to resolve the issue. This makes the entire business process more adaptable and prevents operations from grinding to a halt when faced with the unexpected.
Example: A customer support chatbot can handle 90% of inquiries. When a customer presents a unique, multi-faceted complaint involving a product defect and a billing error, the system seamlessly transfers the entire context to a senior support agent who can solve the complex problem and retain the customer.
Over-reliance on a single AI provider is a significant risk. If that provider has an outage, changes its terms, or goes out of business, your operations could be crippled. In the same way you have role redundancy with staff, consider an ensemble approach to how you orchestrate multiple Ais to work for you.
Businesses with strong HITL models maintain deep institutional knowledge of their processes. The humans in the loop aren’t just passive reviewers; they are active participants who understand the task’s inputs, logic, and desired outcomes. If the AI tool fails, this team is already trained and positioned to revert to a manual or semi-manual workflow, ensuring business continuity. They are your ultimate insurance policy against AI provider failure.
When a task is fully automated and the people who once performed it are gone, the underlying knowledge of how and why that task is done can erode. The process becomes a black box that no one in the company truly understands, making it nearly impossible to improve or fix.
HITL keeps this vital knowledge alive and evolving. The humans in the loop become experts not just on the old manual process, but on the new AI-assisted process. They understand the AI’s strengths and weaknesses and are best positioned to recommend improvements, train new models, and adapt the workflow as business needs change. This transforms your team from simple doers into system stewards.
For government contractors and regulated industries, the conversation around AI is not just about efficiency, it is about compliance, security, and building defensible trust.
The U.S. lacks a single, comprehensive AI law, but a clear framework of expectations is emerging through executive orders and agency guidance. For contractors, this means:
Data Protection is Non-Negotiable: Public AI tools are generally unsuitable for CUI. Any AI system touching sensitive data must be auditable and governed under CMMC 2.0 and DFARS.
AI-Specific Requirements are Coming: The 2026 National Defense Authorization Act (NDAA) directs the DoD to integrate a formal AI security framework into CMMC, making AI an explicit part of compliance.
Fairness and Governance are Mandated: Federal guidance requires that AI used in hiring be monitored for bias and that all AI systems used by contractors undergo risk assessments.
While the U.S. approach evolves, the EU AI Act has set the global gold standard. It classifies AI by risk and imposes strict controls on high-risk systems used in employment, critical infrastructure, and safety. For a U.S. executive, the EU AI Act isn’t just foreign policy; it’s a preview of future client expectations. Your customers will soon demand this level of governance, regardless of geography.
As regulations evolve, ISO/IEC 42001 (AI Management Systems) is emerging as a key framework for demonstrating responsible governance. ISO 42001 is more than a certificate for your wall; it’s a strategic tool to prove your trustworthiness in a market filled with AI uncertainty. It’s how you move from saying you’re responsible to proving it.
If AI is commoditizing execution, then your most valuable, non-replicable asset becomes your team’s judgment. The winning playbook isn’t about replacing people; it’s about redesigning work to amplify their uniquely human skills. Here’s the blueprint:
AI is not just optimizing service delivery, it is redefining it. Traditional labor-based models are giving way to outcome-based services powered by technology-enabled delivery and continuous, data-driven improvement. The value of service is moving away from execution and toward insight, trust, and relationship.
Our original motto emphasized a timeless truth: people matter.
That has not changed. What has changed is where and how they matter most.
Humans are the stewards of:
AI may answer the phone. AI may write the report. AI may even recommend the strategy or my new company motto. But when something goes wrong, when something is unclear, when something truly matters, people still want a human.
The enduring businesses of tomorrow won’t just use AI to make their people more efficient. They will use AI to make their people more human.