Culture Determines Whether People Click or Not—Not Just Training
Training Teaches. Culture Decides.
When most organizations talk about “culture,” it can feel abstract: values on a wall, slogans in the all-hands deck, a paragraph in the annual report.
We prefer a more practical lens.
We think of your people as running a Human Operating System—the Human OS.
The Human OS is the set of knowledge, skills, awareness, understanding, psychology and behaviors that shape how an individual operates with technology, data, information and systems at work.
It’s how each person:
Looks at the world
Processes information
Makes choices under pressure
Interprets and applies rules
Interacts with colleagues, customers and tools
When you put thousands of Human OSs together in one organization, culture is what emerges from how they interact. Culture is the collective effect of all those individual operating systems bumping into each other every day.
And in a world of always-on connectivity, SaaS everything, and AI everywhere, that Human OS is now one of your most critical security control surfaces.
We’ve spent decades treating laptops, phones, servers and apps as endpoints that need to be inventoried, monitored, patched and protected.
Humans are also endpoints. Just a different kind.
Each person in your organization is an endpoint in the network of:
Access – accounts, roles, permissions, data visibility
Decisions – approvals, exceptions, escalations, “just this once”
Interpretation – how they read alerts, policies, dashboards and AI outputs
Influence – what they tell their team is “normal” or “fine to do here”
If that endpoint’s Human OS is outdated, misaligned or misinformed, you get:
Brilliant people making risky choices with data and AI
Workarounds and shadow workflows that quietly reshape your risk surface
Friction between formal policy and lived reality (“this is how we really do it”)
We like to say: if you don’t patch the Human OS, you’re leaving vulnerabilities all over the environment.
Updating the Human OS is no longer a “nice to have awareness initiative.”
It’s a critical process for any organization that is digital, cloud-based, or AI-enabled (which is… everyone).
If we pop the hood on an individual’s Human OS in the workplace, we see at least four big components:
Knowledge & Skills
What they actually know about security, privacy, AI, data and risk in their role
How comfortable they are with the tools and systems they’re supposed to use
Awareness & Understanding
Do they recognize risky situations when they see them?
Do they understand why the rules are there, not just what the rules say?
Psychology
How they respond to pressure, deadlines, authority and uncertainty
Their default trust level in systems, leaders and tools (including AI)
Their appetite for risk, need for control, fear of getting in trouble
Behaviour
What they actually do in real life—especially when stressed, tired or overloaded
The shortcuts they take, the workarounds they adopt, the norms they reinforce
All of this plays out in the way someone uses technology, handles data, interprets dashboards and interacts with AI systems day to day.
In other words: Human OS = how a person “runs” in your environment.
Now multiply that across hundreds or thousands of people.
Culture is what emerges when lots of Human OSs interact:
The unwritten rules of “how we really do things here”
Who gets listened to, and who gets ignored
How conflict, risk and mistakes are handled
Whether security and AI governance feel like help or hindrance
Your Psychological Perimeter—the boundary where human cognition, emotion and behavior meet systems, data and AI—lives in this interplay between:
Individual Human OSs, and
The culture they collectively create
If your Human OSs are:
Knowledgeable, aware and skilled
Supported by a culture that values secure, thoughtful work
Clear on how to use AI safely and when to question it
…then the Psychological Perimeter becomes a resilient, intelligent control surface.
If your Human OSs are:
Undertrained and overloaded
Operating in a culture of shortcuts, heroics and “don’t ask, just ship”
Unsure about AI rules, but under pressure to use it anyway
…then the Psychological Perimeter becomes a fragile, easily exploited boundary.
AI doesn’t make the Human OS less important—it makes it central.
In an AI-first organization, people are:
Designing and approving AI use cases
Feeding data into AI tools (sometimes safely, sometimes not)
Interpreting AI outputs in decisions, reports and customer interactions
Building and tweaking agentic workflows that may outlive their creators
Without an updated Human OS, you see:
Over-trust in AI (“it must be right”) or total rejection (“I don’t use any of this”)
Shadow AI workflows created by well-meaning, under-supported teams
Misconfigurations in AI platforms that come from uncertainty, not malice
Silent normalization of “everyone pastes a little sensitive data into that free tool”
This is exactly what we describe in our AI Workforce Risk work:
the risk that lives in the gap between what the AI stack can do and what the Human OS is ready for.
If you don’t deliberately update the Human OS, AI adoption will pull it out of date in real time.
Treating the Human OS as an operating system means treating updates as a process, not an occasional event.
That looks like:
Continuous learning, not one-off courses
Role-specific enablement for high-risk roles and AI-heavy workflows
Mindset and decision-making support, not just “don’t click this” rules
Feedback loops from incidents, near-misses and Shadow AI back into training and design
Alignment with culture so that the behaviors you want are actually rewarded
This is where culture work and Human Risk Management Programs meet:
Human Risk Programs define the architecture: what needs to change in knowledge, skills, behaviors and culture.
Culture and leadership shape the runtime: what actually gets reinforced day to day.
The Psychological Perimeter and AI Workforce Risk give you the threat model: where and how human-factored risk will show up.
Updating the Human OS is how you close the loop.
If you want to see how the Human OS fits into the bigger picture:
For the full perimeter view:
Read “The Psychological Perimeter: Human Risk, AI, and the New Frontline of Cybersecurity” for the full model of cognition, culture and AI at the edge of your organization.
For the AI adoption angle:
Read “AI Workforce Risk: The Problem You’ll Only See When It’s Too Late” to understand how unpatched Human OSs quietly undermine AI transformation.
Together, these pieces give you a simple but powerful idea:
your Human OS is now as important to patch and maintain as any technology stack. If you want a secure, AI-enabled, resilient organization, start by updating the operating systems between the keyboards and the chairs.
Training Teaches. Culture Decides.
4 min read
In the intricate labyrinth of cybersecurity, where threats lurk in the shadows of every digital transaction, policies emerge as the guiding beacons,...
6 min read
We’ve spent years building IT operations, security operations and now AI operations. But there’s a missing layer: the operational capability that...
4 min read
Subscribe to our newsletters for the latest news and insights.
Stay updated with best practices to enhance your workforce.
Get the latest on strategic risk for Executives and Managers.