Malicious Insiders & the Google AI Case- What You Need to Do To Mitigate Insider Threats Through Positive Cyber Culture
The New Frontline in Cybersecurity
Kate Goldman Apr 26, 2024 12:00:00 AM
As we all know, the need for cybersecurity is still on the rise, and I’d argue that it's become increasingly clear how our traditional, tech-centric defenses aren't enough. The recent report from the Cyber Safety Review Board on the Microsoft breach brought this into sharp focus. This wasn't just about technological failures; it was also about human errors and organizational culture—factors that traditional cybersecurity strategies often overlook, push to the side, or discount as ‘not as important to the mission of information security’.
I encourage every leader in the cybersecurity field to take a moment in the next few months to re-evaluate the very foundation of your human risk management, cyberculture, and cyber awareness programs. It’s time to shift the perspective from viewing the human aspect of digital protection as an add-on to recognizing it as a cornerstone of our security posture. The insights from recent reports make it clear that enhancing our understanding of human factors isn't just beneficial—it's essential. As budgets are prepared for 2025, now is the time to consider what resources and changes are necessary to elevate your approach.
The report contained several key points regarding organizational and cybersecurity culture, human risk, psychology, decision-making, and cybersecurity policy, especially highlighting the broader implications of Microsoft's security practices and transparency. But, we’re also talking about a major tech company that, I guarantee, dedicates more time and money to cybersecurity than most other companies out there. It’s a pickle.
Here are some key takeaways to consider:
This all brings to light the complex interplay between cybersecurity, corporate culture, and the broader implications of security breaches in the modern digital landscape. But to frame this in the context of human risk, human factors, and human error, let’s go one layer deeper.
Unpacking a few key phrases from the CRSB Report on the Microsoft Breach:
This leads to a few particular questions:
So where to begin? First, is to understand that answering these questions can be done, should be done, and can be quantified, qualified, and analyzed for key risk factors within your organization.
The study of human risk and error is built upon a foundation laid by influential thinkers in psychology and safety science. James Reason's Swiss Cheese Model revolutionized our understanding of systemic failures, illustrating how human errors pass through multiple layers of defenses, ultimately leading to accidents. Daniel Kahneman’s exploration of cognitive biases through his Dual-Process Theory shed light on how quick, intuitive decisions (System 1) and slower, analytical thinking (System 2) can lead to errors in high-stakes environments. Meanwhile, Charles Perrow's Normal Accident Theory and Sidney Dekker's work on human factors have challenged and expanded our notions of error and blame, suggesting that what often appears as human error is a consequence of complex system designs and interactions.
While these frameworks provide invaluable insights, they also reflect a tension between understanding human behavior as a variable we can control and as an intrinsic, often unpredictable part of complex systems. This historical perspective shows that while the science behind human risk and error is robust, it’s not infallible, urging us to continually challenge and refine our approaches to human risk thinking and change in the organizational context.
In the landscape of cybersecurity, understanding the nature of errors—whether intentional or unintentional—is crucial for developing effective security measures. The distinction between these types of errors is not just academic but has significant practical implications for how we design our insider threat programs and overall security posture.
For instance, Intentional vs. Unintentional Actions:
When the Cyber Safety Review Board mentions "avoidable errors," it raises questions about the nature of the errors and the measures that could have been in place to prevent them. In a legal context, the term often implies negligence or a failure to act with the prudence that a reasonable person would exercise under similar circumstances. This is critical for CISOs to understand as it directly relates to the accountability and liability of the organization.
To effectively manage human risks, CISOs should adopt a more nuanced approach that considers psychological, organizational, and cultural factors. This involves:
By viewing "avoidable" and "preventable" errors through this comprehensive lens, CISOs can better prepare their organizations to handle the complex human factors that influence cybersecurity. This approach not only addresses the immediate risks but also contributes to building a more robust and resilient security culture.
As we reflect on the evolution of cybersecurity, it's clear that our traditional approaches have primarily focused on technological defenses—firewalls, anti-virus software, and encryption. This perspective often attributes cybersecurity failures predominantly to human error, with an emphasis on therefore controlling employee behavior through stringent rules and punitive measures. It's a reactive approach that treats symptoms rather than causes, assuming that tighter control and harsher penalties will minimize risks.
However, this is inadequate for addressing the complexities and realities of modern digital environments. We’re here to advocate for a more holistic approach. By recognizing human behavior not just as a potential liability but as a pivotal element in the cybersecurity ecosystem, we acknowledge that behaviors are often symptoms of broader systemic issues—be they poorly designed systems that invite error, unrealistic policy expectations, or a workplace culture that prioritizes speed over security.
The overwhelming focus in cybersecurity has traditionally been on the technical layers of defense. It’s estimated that 99.9% of professionals in the field come from technical backgrounds, often concentrating on the seven layers of the technology stack to defend against cyber adversaries. This technical bias overlooks the crucial eighth layer: the human layer. Here, human decisions and actions play a significant role in both creating vulnerabilities and mitigating them.
Understanding the 'why' behind actions is essential for developing more effective and sustainable security strategies. By investigating the root causes of behaviors, organizations can design systems and processes that naturally encourage secure practices. This might include more intuitive system interfaces that make secure choices easy and natural for users, or policy frameworks that are adaptable to the real-world pressures employees face.
Cultivating a positive security culture is central to this new approach. Instead of relying solely on compliance and fear of retribution, a positive culture encourages engagement with security practices through understanding, commitment, and shared responsibility.
As we move forward, CISOs and cybersecurity professionals must integrate these new perspectives into their strategic planning. This means prioritizing investments not only in technological defenses but also in understanding and shaping human behavior within digital systems. By aligning our cybersecurity strategies with the complexities of human behavior, we can build more resilient organizations that are better equipped to face the challenges of the digital age.
Transitioning from old to new thinking in cybersecurity isn't just a shift in tactics; it's a fundamental change in philosophy. By embracing this change, we can transform our approach from one that merely reacts to breaches to one that proactively strengthens the human foundations of our digital defenses.
The New Frontline in Cybersecurity
4 min read
In the dynamic realm of cybersecurity, the term "data breach" echoes the potential for causing harm, chaos, and compromise. It's more than just a...
7 min read
Rational Choices vs. Emotional Undertones: Navigating Human Decision Making What are human risks in cyber security management? To make models work,...
8 min read