Security | Threat Detection | Cyberattacks | DevSecOps | Compliance

How to Detect Threats to AI Systems with MITRE ATLAS Framework

Cyber threats against AI systems are on the rise, and today’s AI developers need a robust approach to securing AI applications that address the unique vulnerabilities and attack patterns associated with AI systems and ML models deployed in production environments. In this blog, we’re taking a closer look at two specific tools that AI developers can use to help detect cyber threats against AI systems.

Addressing Cyber Risk and the Rise of AI

In this episode of CISO Conversations: EU Data Regulations, Pierre-François Guglielmi, EMEA Field CISO at Rubrik, is joined by Trish McGill, an Executive Subject Matter Expert for Cyber Security IT/OT at De Heus Voeders and Nobian, Brian Wagner, Chief Technology Officer at Revenir, and Tim Clements, Owner of Purpose and Means. Together, they explore the impact of cyber-attacks and data regulations on business resilience, particularly concerning critical infrastructure, and how these factors ultimately affect profits.

AI-Enhanced Cyber Attacks Top the List of Potential Threats Facing Data Security

AI is quickly becoming the basis for more cyber attacks, leading organizations to realize the risk it presents. A new report now shows that AI-enhanced cyber attacks are now the top concern of security leaders. I recently wrote about how prolific ransomware attacks are and what the outcomes were for those experiencing attacks. In the same report - GetApp’s 2024 Data Security report – I also found some interesting data around where AI sits in the list of concerns for cybersecurity leaders.

Using AI Detectors to Identify and Mitigate Harmful Online Content

In today's digital age, online content is generated and shared at an unprecedented rate, leading to a landscape filled with diverse and rich information. However, this constant influx of data has also brought about a surge in harmful content, including hate speech, fake news, cyberbullying, and violent material. These forms of content not only jeopardize user safety but also threaten societal well-being. Addressing this issue requires innovative solutions, and artificial intelligence (AI) detectors have emerged as powerful tools in identifying and mitigating harmful online content.

How to Execute a Secure M365 Copilot Deployment

Microsoft Copilot is a powerful AI assistant that can leverage Microsoft 365 (M365) data from across an organization to generate accurate and relevant insights. But some of that data should be under special lock: you do not want sensitive enterprise information to be used as part of a large language model (LLM). And the reality is that common misconfigurations—such as mislabeled files and overly broad user permissions—can lead to sensitive data exposure to unauthorized users.

What is India's Digital Personal Data Protection (DPDP) Act? Everything You Need to Know!

Data protection has become a critical concern worldwide as digital transactions and data exchanges grow. Countries are establishing strict data protection laws to safeguard personal information, and India is no exception. The Digital Personal Data Protection (DPDP) Act is India’s response to growing privacy concerns and the need for robust regulations around personal data usage.

Massive AI Call Center Data Breach: 10 million Conversations Leaked, Heightening Fraud Risks

In a significant breach, over 10 million customer conversations from an AI-powered call center platform in the Middle East have been exposed. This incident has raised alarm bells regarding the security vulnerabilities of AI platforms widely used in sectors such as fintech and e-commerce. As AI platforms become integral to business operations, the risks of compromised data tracking and brand impersonation have also escalated.