22-25 April 2026

Danger in Dialogue: The Security Risks of Large Language Models

2025

TL; DR

Discover the hidden security risks of Large Language Models and learn practical strategies to protect your organization. This session covers threat vectors, real-world examples, and actionable frameworks for secure LLM implementation.

Session Details

In this comprehensive session, we'll explore the critical security challenges posed by Large Language Models (LLMs) as they become increasingly integrated into business operations and daily life.

From prompt injection attacks to data privacy concerns, we'll examine real-world examples of LLM vulnerabilities and their potential impact on organizations. Participants will learn what strategies exist for identifying, assessing, and mitigating these risks while maintaining the benefits of LLM adoption.

Through demonstrations and case studies, attendees will gain hands-on experience in recognizing common attack vectors and implementing effective security measures.

3 things you'll get out of this session

Understand the major categories of LLM security threats and their potential impact on organizations Learn to identify and analyze common attack vectors in LLM implementations Create an effective risk assessment framework for LLM deployment