AI and cybersecurity

De Wiki Campus Cyber
Aller à :navigation, rechercher
Cette page est une version traduite de la page IA et cybersécurité et la traduction est terminée à 100 %.

Groupe de travail


Working Group objectives and deliverables

The mission of the Working Group, launched as part of Cycle 2 of the Cyber Campus, is to analyze the areas in which cybersecurity and artificial intelligence intersect.

Initial objectives

  • Catalog the potential uses of AI for cybersecurity
  • Create a CYLVIA sandbox environment enabling Cyber Campus members to experiment and evolve in a restricted environment.
  • Create AI training courses for cybersecurity professionals at various levels.
  • Identify and map startups specializing in AI solutions for cybersecurity.
  • Identify cyberthreats and possible mitigations on AI application systems

The security of intelligent systems and their uses in cybersecurity, state of knowledge and skills

GT IA & Cyber - Focus du Groupe de Travail.png

Artificial Intelligence systems are becoming more and more widespread in modern information technology, and have made many advances in recent years. With a view to preparing the next generations of applications and defense systems against cyber threats, the Cyber Campus, through the Artificial Intelligence and Cybersecurity Working Group, is highlighting the main uses of AI to benefit cybersecurity, the limits identified as well as the risks and main measures recommended for the security of these systems.

Which players are involved in Artificial Intelligence projects?

Artificial Intelligence projects involve a wide range of players in order to make the most of the data and respond to the use case identified in advance. Identifying key resources is a major challenge for any cybersecurity expert at the start of a project, and a succinct description of roles and responsibilities helps to quickly identify the stakeholders. The various phases of an AI project, including the development, production launch and monitoring of an AI application, involve many players, some of whom are data science specialists, others not. The diagram below shows the four main families of players identified:

  • "Business" profiles, in charge of the use case and its validation in relation to the company's objectives;
  • Data science specialists, in charge of implementing Artificial Intelligence techniques to meet the needs identified by the "business";
  • IT profiles, in charge of developing and putting into production the use cases designed by the data science specialists;
  • Risk profiles, in charge of monitoring and ensuring compliance with internal policies and external regulations.

It should be noted that each company may have slightly different names for these profiles, and may even divide the roles between different players in different ways. Some of these profiles are cross-functional and work on numerous projects (business sponsor, production platform administrator, data architect and risk profiles), while others are dedicated to a single project and form a project team. Risk profiles must therefore work alongside AI profiles in mixed project teams.

The Working Group also recommends the creation of two complementary positions, référents sécurité en Intelligence Artificielle and Facilitateurs en Intelligence Artificielle.

What contribution can artificial intelligence make to cybersecurity?

Intelligence Artificielle, as a scientific field, has benefited from numerous advances over the last few years, notably thanks to the pooling of techniques, capabilities and data enabling the use of statistical Machine Learning approaches to the detriment of deterministic Expert Systems approaches. This approach enables generalization and brings benefits when a significant amount of data is pooled to limit statistical effects. As a result, cybersecurity research is naturally focused on the following areas:

  • Identification and localization of corporate information, starting with its classification, and the application of dynamic policies;
  • Protection by analyzing weak signals and common behaviors, AI-based systems enable security policies and mechanisms to be configured as effectively as possible;
  • Detection by analyzing the slightest variation in behaviors and changes to infrastructures to dynamically build detection scenarios;
  • 'Recovery through decision support, with detailed knowledge of the information system, to establish the response and recovery strategy associated with a cybersecurity incident.

Based on the challenges and experiments carried out in a number of contexts, the Working Group has identified a set of cybersecurity use cases that can be accelerated or enhanced by the use of Artificial Intelligence. To this end, a framework has been defined to classify and enrich the use cases.

This Use Case Classification Framework is an initiative of the Working Group: it is based on guides and standards in the field, and is intended to be a "practical" way of simply categorizing use cases and encouraging the identification of new ones. In addition to helping to organize knowledge, this framework has been designed to facilitate sharing with international entities. It is built around three axes:

Framework des cas d'usages IA & Cybersécurité
  • categorization of use cases according to the activities of cybersecurity teams, following a typical cybersecurity organization breakdown;
  • categorization of the use case according to the security objectives it seeks to address. The NIST Cybersecurity Framework is used here;
  • reconciling the use case with the attack treatment of MITRE ATT&CK or the implementation of defense elements of MITRE D3FEND.

Cybersecurity activities

The categorization of use cases in relation to the activities of cybersecurity teams has resulted in 5 main categories:

  • Risk and compliance management, including all activities relating to the definition, treatment and management of risks and regulations applicable to organizations;
  • The security assessment, often identified with the work of Red Teams, performing auditing, penetration testing, and organizational compromise activities;
  • Application security, including identity and access management, and application protection (Security by Design, SDLC, DevSecOps, etc.);
  • Infrastructure security, including the implementation of infrastructure cybersecurity solutions (FW, IPS, EDR, Proxy, Directories, etc.) required for in-depth information system security;
  • Cyber defense activities, often identified with the work of Blue Teams, ensuring the operational security of organizations' Information Systems.

Security objectives

Security objectives aim to determine the activities to be implemented in order to control cybersecurity risks within organizations. These are divided into 5 types according to the NIST framework:

  • IDENTIFY with the aim of identifying, evaluating and implementing the appropriate organization to control the information system and deal with risks;
  • PROTECT with the aim of implementing the activities required to protect and maintain the security of information systems;
  • DETECT to identify and qualify cybersecurity incidents;
  • RESPOND with the aim of dealing with cybersecurity incidents: stopping or containing the attack, adapting the security policy or its implementation ;
  • RESET with the aim of returning to a normal operating state following a cybersecurity incident.

Case studies presented by members of the Artificial Intelligence and Cybersecurity Working Group

How to secure applications using artificial intelligence systems

An AI system as seen by cybersecurity experts

The Working Group has identified only the common building blocks required for any AI project, to provide readers with a succinct, pragmatic vision.

An Artificial Intelligence system can be represented as a pipeline made up of 4 key stages:

Modélisation d'un pipeline d'Intelligence Artificielle
  • Exploration is a set of technical components used to analyze data in order to understand its meaning and determine the axes of analysis for the following phases;
  • Training is a set of technical components for modeling data and training an AI model;
  • Evaluation is a set of technical components used to validate the relevance of a trained model in relation to the modeling objectives;
  • Production is a set of technical components enabling the model to infer results from production data.

Applying EBIOS RM to an Artificial Intelligence system

GT IA & Cyber - Les questions à se poser.png

Attacks against ML can be grouped into 3 main families (poisoning, oracle and evasion) and cover the entire lifecycle of an ML-based system.

Attaques sur un systeme IA.png

The working group's research identified 5 different attack patterns applicable to any Artificial Intelligence system:

  • Training data poisoning attack (targeted or untargeted)
  • Data supply attack
  • Attack on library supply chain
  • Oracle attack on production model
  • Evasion attack on production model

Practical solutions to reduce the risk of system compromise

The 4 best practices to put in place to limit the risks associated with AI systems are :

  • Set up an ethics and risk committee to select the use cases for artificial intelligence that are relevant to the company and the authorized data.
  • Inform artificial intelligence and data teams of their rights and obligations when handling data, and the associated risks.
  • Train artificial intelligence and data teams in best practices for secure data handling (anonymization, encryption, etc.), distributed learning techniques and secure development.
  • Define for each project: data that cannot be transmitted to the system, acceptable system deviation/relevance limits and data theft signals (query volume, targeted queries on critical data, etc.).

Measures specific to development, training, validation and production environments:

  • Standardize the anonymization of data collection, storage and transmission encryption.
  • Implement information leak detection scenarios for the entire data pipeline
  • Provide a specific, isolated training environment with restricted access to libraries for those handling the data sets.
  • Regularly validate the security level of open-source libraries and components used in training and validation environments.
  • Build adversarial datasets with identified risk scenarios to validate model changes
  • Launch models in production in secure enclaves with limited execution rights.
  • Monitor systems for deviations