Skip to content
/ PILLAR Public

An AI-Powered Privacy Threat Modeling tool based on the LINDDUN framework by leveraging Large Language Models.

License

Notifications You must be signed in to change notification settings

stfbk/PILLAR

Repository files navigation

P.I.L.L.A.R.

P.I.L.L.A.R. (Privacy risk Identification with LINDDUN and LLM Analysis Report) is a tool which helps developers to identify privacy threats in their software. It uses the LINDDUN methodology executed by Large Language Models (LLMs) to identify privacy threats given the description of the application to analyze or its Data Flow Diagram (DFD). Once privacy threats have been elicited, P.I.L.L.A.R. offers the possibility to assess their impact and suggest control measures, based on the privacy patterns website. Finally, users can download a report of their privacy threat modeling.

The tool is designed to be used by developers and security analysts to easily evaluate the privacy risks of their software and take the necessary actions to mitigate them, offering a good starting point for privacy analysis.

Installation

To install the project, you need to clone the repository and install the dependencies. You can do this by running the following commands:

git clone https://github.com/AndreaBissoli/pillar.git 
cd pillar 
pip install -r requirements.txt 
streamlit run main.py 

Usage

To use the tool, follow the instructions on the web interface. You will need API keys of some LLM providers to use the tool, especially an OpenAI API key, as some features only work with it (for now). You can find more information on how to get an API key in the OpenAI website. Other supported LLM providers are MistralAI and Google Gemini. If you would like to see more LLM providers supported, feel free to open an issue or a pull request.

Features

  • Application Description: Provide a natural language description of your application, specifying the type (e.g., mobile app, web app), the data it collects, and the data policy.
  • Data Flow Diagram (DFD) Management: Upload, edit, generate from an image or description, and download a Data Flow Diagram (DFD) of the application to identify privacy threats.
  • LINDDUN Analysis: Perform a simple LINDDUN analysis of the application in the Threat Model tab.
  • LINDDUN Go Simulation: Conduct a LINDDUN Go simulation for the application in the LINDDUN Go tab, utilizing one or multiple LLM agents.
  • LINDDUN Pro Methodology: Use the LINDDUN Pro methodology with the provided DFD to identify precise privacy threats in the LINDDUN Pro tab.
  • Impact Assessment and Control Measures: Assess the impact of identified privacy threats and suggest control measures based on the privacy patterns website.
  • Report Generation: Download a comprehensive report on the privacy threat modeling of the application.

Roadmap

Here’s an outline of our upcoming features and enhancements for PILLAR:

  • Support for Open Threat Modeling (OTM) Format by adding the ability to export threat modeling results in the widely-adopted OTM format to enhance compatibility with other tools.

  • Locally Hosted Models by implementing support for locally hosted models, such as Ollama, to give users more control over privacy and security when running AI models.

  • Integration with OWASP DFD Generation Framework by exploring integration with the OWASP DFD generating framework (pytm) to improve and extend PILLAR's DFD generation capabilities.


Citation

If you use PILLAR in your research or project, please cite our paper as follows:

@article{pillar,
title={PILLAR: an AI-Powered Privacy Threat Modeling Tool},
author={Mollaeefar, Majid, Bissoli, Andrea and Ranise, Silvio},
journal={arXiv preprint arXiv:2410.08755},
url={https://arxiv.org/abs/2410.08755},
year={2024}
}

License

This project is licensed under the Apache License, Version 2.0 - see the LICENSE file for details.

Contacts

This project is developed by Andrea Bissoli and Majid Mollaeefar as an internship project for Fondazione Bruno Kessler. For any questions or feedback, feel free to contact us.