Skip to content

Interactive workflows for creating AI intelligence reports from real-world data sources

License

Notifications You must be signed in to change notification settings

OutsideCollective/intelligence-toolkit

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Intelligence Toolkit

The Intelligence Toolkit is a suite of interactive workflows for creating AI intelligence reports from real-world data sources. It helps users to identify patterns, themes, relationships, and risks within complex datasets, with generative AI (GPT models) used to create reports on findings of interest.

The project page can be found at github.com/microsoft/intelligence-toolkit or aka.ms/itk.

Instructions on how to run and deploy Intelligence Toolkit can be found here.

Transparency FAQ

What can Intelligence Toolkit do?

The Intelligence Toolkit aims to help domain experts make sense of real-world data at a speed and scale that wouldn't otherwise be possible. It was specifically designed for analysis of case, entity, and text data:

  • Case Data
    • Units are structured records describing individual people.
    • Analysis aims to inform policy while preserving privacy.
  • Entity Data
    • Units are structured records describing real-world entities.
    • Analysis aims to understand risks represented by relationships.
  • Text Data
    • Units are collections or instances of unstructured text documents.
    • Analysis aims to retrieve information and summarize themes.

What are Intelligence Toolkit's intended uses?

The Intelligence Toolkit is designed to be used by domain experts who are familiar with the data and the intelligence they want to derive from it. Users should be independently capable of evaluating the quality of data insights and AI interpretations before taking action, e.g., sharing intelligence outputs or making decisions informed by these outputs.

It supports a variety of interactive workflows, each designed to address a specific type of intelligence task:

  • Case Intelligence Workflows
    • Anonymize Case Data generates differentially-private datasets and summaries from sensitive case records.
    • Detect Case Patterns generates reports on patterns of attribute values detected in streams of case records.
    • Compare Case Groups generates reports by defining and comparing groups of case records.
  • Entity Intelligence Workflows
  • Text Intelligence Workflows
    • Query Text Data generates reports from a collection of text documents.
    • Extract Record Data generate schema-aligned JSON objects and CSV records from unstructured text.
    • Generate Mock Data generates mock records and texts from a JSON schema defined or uploaded by the user.

All tutorial data and examples used in Intelligence Toolkit were created for this purpose using the Generate Mock Data workflow.

What workflow should I use?

Use the diagram to identify an appropriate workflow, which can be opened from the left sidebar while running the application.

%%{init: {
  "flowchart": {"htmlLabels": true}} }%%
flowchart TD
    NoData["<b>Input</b>: None"] --> |"<b>Generate Mock Data</b><br/>workflow"| MockData["AI-Generated Records"]
    NoData["<b>Input</b>: None"] --> |"<b>Generate Mock Data</b><br/>workflow"| MockText["AI-Generated Texts"]
    MockText["AI-Generated Texts"] -->  TextDocs["<b>Input:</b> Text Data"]
    MockData["AI-Generated Records"] --> PersonalData["<b>Input</b>: Personal Case Records"]
    MockData["AI-Generated Records"] --> CaseRecords["<b>Input</b>: Case Records"]
    MockData["AI-Generated Records"] --> EntityData["<b>Input</b>: Entity Records"]
    PersonalData["<b>Input</b>: Personal Case Records"] ----> |"<b>Anonymize Case Data</b><br/>workflow"| AnonData["Anonymous Case Records"]
    CaseRecords["<b>Input</b>: Case Records"] ---> HasTime{"Time<br/>Attributes?"}
    HasTime{"Time<br/>Attributes?"} --> |"<b>Detect Case Patterns</b><br/>workflow"| CasePatterns["AI Pattern Reports"]
    CaseRecords["<b>Input</b>: Case Records"] ---> HasGroups{"Grouping<br/>Attributes?"}
    HasGroups{"Grouping<br/>Attributes?"} --> |"<b>Compare Case Groups</b><br/>workflow"| MatchedEntities["AI Group Reports"]
    EntityData["<b>Input</b>: Entity Records"] ---> HasInconsistencies{"Inconsistent<br/>Attributes?"} --> |"<b>Match Entity Records</b><br/>workflow"| RecordLinking["AI-Matched Records"]
    EntityData["<b>Input</b>: Entity Records"] ---> HasIdentifiers{"Identifying<br/>Attributes?"} --> |"<b>Detect Entity Networks</b><br/>workflow"| NetworkAnalysis["AI Network Reports"]
    TextDocs["<b>Input:</b> Text Data"] ---> NeedRecords{"Need<br/>Records?"} --> |"<b>Extract Record Data</b><br/>workflow"| ExtractedRecords["AI-Extracted Records"]
    TextDocs["<b>Input:</b> Text Data"] ---> NeedAnswers{"Need<br/>Answers?"} --> |"<b>Query Text Data</b><br/>workflow"| AnswerReports["AI Answer Reports"]
Loading

How was Intelligence Toolkit evaluated?

The Intelligence Toolkit was designed, refined, and evaluated in the context of the Tech Against Trafficking (TAT) accelerator program with Issara Institute and Polaris (2023-2024). It includes and builds on prior accelerator outputs developed with Unseen (2021-2022) and IOM/CTDC (2019-2020). See this launch blog for more information.

Additionally, a comprehensive system evaluation was performed from the standpoint of Responsible Artificial Intelligence (RAI). This evaluation was carried out utilizing the GPT-4 model. It is important to note that the choice of model plays a significant role in the evaluation process. Consequently, employing a model different from GPT-4 may yield varying results, as each model possesses unique characteristics and processing methodologies that can influence the outcome of the evaluation. Please refer to this Overview of Responsible AI practices for more information.

What are the limitations of Intelligence Toolkit? How can users minimize the impact of these limitations when using the system?

  • The Intelligence toolkit aims to detect and explain patterns, relationships, and risks in data provided by the user. It is not designed to make decisions or take actions based on these findings.
  • The statistical "insights" that it detects may not be insightful or useful in practice, and will inherit any biases, errors, or omissions present in the data collecting/generating process. These may be further amplified by the AI interpretations and reports generated by the toolkit.
  • The generative AI model may itself introduce additional statistical or societal biases, or fabricate information not present in its grounding data, as a consequence of its training and design.
  • Users should be experts in their domain, familiar with the data, and both able and willing to evaluate the quality of the insights and AI interpretations before taking action.
  • The system was designed and tested for the processing of English language data and the creation of English language outputs. Performance in other languages may vary and should be assessed by someone who is both an expert on the data and a native speaker of that language.

What operational factors and settings allow for effective and responsible use of Intelligence Toolkit?

  • The Intelligence Toolkit is designed for moderate-sized datasets (e.g., 100s of thousands of records, 100s of PDF documents). Larger datasets will require longer to process and may exceed the memory limits of the execution environment.
  • Responsible use of personal case data requires that the data be deidentified prior to uploading and then converted into anonymous data using the Anonymize Case Data workflow. Any subsequent analysis of the case data should be done using the anonym case data, not the original (sensitive/personal) case data.
  • It is the user's responsibility to ensure that any data sent to generative AI models is not personal/sensitive/secret/confidential, that use of generative AI models is consistent with the terms of service of the model provider, and that such use incurs per-token costs charged to the account linked to the user-provided API key. Understanding usage costs (OpenAI, Azure) and setting a billing cap (OpenAI) or budget (Azure) is recommended.

What data is collected?

Intelligence Toolkit may be deployed as a desktop application or a cloud service. The application supports short, end-to-end workflows from input data to output reports. As such, it stores no data beyond the use of a caching mechanism for text embeddings that avoids unnecessary recomputation costs. No data is collected by Microsoft or sent to any other service other than the selected AI model API.

What AI model is used?

The system uses the GPT-4 model from OpenAI, either via OpenAI or Azure OpenAI APIs. See the GPT-4 System Card to understand the capabilities and limitations of this model. For models hosted on Azure OpenAI, also see the accompanying Transparency Note.

Disclaimer

  • Intelligence Toolkit is an AI system that generates text.
  • System performance may vary by workflow, dataset, query, and response.
  • Outputs may include factual errors, fabrication, or speculation.
  • Users are responsible for determining the accuracy of generated content.
  • System outputs do not represent the opinions of Microsoft.
  • All decisions leveraging outputs of the system should be made with human oversight and not be based solely on system outputs.
  • The system is only intended to be used for analysis by domain experts capable of evaluating the quality of data insights it generates.
  • Use of the system must comply with all applicable laws, regulations, and policies, including those pertaining to privacy and security.
  • The system should not be used in highly regulated domains where inaccurate outputs could suggest actions that lead to injury or negatively impact an individual's legal, financial, or life opportunities.
  • Intelligence Toolkit is meant to be used to evaluate populations and entities, not individuals, identifying areas for further investigation by human experts.
  • Intelligence Toolkit is not meant to be used as per se evidence of a crime or to establish criminal activity.

Prohibited Uses

All use of Intelligence Toolkit should be consistent with this documentation. In addition, using the system in any of the following ways is strictly prohibited:

  • Pursuing any illegal purpose.
  • Identifying or evaluating individuals.
  • Establishing criminal activity.

Getting Started

You can start using the Intelligence Toolkit as either a web application (with a tool called Docker) or a Python package (via PyPI). Choose one of the options below based on your needs.

Option 1: Using Intelligence Toolkit as a Web Application (via Docker)

To use the Intelligence Toolkit as a web application, you can download and run it using Docker.

1. Install Docker:

Download and install Docker Desktop from docker.com.

Start the Docker Desktop app and make sure it’s running before proceeding.

2. Open Terminal:

Open a terminal according to your OS:

  • If you are using Windows, search for and open the app Windows Powershell in the Windows start menu.

  • If you are using Linux or Mac, search for and open Terminal.

3. Pull the Docker Container:

Download a copy of the Intelligence Toolkit application from GitHub:

docker pull ghcr.io/microsoft/intelligence-toolkit:latest

Note: The image is approximately 2GB, so the download may take some time depending on your internet speed.

4. Run the Docker Container:

Once the download is finished, run the Intelligence Toolkit application using Docker by pasting the following command into your terminal and pressing enter:

docker run -d --name intelligence-toolkit -p 80:80 ghcr.io/microsoft/intelligence-toolkit:latest

5. Access the Web Application:

Open http://localhost:80 in your web browser to start using Intelligence Toolkit.

Note: Docker Desktop App may enter sleep mode if inactive. In this case, open Docker Desktop, select Container in the left menu, then press play on intelligence-toolkit.

6. Setting up the AI model:

Intelligence Toolkit can be used with either OpenAI or Azure OpenAI as the generative AI API.

The Generate Mock Data and Extract Record Data workflows additionally use OpenAI's Structured Outputs API, which requires a gpt-4o model as follows:

  • gpt-4o-mini
  • gpt-4o

You can access the Settings page on the left sidebar when running the application:

  • For OpenAI, you will need an active OpenAI account (create here) and API key (create here).

  • For Azure OpenAI, you will need an active Azure account (create here), endpoint, key and version for the AI Service (create here).

Option 2: Using Intelligence Toolkit as a Python Package (via PyPI)

If you prefer to use Intelligence Toolkit as a Python package, install it directly from PyPI:

pip install intelligence-toolkit

After installation, explore the examples in the example_notebooks folder to get started with various functionalities.


If you have any questions or need further assistance, you can reach out to the project maintainers at [email protected].


NGOs

Non-profit organizations can apply for an annual Azure credit grant of $2,000, which can be used to set up and run an instance of the intelligence-toolkit app for your organization.

Read more about eligibility and registration here

Trademarks

  • This project may contain trademarks or logos for projects, products, or services.
  • Authorized use of Microsoft trademarks or logos is subject to and must follow Microsoft's Trademark & Brand Guidelines.
  • Use of Microsoft trademarks or logos in modified versions of this project must not cause confusion or imply Microsoft sponsorship.
  • Any use of third-party trademarks or logos are subject to those third-party's policies.

Privacy

See Microsoft Privacy Statement.

About

Interactive workflows for creating AI intelligence reports from real-world data sources

Resources

License

Code of conduct

Security policy

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Python 99.8%
  • Other 0.2%