Designing an AI Driven Sensitive Data Security Tool for Striim

Led the design of Striim's first AI-powered governance feature that automatically discovers, classifies, and protects sensitive data with confidence scoring to build user trust.

Led the design of Striim's first AI-powered governance feature that automatically discovers, classifies, and protects sensitive data with confidence scoring to build user trust.

Led the design of Striim's first AI-powered governance feature that automatically discovers, classifies, and protects sensitive data with confidence scoring to build user trust.

Overview

In today's data-driven landscape, organizations process massive volumes of streaming data containing sensitive information such as personally identifiable information (PII), payment card data, social security numbers, and other regulated data types.

As data flows through real-time streaming pipelines, companies face a critical challenge: how to automatically discover, classify, and protect sensitive data in streaming data pipelines.

With Striim serving customers in the finance and healthcare sectors it was uniquely positioned to help prevent hefty fines by improving governance and security postures when moving data from source to target systems.

What is Striim ?

Striim is a platform for building data pipelines—connecting sources, transformations, and targets. Like a conveyor belt, it moves data between systems (such as databases and cloud services) in real time to power analytics and AI.

My role

My Role: Lead UX Designer

Team: Product Manager, Engineering Team, User Researcher

Key Contributions: Design strategy, AI design patterns, interaction design, content design, visual design, user research consultation.

Research & Discovery

Background Research

As this feature encompassed two areas that were unfamiliar to me—sensitive data governance and AI—I started with extensive research. This included reviewing documentation, watching video tutorials on Data Security Posture Management and AI tools, and conducting a competitive analysis to establish our initial product vision.

User Research

We conducted interviews with customers in the financial sector to understand:

  • How organizations identify and classify sensitive data

  • Storage practices, user roles, and access levels

  • Breach implications and regulatory requirements

  • How Striim could enhance sensitive data handling

The Key Findings

Organizations struggle with these critical problems when handling sensitive data in streaming pipelines:

  1. Invisible sensitive data - Limited visibility into what sensitive data exists within sources. This can pose a significant challenge if there are data quality issues.

  2. Compliance risk - Potential violations of GDPR, CCPA, HIPAA, and PCI-DSS regulations. Companies must constantly adapt to regulatory changes.

  3. Manual, inconsistent classification – Identifying and protecting sensitive data is time-consuming and error-prone. Organizations rely on a mix of industry standards and custom policies, with classification levels that vary by sector and region.

Poor handling of sensitive data leads to reputation loss, diminished client trust, and significant legal and financial consequences.

Usability Testing Approach

Due to tight timelines and limited customer access, we combined discovery research with usability testing during rapid design iterations.

Research Methods:

  • Structured interviews for initial insights and context

  • Think-aloud exercises with UI mockups for unbiased user perceptions

  • Modified desirability study using 34 positive/negative words to describe design feelings

This approach revealed customer feedback and uncovered previously unknown user needs.

AI Design Strategy

My design approach was heavily influenced by Google's AI design guidelines, focusing on three key principles:

1. Help Users Calibrate Trust

Challenge: Users need to trust AI results without blind acceptance

Solution:

  • Communicate product capabilities and limitations upfront

  • Display privacy and security settings prominently

  • Include warnings about not blindly trusting results

  • Establish trust from the beginning and maintain it throughout the experience

2. Optimize for Understanding

Challenge: Users need to understand how AI classifications work

Solution:

  • Provide in-the-moment explanations of AI system functionality

  • Explain why the AI provided specific outputs

  • Use clear, comprehensive language for complex AI concepts

3. Manage Influence on User Decisions

Challenge: Users need to know when to trust system predictions

Solution: Display model confidence to help users make informed decisions

Key Design Innovation: Confidence Scoring

Through usability testing, we discovered users needed confidence scores to calibrate their trust in AI results. I spearheaded the initiative to implement this feature, working closely with engineering and product management.

Our Implementation:

  1. Clear sample size display - Show the amount of data analyzed

  2. Statistical confidence measure - For example: "Column A contains SSNs with 80% confidence based on analysis of 100 records where 80 contained SSNs"

This approach provided users with the context needed to make informed decisions about sensitive data classification.

The Designs

Sherlock: The Discovery Experience

User Story: As a Striim user with appropriate privileges, I want Striim to discover and report on the sensitive data that can flow into my pipeline from my data sources.

The discovery interface provides:

  • Automated scanning of sources in pipelines

  • Clear visualization of sensitive data findings

  • Confidence scores for each classification

  • Comprehensive reporting capabilities

Sentinel: The Protection Experience

User Story: As a Striim user with appropriate privileges, I want to protect the sensitive data that flows in my pipelines and use the sensitive data discovery report to do so.

The protection interface enables:

  • Policy-based data protection rules

  • Real-time masking and encryption options

  • Automated configuration based on the discovery findings

Use the discovery report to inform the pipeline design
Add a Sentinel that continuously detects sensitive data and protects it
Monitor the actions taken by Sentinel in real-time
Explanation of how Sentinel works (Source: Striim documentation)

Creative Constraints

Working within Striim's existing canvas experience and limited configuration panel space required creative solutions for complex data governance configurations while maintaining usability.

I took the opportunity to introduce a resizable configuration panel, allowing users to expand it and better focus on their configuration tasks. View my case study on how I reimagined the entire data pipeline creation experience here.

Search for a sensitive data category (identifier)
Adding an identifier
Removing an identifier

Impact & Results

The AI-driven sensitive data discovery platform delivered significant value:

Business Impact:

  • Positioned Striim as an AI-forward platform in the data integration space

  • Provided customers with a competitive advantage in data governance

User Experience Impact:

  • Transformed complex compliance tasks into intuitive workflows

  • Increased user confidence through transparent AI decision-making

  • Reduced time-to-value for sensitive data protection implementation

  • Set the foundation for future user experience enhancements in the pipeline creation workflow.

Reflections

This project presented unique challenges that informed my approach to product design:

What I Learned:

  • Tight timelines require creative research methods - Combining discovery and usability testing was effective but required careful planning

  • AI design requires extra attention to trust-building - Users need confidence indicators and clear explanations to trust automated systems

  • Cross-functional alignment is crucial - When everyone is learning a new domain, clear communication and shared objectives become even more important

What I'd Do Differently:

If I were to revisit this project, I’d prioritize dedicated user research phases and establish clearer frameworks for cross-functional alignment—especially when working in unfamiliar technical domains. I’d also streamline the configuration experience by starting with a simple setup and introducing complexity gradually, based on user feedback, to avoid overwhelming users.


While there's room for improving both the UX process and final designs, this project was an invaluable learning experience in building AI-driven experiences. The tight constraints taught me to rapidly synthesize complex technical requirements into user-centered solutions and effectively communicate concepts to stakeholders.