Classification Technical Overview

Transcend uncovers and classify sensitive data across structured, semi-structured, and unstructured data systems, enabling unparalled coverage of data estate and giving you peace of mind.

Structured data is anything that lives in a tradiitonal relational database, characterized by tables with roles and columns. Many companies

Structured Discovery employs an ensemble of five methods for classifying data, applied in the follow˚√ing order:

  1. Property Name Matching: Identifies data categories based on column/field names
  2. Regular Expression Content Matching: Uses pattern matching to identify data in specific formats
  3. Sombra™ Large Language Model (LLM): Uses AI within your environment to analyze data content
  4. Transcend-Hosted Metadata-Only LLM: Uses AI to analyze metadata without accessing content
  5. Random Forest ML Model: (deprecated) A legacy classifier which has been phased out in favor of LLM-based approaches

This sequential approach is designed for both efficiency and accuracy. We begin with simpler, more computationally efficient methods for straightforward classifications, and only progress to more complex methods when needed:

  • Optimize Resource Usage: Simple methods like Property Name Matching require minimal computational resources and can quickly identify common data types.
  • Progressive Complexity: Each subsequent method employs more sophisticated analysis techniques, using increasingly detailed information about the data.
  • Fall-Through Design: If a datapoint isn't classified with high confidence by one method, it "falls through" to the next, more nuanced method.
  • Balance Speed and Accuracy: This approach ensures we use the least intensive method capable of making an accurate classification.

This method examines the name of each field (column) and matches it against known data categories. For example:

  • A column named email is classified as containing EMAIL
  • A field called birth_date matches the DATE_OF_BIRTH category
  • Common variations and format differences are considered
  • System-specific prefixes and suffixes (like Salesforce's __c) are accounted for

Property Name Matching provides high specificity (few false positives) but lower sensitivity (may miss some matches). It typically achieves about 72% accuracy but requires no access to the actual data content.

You can define custom pattern-matching rules to identify specific data formats. To manage these patterns:

  1. Navigate to the Data Inventory > Data Categories page
  2. Select a data category
  3. Add or edit regular expressions for that category
Adding custom regular expressions to help with classification

Best practice: Create specific patterns that avoid false positives. For example:

  • Good: /[A-Z0-9a-z._%+-]+@[A-Z0-9a-z.-]+\.[A-Za-z]{2,}/ for email addresses
  • Problematic: \d+-\d+-\d+-\d+ for phone numbers (too generic, would match many non-phone numbers)

These custom regular expressions work for both Structured and Unstructured Discovery. Columns matching your patterns will appear with "Regex Matching" as the classification method.

For data not classified by the previous methods, our advanced LLM classifier analyzes both the metadata and content samples. This model:

  • Runs within your environment for maximum data security
  • Analyzes up to 200 characters of sample data
  • Considers column names, table context, and data patterns
  • Achieves up to 82% classification accuracy

This approach requires:

  • Minimum Classifier version 3.0.6 and Sombra version 7.226.3
  • GPU-enabled infrastructure (recommended: AWS g2.5xlarge instance)

See our Hosted LLM Classifier documentation for deployment details.

For organizations not ready to host their own LLM, we offer a cloud-based classification option that:

  • Analyzes only metadata (table and column names, descriptions)
  • Does not process actual data content
  • Achieves up to 70% classification accuracy
  • Requires no additional infrastructure

This becomes the fallback if the self-hosted LLM is unavailable. While effective, it may occasionally misclassify fields where the name is ambiguous without seeing sample data.

Our legacy classification approach used a statistical machine learning technique called random forests. We're phasing out this method, but you may still see some predictions from it in your results.

Each classification method uses different information sources, as summarized below:

MethodDatapoint NameDatapoint DescriptionCategory DescriptionData Content Samples
Property Name Matching
Regular Expression Content Matching
Self-Hosted Large Language Model
Transcend-Hosted Metadata-Only LLM

Many data systems store semi-structured data, such as blobs of JSON or Parquet files. Where possible, Transcend flattens or "unnests" semi-structured data into a more "structured" format, enabling greater granularity in classification.

Any field containing data with a JSON structure will be expanded to generate new derived datapoints. The following steps are involved in the generation of derived datapoints:

  1. During discovery, we capture the data type of each datapoint.
  2. During the sampling process, if a datapoint's content is of type JSON, we flatten each sample into unique paths to fields that contain primitive types. Note: we only traverse the JSON structure up to 5 levels deep or until the path length is less than 100 characters.
  3. We merge paths and their associated values from all samples.
  4. Each unique path is now considered a derived datapoint and is independently classified to achieve more accurate predictions.
  5. These derived datapoints are shown as separate datapoints on the dashboard.
  6. When a parent datapoint is deleted, all of its derived datapoints are also deleted.
  7. When a derived datapoint is triggered for re-classification, we resample its parent datapoint and regenerate the derived datapoints based on the samples collected.

In the following picture, you can see expanded derived datapoints from c2 datapoint which is a column from a MySQL table containing JSON data.

Derived datapoints from a JSON column

Unstructured Discovery employs two primary methods to classify personal data in file contents:

This method uses pattern matching to identify specific data formats within file contents. You can configure your own regular expressions to recognize data patterns specific to your organization:

  1. Navigate to Data Inventory > Data Categories
  2. Select a data category
  3. Add or edit regular expressions for that category
Adding custom regular expressions to help with classification

Best practice: Create specific patterns that avoid false positives. For example:

  • Good: /[A-Z0-9a-z._%+-]+@[A-Z0-9a-z.-]+\.[A-Za-z]{2,}/ for email addresses
  • Problematic: \d+-\d+-\d+-\d+ for phone numbers (too generic, would match many non-phone numbers)

When the system finds content matching your defined patterns, it classifies that content according to the associated data category.

For more sophisticated identification of personal data, Unstructured Discovery can use Named Entity Recognition:

  • Uses AI to identify personal data that may not follow strict patterns
  • Runs in your environment for maximum data security
  • Can identify names, addresses, IDs, and other personal information in context
  • Achieves approximately 70% precision and 71% recall in testing

This approach requires:

  • Minimum Classifier version 3.3.1 and Sombra™ version 7.261.0
  • GPU-enabled infrastructure (recommended: AWS g2.5xlarge instance)

For deployment details, see our Hosted LLM Classifier documentation.

After scanning and classifying your unstructured data, you'll see results organized by:

  • File location: Where the file containing personal data was found
  • File type: The format of the file (PDF, DOCX, etc.)
  • Data categories: What types of personal data were identified
  • Classification method: Which method identified the personal data

This information helps you understand where sensitive data exists in your unstructured content and take appropriate actions for privacy compliance.