Classification Technical Overview
Transcend uncovers and classify sensitive data across structured, semi-structured, and unstructured data systems, enabling unparalled coverage of data estate and giving you peace of mind.
Structured data is anything that lives in a tradiitonal relational database, characterized by tables with roles and columns. Many companies
Structured Discovery employs an ensemble of five methods for classifying data, applied in the follow˚√ing order:
- Property Name Matching: Identifies data categories based on column/field names
- Regular Expression Content Matching: Uses pattern matching to identify data in specific formats
- Sombra™ Large Language Model (LLM): Uses AI within your environment to analyze data content
- Transcend-Hosted Metadata-Only LLM: Uses AI to analyze metadata without accessing content
- Random Forest ML Model: (deprecated) A legacy classifier which has been phased out in favor of LLM-based approaches
This sequential approach is designed for both efficiency and accuracy. We begin with simpler, more computationally efficient methods for straightforward classifications, and only progress to more complex methods when needed:
- Optimize Resource Usage: Simple methods like Property Name Matching require minimal computational resources and can quickly identify common data types.
- Progressive Complexity: Each subsequent method employs more sophisticated analysis techniques, using increasingly detailed information about the data.
- Fall-Through Design: If a datapoint isn't classified with high confidence by one method, it "falls through" to the next, more nuanced method.
- Balance Speed and Accuracy: This approach ensures we use the least intensive method capable of making an accurate classification.
This method examines the name of each field (column) and matches it against known data categories. For example:
- A column named
email
is classified as containingEMAIL
- A field called
birth_date
matches theDATE_OF_BIRTH
category - Common variations and format differences are considered
- System-specific prefixes and suffixes (like Salesforce's
__c
) are accounted for
Property Name Matching provides high specificity (few false positives) but lower sensitivity (may miss some matches). It typically achieves about 72% accuracy but requires no access to the actual data content.
You can define custom pattern-matching rules to identify specific data formats. To manage these patterns:
- Navigate to the Data Inventory > Data Categories page
- Select a data category
- Add or edit regular expressions for that category

Best practice: Create specific patterns that avoid false positives. For example:
- Good:
/[A-Z0-9a-z._%+-]+@[A-Z0-9a-z.-]+\.[A-Za-z]{2,}/
for email addresses - Problematic:
\d+-\d+-\d+-\d+
for phone numbers (too generic, would match many non-phone numbers)
These custom regular expressions work for both Structured and Unstructured Discovery. Columns matching your patterns will appear with "Regex Matching" as the classification method.
For data not classified by the previous methods, our advanced LLM classifier analyzes both the metadata and content samples. This model:
- Runs within your environment for maximum data security
- Analyzes up to 200 characters of sample data
- Considers column names, table context, and data patterns
- Achieves up to 82% classification accuracy
This approach requires:
- Minimum Classifier version 3.0.6 and Sombra version 7.226.3
- GPU-enabled infrastructure (recommended: AWS
g2.5xlarge
instance)
See our Hosted LLM Classifier documentation for deployment details.
For organizations not ready to host their own LLM, we offer a cloud-based classification option that:
- Analyzes only metadata (table and column names, descriptions)
- Does not process actual data content
- Achieves up to 70% classification accuracy
- Requires no additional infrastructure
This becomes the fallback if the self-hosted LLM is unavailable. While effective, it may occasionally misclassify fields where the name is ambiguous without seeing sample data.
Our legacy classification approach used a statistical machine learning technique called random forests. We're phasing out this method, but you may still see some predictions from it in your results.
Each classification method uses different information sources, as summarized below:
Method | Datapoint Name | Datapoint Description | Category Description | Data Content Samples |
---|---|---|---|---|
Property Name Matching | ✓ | |||
Regular Expression Content Matching | ✓ | |||
Self-Hosted Large Language Model | ✓ | ✓ | ✓ | ✓ |
Transcend-Hosted Metadata-Only LLM | ✓ | ✓ | ✓ |
Many data systems store semi-structured data, such as blobs of JSON or Parquet files. Where possible, Transcend flattens or "unnests" semi-structured data into a more "structured" format, enabling greater granularity in classification.
Any field containing data with a JSON structure will be expanded to generate new derived datapoints. The following steps are involved in the generation of derived datapoints:
- During discovery, we capture the data type of each datapoint.
- During the sampling process, if a datapoint's content is of type JSON, we flatten each sample into unique paths to fields that contain primitive types. Note: we only traverse the JSON structure up to 5 levels deep or until the path length is less than 100 characters.
- We merge paths and their associated values from all samples.
- Each unique path is now considered a derived datapoint and is independently classified to achieve more accurate predictions.
- These derived datapoints are shown as separate datapoints on the dashboard.
- When a parent datapoint is deleted, all of its derived datapoints are also deleted.
- When a derived datapoint is triggered for re-classification, we resample its parent datapoint and regenerate the derived datapoints based on the samples collected.
In the following picture, you can see expanded derived datapoints from c2
datapoint which is a column from a MySQL table containing JSON data.

Unstructured Discovery employs two primary methods to classify personal data in file contents:
This method uses pattern matching to identify specific data formats within file contents. You can configure your own regular expressions to recognize data patterns specific to your organization:
- Navigate to Data Inventory > Data Categories
- Select a data category
- Add or edit regular expressions for that category

Best practice: Create specific patterns that avoid false positives. For example:
- Good:
/[A-Z0-9a-z._%+-]+@[A-Z0-9a-z.-]+\.[A-Za-z]{2,}/
for email addresses - Problematic:
\d+-\d+-\d+-\d+
for phone numbers (too generic, would match many non-phone numbers)
When the system finds content matching your defined patterns, it classifies that content according to the associated data category.
For more sophisticated identification of personal data, Unstructured Discovery can use Named Entity Recognition:
- Uses AI to identify personal data that may not follow strict patterns
- Runs in your environment for maximum data security
- Can identify names, addresses, IDs, and other personal information in context
- Achieves approximately 70% precision and 71% recall in testing
This approach requires:
- Minimum Classifier version 3.3.1 and Sombra™ version 7.261.0
- GPU-enabled infrastructure (recommended: AWS
g2.5xlarge
instance)
For deployment details, see our Hosted LLM Classifier documentation.
After scanning and classifying your unstructured data, you'll see results organized by:
- File location: Where the file containing personal data was found
- File type: The format of the file (PDF, DOCX, etc.)
- Data categories: What types of personal data were identified
- Classification method: Which method identified the personal data
This information helps you understand where sensitive data exists in your unstructured content and take appropriate actions for privacy compliance.