Bulk Enrichment
Enrich existing entities at scale
Learn how to enrich large datasets efficiently, including matching strategies, handling partial matches, and considerations for processing high-volume data.
Prerequisites
Before starting, ensure you have:
- Completed the CSV Import guide
- Understanding of Entities and Sheets
- Your API key (see Authentication)
Understanding Bulk Enrichment
Bulk enrichment adds data to entities that already exist in your sheets or are being imported from CSV files. Unlike discovery searches, enrichment focuses on gathering additional information about known entities.
Enrichment vs Discovery
| Aspect | Discovery | Enrichment |
|---|---|---|
| Entity source | Search criteria | CSV file or existing sheet |
| Primary goal | Find new entities | Add data to known entities |
| Entity matching | N/A | By name, domain, or identifier |
| Typical use | Lead generation | Data enhancement |
Enrichment via ICP Configuration
Define enrichment fields in your ICP's entity target description:
Enrichment-Only ICPs
For pure enrichment (no filtering), omit the Criteria section and only include Enrichment Fields. The AI will focus on researching and populating those fields.
Matching Strategies
When importing CSV data, Linkt matches rows to real-world entities using various strategies.
Match by Company Name
The primary column contains company names:
CSV Example:
Match by Domain
Include website domains for more accurate matching:
Linkt uses the domain to verify entity identity and improve match accuracy.
Match by LinkedIn URL
For person entities, LinkedIn URLs provide the most reliable matching:
Matching Best Practices
| Strategy | Accuracy | Best For |
|---|---|---|
| Name + Domain | Highest | Company enrichment |
| LinkedIn URL | Highest | Person enrichment |
| Name only | Good | When domain unavailable |
| Name + Location | Good | Common company names |
Handling Partial Matches
Not all CSV rows will match entities perfectly. Handle partial matches appropriately.
Match States
| State | Description | Action |
|---|---|---|
completed | Entity matched and enriched | Data saved to sheet |
discarded | No match found | Skipped, logged in queue |
processing | Currently being matched | In progress |
Monitoring Match Status
Check the run queue for match results:
Response:
Reviewing Discarded Rows
Filter to see only discarded items:
Improving Match Rates
- Clean data before import — Fix typos, standardize names
- Include domains — Add website column for companies
- Use LinkedIn URLs — Most reliable for people
- Remove duplicates — Dedupe CSV before upload
- Verify entity existence — Some companies may have been acquired or renamed
Enrichment Data Handling
Understand how enrichment data is stored and merged.
Data Storage Pattern
Each enriched field follows the EntityAttribute structure:
Field Update Behavior
| Scenario | Behavior |
|---|---|
| New field | Created with value |
| Existing field, new value | Updated with new value |
| Existing field, null result | Original value preserved |
| Re-enrichment | Updates timestamps |
Preserving Original Data
Original CSV data is preserved alongside enriched data:
Large Dataset Considerations
Handle high-volume enrichment efficiently.
Batch Size Guidelines
| Dataset Size | Approach | Notes |
|---|---|---|
| 1-100 rows | Single import | Direct processing |
| 100-500 rows | Single import | Monitor queue progress |
| 500-1000 rows | Consider splitting | Better error handling |
| 1000+ rows | Multiple batches | Recommended |
Splitting Large Files
Parallel Processing
Process multiple batches concurrently:
Rate Limiting
Be mindful of API rate limits:
Progress Monitoring
Track progress across large imports:
Complete Example
Full bulk enrichment workflow:
Best Practices
Data Preparation
- Clean your CSV — Remove duplicates, fix obvious errors
- Include identifiers — Add domain or LinkedIn URL when available
- Standardize formats — Consistent company name formatting
- Validate encoding — Ensure UTF-8 encoding
Enrichment Efficiency
- Focus on high-value fields — Don't request unnecessary data
- Batch appropriately — 500-row batches work well
- Monitor progress — Track queue completion rates
- Handle failures — Implement retry logic for transient errors
Quality Assurance
- Spot check results — Review a sample of enriched entities
- Track match rates — Monitor discarded rows
- Validate data quality — Check enrichment accuracy
- Export and backup — Save enriched data externally
Next Steps
- Files Reference — File upload and management API
- CSV Import — Basic import workflow
- Entities Reference — Entity data structure
- Custom Fields — Adding enrichment fields