Skip to main content

2 posts tagged with "deduplication"

View All Tags

· 15 min read
Oleg Kulyk

Building a Web Data Quality Layer: Deduping, Canonicalization, and Drift Alerts

High‑stakes applications of web data – such as pricing intelligence, financial signals, compliance monitoring, and risk analytics – rely not only on acquiring data at scale but on maintaining a high‑quality, stable, and interpretable data layer. Raw HTML or JSON scraped from the web is often noisy, duplicated, and structurally unstable due to frequent site changes. Without a robust quality layer, downstream analytics, ML models, and dashboards are vulnerable to silent corruption.

· 17 min read
Oleg Kulyk

Data Deduplication and Canonicalization in Scraped Knowledge Graphs

As organizations ingest ever-larger volumes of data from the web, they increasingly rely on knowledge graphs (KGs) to model entities (people, organizations, products, places) and their relationships in a structured way. However, web data is heterogeneous, noisy, and heavily duplicated. The same entity may appear thousands of times across sites, with different names, formats, partial data, or conflicting attributes. Without robust deduplication and canonicalization, a scraped knowledge graph quickly becomes fragmented, inaccurate, and operationally useless.