Data Quality (DQ)

When it comes to data quality, how good is good enough?

The hallmark of data quality is how well data supports the context in which it’s consumed. Your legal department, for example, may use "Informatica Company” while your finance department uses "INFA,” and both records are of equal quality.

Quality is a relative and never-ending judgment, one that needs to be defined by the business (or business unit) that’s consuming the data. An essential element of holistic data governance, trustworthy data serves critical business needs across the enterprise—from legal to finance to marketing and beyond.

Driving data quality requires a repeatable process that includes:


  • - Defining the specific requirements for "good data,” wherever it’s used.
  • - Establishing rules for certifying the quality of that data.
  • Integrating those rules into an existing workflow to both test and allow for exception handling.
  • - Continuing to monitor and measure data quality during its lifecycle (usually done by data stewards).


And because rules and needs change and new systems can be added to the mix, truly successful data quality initiatives need to be scalable to address those new requirements.


Bad data costs U.S. businesses $600 billion a year, according to the research firm TDWI.

Why data quality matters

Why does quality data matter? An often-cited statistic puts the cost of "bad” data to U.S. businesses at $600 billion annually1. Whether bad data causes you to lose revenue, damages your brand, reduces your competitive edge, or simply results in bad decision-making, the costs are significant.

When looking for a data quality solution, we recommend you put the following at the top of your "must-haves” list:


  • Reliability: The demand for reliable data quality is a critical—and persistent—need. Your business will want a vendor with a successful track record as well as a comprehensive approach to ensure a holistic data quality process that supports the key steps around discovering, defining, applying, monitoring, and measuring progress.
  • Portability: Whether you’re using legacy mainframes or the latest technology, you’ll want a data quality tool that can evolve as your business does. That means the data quality solution you deploy should allow you to scale across platforms—from on-premise to the cloud, in Hadoop, or in a hybrid ecosystem.


Why Informatica?

We’ve pioneered the categories of data integration and data quality since 1993 and offer a full suite of data quality and data enrichment—or Data as a Service(DaaS)—products that ensures consistent quality throughout data’s lifecycle—as it enters your systems, is analyzed mid-stream, and when it’s stored or archived—on premise, in the cloud, or on Hadoop.

  • Embedded data stewardship: Our data management platform enables collaboration across your entire infrastructure, not just within specific tools and use cases.
  • "Define once, deploy anywhere”: With Informatica Data Quality, you only need to define your business rules once, and they’ll be applied consistently anywhere—on premises, in the cloud, on Hadoop.
  • DaaS excellence: We ensure you can communicate with your customers using accurate contact record verification and superior data enrichment services.