banner



Which Of The Following Are Genuine Data-mining Procedures? (Please Check All That Apply)

Epitome by Gerd Altmann from Pixabay (CC0)

vii Steps to Ensure and Sustain Data Quality

Several years ago, I met a senior managing director from a large company. He mentioned the company he worked for was facing information quality issues that eroded customer satisfaction, and he had spent months investigating the potential causes and how to fix them. "What have you found?" I asked eagerly. "It is a tough consequence. I did not find a single crusade, on the contrary, many things went wrong," he replied. He and then started citing a long listing of what contributed to the data quality problems — almost every section in the company was involved and it was hard for him to decide where to brainstorm next. This is a typical example when dealing with Data Quality, which direct relates to how an organization is doing its business organization and the entire life bike of the data itself.

Earlier data scientific discipline became mainstream, data quality was mostly mentioned for the reports delivered to internal or external clients. Nowadays, because machine learning requires a large corporeality of training data, the internal datasets inside an organization are in high need. In addition, the analytics are always hungry for information and constantly search for data avails that tin potentially add value, which has led to quick adoption of new datasets or data sources non explored or used before. This trend has made data management and good practices of ensuring good data quality more important than ever.

The goal of this article is to give you a articulate idea of how to build a information pipeline that creates and sustains good data quality from the offset. In other words, data quality is not something that tin can be fundamentally improved past finding bug and fixing them. Instead, every arrangement should start past producing data with skilful quality in the first place.

First of all, what is Data Quality? Generally speaking, information is of loftier quality when it satisfies the requirements of its intended apply for clients, decision-makers, downstream applications and processes. A practiced illustration is the quality of a product produced past a manufacturer, for which adept product quality is non the business outcome, but drives customer satisfaction and impacts the value and life cycle of the product itself. Similarly, the quality of the data is an important attribute that could drive the value of the data and, hence, touch on aspects of the business consequence, such every bit regulatory compliance, client satisfaction, or accuracy of decision making. Below lists v main criteria used to measure data quality:

  • Accurateness: for any data described, it needs to be authentic.
  • Relevancy: the data should come across the requirements for the intended use.
  • Completeness: the data should non accept missing values or miss data records.
  • Timeliness: the data should be up to date.
  • Consistency:the data should have the information format as expected and tin be cross reference-able with the same results.

The standard for good data quality tin differ depending on the requirement and the nature of the data itself. For example, the core customer dataset of a company needs to encounter very loftier standards for the above criteria, while there could exist higher tolerance of errors or incompleteness for a third-party information source. For an organisation to deliver data with expert quality, it needs to manage and control each data storage created in the pipeline from the beginning to the end. Many organizations only focus on the concluding data and invest in data quality control endeavour right before it is delivered. This is non adept enough and too often, when an event is found in the finish, it is already too belatedly — either it takes a long fourth dimension to find out where the problem came from, or it becomes too costly and time consuming to prepare the issue. However, if a visitor can manage the information quality of each dataset at the time when it is received or created, the data quality is naturally guaranteed. There are 7 essential steps to making that happen:

one. Rigorous information profiling and control of incoming data

In most cases, bad data comes from data receiving. In an organization, the data usually comes from other sources outside the command of the company or section. It could be the data sent from another organization, or, in many cases, collected by third-party software. Therefore, its data quality cannot exist guaranteed, and a rigorous information quality control of incoming data is maybe the most important aspect among all information quality control tasks. A good data profiling tool so comes in handy; such a tool should be capable of examining the following aspects of the data:

  • Data format and information patterns
  • Data consistency on each record
  • Data value distributions and abnormalies
  • Completeness of the information

It is as well essential to automate the information profiling and data quality alerts then that the quality of incoming data is consistently controlled and managed whenever it is received — never presume an incoming data is as expert as expected without profiling and checks. Lastly, each piece of incoming data should exist managed using the same standards and best practices, and a centralized catalog and KPI dashboard should be established to accurately record and monitor the quality of the data.

ii. Careful data pipeline design to avoid duplicate data

Indistinguishable information refers to when the whole or part of information is created from the same data source, using the same logic, but by different people or teams likely for different downstream purposes. When a duplicate data is created, it is very likely out of sync and leads to different results, with cascading effects throughout multiple systems or databases. At the terminate, when a data issue arises, it becomes hard or time-consuming to trace the root cause, non to mention fixing it.

In order for an organization to prevent this from happening, a data pipeline needs to be clearly divers and carefully designed in areas including information assets, information modeling, business concern rules, and compages. Constructive communication is also needed to promote and enforce data sharing within the arrangement, which will ameliorate overall efficiency and reduce any potential information quality issues caused by data duplications. This gets into the cadre of data management, the details of which are beyond the telescopic of this article. On a high level, there are 3 areas that need to exist established to foreclose duplicate data from being created:

  1. A data governance programme, which conspicuously defines the ownership of a dataset and effectively communicates and promotes dataset sharing to avert any department silos.
  2. Centralized data assets management and data modeling, which are reviewed and audited regularly.
  3. Clear logical design of data pipelines at the enterprise level, which is shared across the system.

With today's rapid changes in technology platforms, solid information management and enterprise-level data governance are essential for hereafter successful platform migrations.

iii. Accurate gathering of data requirements

An important attribute of having adept data quality is to satisfy the requirements and deliver the information to clients and users for what the information is intended for. It is non every bit simple as it first sounds, because:

  • It is non easy to properly present the data. Truly agreement what a customer is looking for requires thorough data discoveries, data assay, and clear communications, often via data examples and visualizations.
  • The requirement should capture all data conditions and scenarios — it is considered incomplete if all the dependencies or weather condition are not reviewed and documented.
  • Clear documentation of the requirements, with piece of cake access and sharing, is another important aspect, which should be enforced past the Data Governance Committee.

The role of Business Analyst is essential in requirement gathering. Their agreement of the clients, as well every bit current systems, allows them to speak both sides' languages. Later on gathering the requirements, business analysts too perform impact analysis and help to come upwards with exam plans to make sure the data produced meets the requirements.

4. Enforcement of data integrity

An of import characteristic of the relational database is the power to enforce data Integrity using techniques such as foreign keys, check constraints, and triggers. When the data volume grows, along with more and more data sources and deliverables, not all datasets can live in a unmarried database system. The referential integrity of the data, therefore, needs to be enforced by applications and processes, which need to be defined by all-time practices of data governance and included in the design for implementation. In today's big information world, referential enforcement has become more and more difficult. Without the mindset of enforcing integrity in the commencement place, the referenced data could become out of date, incomplete or delayed, which so leads to serious data quality issues.

5. Integration of information lineage traceability into the data pipelines

For a well-designed data pipeline, the fourth dimension to troubleshoot a data issue should non increase with the complexity of the arrangement or the volume of the data. Without the data lineage traceability built into the pipeline, when a information consequence happens, it could take hours or days to track down the cause. Sometimes it could get through multiple teams and crave information engineers to look into the code to investigate.

Data Lineage traceability has 2 aspects:

  • Meta-data: the ability to trace through the relationships betwixt datasets, data fields and the transformation logic in betwixt.
  • Data itself: the ability to trace a information issue quickly to the individual record(southward) in an upstream data source.

Meta-information traceability is an essential part of effective information governance. This is enabled by clear documentation and modeling of each dataset from the beginning, including its fields and structure. When a data pipeline is designed and enforced by the information governance, the meta-data traceability should be established at the same time. Today, meta-data lineage tracking is a must-accept capability for any data governance tool on the market, which makes it easier to store and trace through datasets and fields by a few clicks, instead of having information experts go through documents, databases, and even programs.

Information traceability is more than difficult than meta-information traceability. Beneath lists some common techniques to enable this power:

  1. Trace by unique keys of each dataset: This offset requires each dataset has 1 or a grouping of unique keys, which is then carried downwardly to the downstream dataset through the pipeline. All the same, not every dataset can be traced by unique keys. For example, when a dataset is aggregated, the keys from the source become lost in the aggregated data.
  2. Build a unique sequence number, such as transaction identifier or record identifier when there are no obvious unique keys in the data itself.
  3. Build link tables when there are many-to-many relationships, merely not 1-to-1or 1-to-many.
  4. Add timestamp (or version) to each information record, to indicate when information technology is added or changed.
  5. Log data change in a log table with the value earlier a change and the timestamp when the change happens

Information traceability takes time to design and implement. It is, still, strategically critical for data architects and engineers to build information technology into the pipeline from the kickoff; it is definitely worth the effort considering it will relieve a tremendous amount of time when a information quality consequence does happen. Furthermore, data traceability lays the foundation for farther improving data quality reports and dashboards that enables one to find out data bug earlier earlier the data is delivered to clients or internal users.

vi. Automated regression testing as part of change management

Plainly, data quality issues oft occur when a new dataset is introduced or an existing dataset is modified. For effective alter management, test plans should be built with two themes: 1) confirming the change meets the requirement; two) ensuring the change does non have an unintentional impact on the data in the pipelines that should not be changed. For mission critical datasets, when a change happens, regular regression testing should be implemented for every deliverable and comparisons should be done for every field and every row of a dataset. With the rapid progress of technologies in big data, system migration constantly happens in a few years. Automated regression test with thorough data comparisons is a must to make sure adept data quality is maintained consistently.

seven. Capable information quality control teams

Lastly, 2 types of teams play disquisitional roles to ensure high information quality for an organization:

Quality Assurance: This team checks the quality of software and programs whenever changes happen. Rigorous change management performed by this team is essential to ensure data quality in an organization that undergoes fast transformations and changes with information-intensive applications.

Production Quality Control: Depending on an system, this squad does not have to be a split team by itself. One-time it can be a office of the Quality Assurance or Business organisation Analyst team. The team needs to accept a good understanding of the concern rules and concern requirements, and be equipped past the tools and dashboards to detect abnormalities, outliers, cleaved trends and any other unusual scenarios that happen on Product. The objective of this team is to identify whatever data quality upshot and take information technology stock-still before users and clients do. This team also needs to partner with customer service teams and can become direct feedback from customers and accost their concerns chop-chop. With the advances of mod AI technologies, efficiency can exist potentially improved drastically. Withal, equally stated at the beginning of this commodity, quality control at the terminate is necessary but not sufficient to ensure a visitor creates and sustains adept information quality. The 6 steps stated higher up are besides required.

Summary

In determination, expert data quality requires disciplined data governance, rigorous direction of incoming data, accurate requirement gathering, thorough regression testing for alter direction and conscientious design of data pipelines, in addition to information quality command programs for the data delivered both externally and internally. For all quality problems, information technology is much easier and less costly to prevent the data issue from happening in the first place, rather than relying on defending systems and ad hoc fixes to deal with data quality problems. Finally, by following the 7 steps in this article, proficient data quality can not only be guaranteed and but also sustained.

Which Of The Following Are Genuine Data-mining Procedures? (Please Check All That Apply),

Source: https://towardsdatascience.com/7-steps-to-ensure-and-sustain-data-quality-3c0040591366

Posted by: wagnersubbillson.blogspot.com

0 Response to "Which Of The Following Are Genuine Data-mining Procedures? (Please Check All That Apply)"

Post a Comment

Iklan Atas Artikel

Iklan Tengah Artikel 1

Iklan Tengah Artikel 2

Iklan Bawah Artikel