Discovering Dependencies among Data Quality Dimensions: A Validation of Instrument (original) (raw)
Related papers
A Framework to Construct Data Quality Dimensions Relationships
Data and information obtained from data analysis is an essential asset to construct and support information systems. As data is a significant resource, the quality of data is critical to enhance data quality and increase the effectiveness of business processes. Relationships among all four major data quality dimensions for process improvement are often neglected. For this reason, this study proposes to construct a reliable framework to support process activities in information systems. This study focuses on four critical quality dimensions; accuracy, completeness, consistency, and timeliness. A qualitative approach was conducted using a questionnaire and the responses were assessed to measure reliability and validity of the survey. Factor analysis and Cronbach-alpha test were applied to interpret the results. The results show that the items of each data quality dimension and improvement process are reliable and valid. This framework can be used to evaluate data quality in an information system to improve the involved process. Accuracy AQ1: This information is correct. (0.939) AQ2: This information is incorrect. (R) (0.872) AQ3: This information is accurate. (0.914) AQ4: This information is reliable. (0.797) Completeness ComQ1: This information includes all necessary values. (0.888) ComQ2: This information is incomplete. (R) (0.858) ComQ3: This information is complete. (0.844)
An analysis of data quality dimensions
2015
Data quality (DQ) has been studied in significant depth over the last two decades and has received attention from both the academic and the practitioner community. Over that period of time a large number of data quality dimensions have been identified in due course of research and practice. While it is important to embrace the diversity of views of data quality, it is equally important for the data quality research and practitioner community to be united in the consistent interpretation of this foundational concept. In this paper, we provide a step towards this consistent interpretation. Through a systematic review of research and practitioner literature, we identify previously published data quality dimensions and embark on the analysis and consolidation of the overlapping and inconsistent definitions. We stipulate that the shared understanding facilitated by this consolidation is a necessary prelude to generic and declarative forms of requirements modeling for data quality.
A REVIEW OF DATA QUALITY RESEARCH IN ACHIEVING HIGH DATA QUALITY WITHIN ORGANIZATION
Journal of Theoretical and Applied Information Technology , 2017
The aim of this review is to highlight issues in data quality research and to discuss potential research opportunity to achieve high data quality within an organization. The review adopted systematic literature review method based on research articles published in journals and conference proceedings. We developed a review strategy based on specific themes such as current research area in data quality, critical dimensions in data quality, data quality management model and methodologies and data quality assessment methods. Based on the review strategy, we select relevant research articles, extract and synthesis the information to answer our research questions. The review highlights the advancement of data quality research to resemble its real world application and discuss the available gap for future research. Research area such as organizations management, data quality impact towards the organization and database related technical solutions for data quality dominated the early years of data quality research. However, since the Internet is now taking place as the new information source, the emerging of new research areas such as data quality assessment for web and big data is inevitable. This review also identifies and discusses critical data quality dimensions in organization such as data completeness, consistency, accuracy and timeliness. We also compare and highlight gaps in data quality management model and methodologies. Existing model and methodologies capabilities are restricted to the structured data type and limit its ability to assess data quality in web and big data. Finally, we uncover available methods in data quality assessment and highlight its limitation for future research. This review is important to highlight and analyse limitation of existing data quality research related to the recent needs in data quality such as unstructured data type and big data.
Data Quality:A Survey of Data Quality Dimensions
Nowadays, activities and decisions making in an organization is based on data and information obtained from data analysis, which provides various services for constructing reliable and accurate process. As data are significant resources in all organizations the quality of data is critical for managers and operating processes to identify related performance issues. Moreover, high quality data can increase opportunity for achieving top services in an organization. However, identifying various aspects of data quality from definition, dimensions, types, strategies, techniques are essential to equip methods and processes for improving data. This paper focuses on systematic review of data quality dimensions in order to use at proposed framework which combining data mining and statistical techniques to measure dependencies among dimensions and illustrate how extracting knowledge can increase process quality.
Data quality assessment and improvement
International Journal of Business Information Systems, 2016
Data quality has significance to companies, but is an issue that can be challenging to approach and operationalise. This study focuses on data quality from the perspective of operationalisation by analysing the practices of a company that is a world leader in its business. A model is proposed for managing data quality to enable evaluation and operationalisation. The results indicate that data quality is best ensured when organisation specific aspects are taken into account. The model acknowledges the needs of different data domains, particularly those that have master data characteristics. The proposed model can provide a starting point for operationalising data quality assessment and improvement. The consequent appreciation of data quality improves data maintenance processes, IT solutions, data quality and relevant expertise, all of which form the basis for handling the origins of products.
Data Quality Dimensions, Metrics, and Improvement Techniques
Future Computing and Informatics Journal, 2021
Achieving high level of data quality is considered one of the most important assets for any small, medium and large size organizations. Data quality is the main hype for both practitioners and researchers who deal with traditional or big data. The level of data quality is measured through several quality dimensions. High percentage of the current studies focus on assessing and applying data quality on traditional data. As we are in the era of big data, the attention should be paid to the tremendous volume of generated and processed data in which 80% of all the generated data is unstructured. However, the initiatives for creating big data quality evaluation models are still under development. This paper investigates the data quality dimensions that are mostly used in both traditional and big data to figure out the metrics and techniques that are used to measure and handle each dimension. A complete definition for each traditional and big data quality dimension, metrics and handling t...
2008
The value of management decisions, the security of our nation, and the very foundations of our business integrity are all dependent on the quality of data and information. However, the quality of the data and information is dependent on how that data or information will be used. This paper proposes a theory of data quality based on the five principles defined by J. M. Juran for product and service quality and extends Wang et al's 1995 framework for data quality research. It then examines the data and information quality literature from journals within the context of this framework.
A framework for analysis of data quality research
IEEE Transactions on Knowledge and Data Engineering, 1995
Abstiuct-Organizational databases are pervaded with data of poor quality. However, there has not been an analysis of the data quality literature that provides an overall understanding of the state-of-art research in this area. Using an analogy between product manufacturing and data manufacturing, this paper develops a framework for analyzing data quality research, and uses it as the basis for organizing the data quality literature. This framework consists of seven elements: management responsibilities, operation and assurance costs, research and development, production, distribution, personnel management, and legal function. The analysis reveals that most research efforts focus on operation and assurance costs, research and development, and production of data products. Unexplored research topics and unresolved issues are identified and directions for future research provided.
INFORMATION QUALITY: DEFINITIONS, MEASUREMENT, DIMENSIONS, AND RELATIONSHIP WITH DECISION MAKING
European Centre for Research Training and Development, 2018
Quality data is inevitably an important prerequisite for managerial decision-making, especially when the decisions made can have far-reaching consequences for the organization. Hence, scrutinizing the information obtained and demanding that the information meets certain features are paramount to achieving sustainable organizational performance. The present paper provides a roadmap on the definition of Quality Data, its dimensions and relationship with decision making effectiveness.
MMPRO: A Methodology Based on ISO/IEC 15939 to Draw Up Data Quality Measurement Processes
Iq, 2008
Nowadays, data plays a key role in organizations, and management of its quality is becoming an essential activity. As part of such required management, organizations need to draw up processes for measuring the data quality (DQ) levels of their organizational units, taking into account the particularities of different scenarios, available resources, and characteristics of the data used in them. Given that there are not many works in the literature related to this objective, this paper proposes a methodology-abbreviated MMPROto develop processes for measuring DQ. MMPRO is based on ISO/IEC 15939. Despite being a standard of quality software, we believe it can be successfully applied in this context because of the similarities between software and data. The proposed methodology consists of four activities: (1) Establish and sustain the DQ measurement commitment, (2) Plan the DQ Measurement Process, (3) Perform the DQ Measurement Process, and (4) Evaluate the DQ Measurement Process. These four activities are divided into tasks. For each task, input and output products are listed, as well as a set of useful techniques and tools, many of them borrowed from the Software Engineering field.