Author: Haiyan Yu
Publisher: Springer Nature
ISBN: 3031718712
Category :
Languages : en
Pages : 103
Book Description
Data Quality Management in the Data Age
Author: Haiyan Yu
Publisher: Springer Nature
ISBN: 3031718712
Category :
Languages : en
Pages : 103
Book Description
Publisher: Springer Nature
ISBN: 3031718712
Category :
Languages : en
Pages : 103
Book Description
Data Quality for the Information Age
Author: Thomas C. Redman
Publisher: Artech House Publishers
ISBN:
Category : Computers
Languages : en
Pages : 344
Book Description
All aspects of data management are explored in this title, which provides detailed analyses of quality problems and their impacts, potential solutions and how they are combined to form an overall data quality program, senior management's role, and methods used to make and sustain improvements.
Publisher: Artech House Publishers
ISBN:
Category : Computers
Languages : en
Pages : 344
Book Description
All aspects of data management are explored in this title, which provides detailed analyses of quality problems and their impacts, potential solutions and how they are combined to form an overall data quality program, senior management's role, and methods used to make and sustain improvements.
The Practitioner's Guide to Data Quality Improvement
Author: David Loshin
Publisher: Elsevier
ISBN: 0080920349
Category : Computers
Languages : en
Pages : 423
Book Description
The Practitioner's Guide to Data Quality Improvement offers a comprehensive look at data quality for business and IT, encompassing people, process, and technology. It shares the fundamentals for understanding the impacts of poor data quality, and guides practitioners and managers alike in socializing, gaining sponsorship for, planning, and establishing a data quality program. It demonstrates how to institute and run a data quality program, from first thoughts and justifications to maintenance and ongoing metrics. It includes an in-depth look at the use of data quality tools, including business case templates, and tools for analysis, reporting, and strategic planning. This book is recommended for data management practitioners, including database analysts, information analysts, data administrators, data architects, enterprise architects, data warehouse engineers, and systems analysts, and their managers. - Offers a comprehensive look at data quality for business and IT, encompassing people, process, and technology. - Shows how to institute and run a data quality program, from first thoughts and justifications to maintenance and ongoing metrics. - Includes an in-depth look at the use of data quality tools, including business case templates, and tools for analysis, reporting, and strategic planning.
Publisher: Elsevier
ISBN: 0080920349
Category : Computers
Languages : en
Pages : 423
Book Description
The Practitioner's Guide to Data Quality Improvement offers a comprehensive look at data quality for business and IT, encompassing people, process, and technology. It shares the fundamentals for understanding the impacts of poor data quality, and guides practitioners and managers alike in socializing, gaining sponsorship for, planning, and establishing a data quality program. It demonstrates how to institute and run a data quality program, from first thoughts and justifications to maintenance and ongoing metrics. It includes an in-depth look at the use of data quality tools, including business case templates, and tools for analysis, reporting, and strategic planning. This book is recommended for data management practitioners, including database analysts, information analysts, data administrators, data architects, enterprise architects, data warehouse engineers, and systems analysts, and their managers. - Offers a comprehensive look at data quality for business and IT, encompassing people, process, and technology. - Shows how to institute and run a data quality program, from first thoughts and justifications to maintenance and ongoing metrics. - Includes an in-depth look at the use of data quality tools, including business case templates, and tools for analysis, reporting, and strategic planning.
Foundations of Data Quality Management
Author: Wenfei Fan
Publisher: Springer Nature
ISBN: 3031018923
Category : Computers
Languages : en
Pages : 201
Book Description
Data quality is one of the most important problems in data management. A database system typically aims to support the creation, maintenance, and use of large amount of data, focusing on the quantity of data. However, real-life data are often dirty: inconsistent, duplicated, inaccurate, incomplete, or stale. Dirty data in a database routinely generate misleading or biased analytical results and decisions, and lead to loss of revenues, credibility and customers. With this comes the need for data quality management. In contrast to traditional data management tasks, data quality management enables the detection and correction of errors in the data, syntactic or semantic, in order to improve the quality of the data and hence, add value to business processes. While data quality has been a longstanding problem for decades, the prevalent use of the Web has increased the risks, on an unprecedented scale, of creating and propagating dirty data. This monograph gives an overview of fundamental issues underlying central aspects of data quality, namely, data consistency, data deduplication, data accuracy, data currency, and information completeness. We promote a uniform logical framework for dealing with these issues, based on data quality rules. The text is organized into seven chapters, focusing on relational data. Chapter One introduces data quality issues. A conditional dependency theory is developed in Chapter Two, for capturing data inconsistencies. It is followed by practical techniques in Chapter 2b for discovering conditional dependencies, and for detecting inconsistencies and repairing data based on conditional dependencies. Matching dependencies are introduced in Chapter Three, as matching rules for data deduplication. A theory of relative information completeness is studied in Chapter Four, revising the classical Closed World Assumption and the Open World Assumption, to characterize incomplete information in the real world. A data currency model is presented in Chapter Five, to identify the current values of entities in a database and to answer queries with the current values, in the absence of reliable timestamps. Finally, interactions between these data quality issues are explored in Chapter Six. Important theoretical results and practical algorithms are covered, but formal proofs are omitted. The bibliographical notes contain pointers to papers in which the results were presented and proven, as well as references to materials for further reading. This text is intended for a seminar course at the graduate level. It is also to serve as a useful resource for researchers and practitioners who are interested in the study of data quality. The fundamental research on data quality draws on several areas, including mathematical logic, computational complexity and database theory. It has raised as many questions as it has answered, and is a rich source of questions and vitality. Table of Contents: Data Quality: An Overview / Conditional Dependencies / Cleaning Data with Conditional Dependencies / Data Deduplication / Information Completeness / Data Currency / Interactions between Data Quality Issues
Publisher: Springer Nature
ISBN: 3031018923
Category : Computers
Languages : en
Pages : 201
Book Description
Data quality is one of the most important problems in data management. A database system typically aims to support the creation, maintenance, and use of large amount of data, focusing on the quantity of data. However, real-life data are often dirty: inconsistent, duplicated, inaccurate, incomplete, or stale. Dirty data in a database routinely generate misleading or biased analytical results and decisions, and lead to loss of revenues, credibility and customers. With this comes the need for data quality management. In contrast to traditional data management tasks, data quality management enables the detection and correction of errors in the data, syntactic or semantic, in order to improve the quality of the data and hence, add value to business processes. While data quality has been a longstanding problem for decades, the prevalent use of the Web has increased the risks, on an unprecedented scale, of creating and propagating dirty data. This monograph gives an overview of fundamental issues underlying central aspects of data quality, namely, data consistency, data deduplication, data accuracy, data currency, and information completeness. We promote a uniform logical framework for dealing with these issues, based on data quality rules. The text is organized into seven chapters, focusing on relational data. Chapter One introduces data quality issues. A conditional dependency theory is developed in Chapter Two, for capturing data inconsistencies. It is followed by practical techniques in Chapter 2b for discovering conditional dependencies, and for detecting inconsistencies and repairing data based on conditional dependencies. Matching dependencies are introduced in Chapter Three, as matching rules for data deduplication. A theory of relative information completeness is studied in Chapter Four, revising the classical Closed World Assumption and the Open World Assumption, to characterize incomplete information in the real world. A data currency model is presented in Chapter Five, to identify the current values of entities in a database and to answer queries with the current values, in the absence of reliable timestamps. Finally, interactions between these data quality issues are explored in Chapter Six. Important theoretical results and practical algorithms are covered, but formal proofs are omitted. The bibliographical notes contain pointers to papers in which the results were presented and proven, as well as references to materials for further reading. This text is intended for a seminar course at the graduate level. It is also to serve as a useful resource for researchers and practitioners who are interested in the study of data quality. The fundamental research on data quality draws on several areas, including mathematical logic, computational complexity and database theory. It has raised as many questions as it has answered, and is a rich source of questions and vitality. Table of Contents: Data Quality: An Overview / Conditional Dependencies / Cleaning Data with Conditional Dependencies / Data Deduplication / Information Completeness / Data Currency / Interactions between Data Quality Issues
Multi-Domain Master Data Management
Author: Mark Allen
Publisher: Morgan Kaufmann
ISBN: 0128011475
Category : Computers
Languages : en
Pages : 244
Book Description
Multi-Domain Master Data Management delivers practical guidance and specific instruction to help guide planners and practitioners through the challenges of a multi-domain master data management (MDM) implementation. Authors Mark Allen and Dalton Cervo bring their expertise to you in the only reference you need to help your organization take master data management to the next level by incorporating it across multiple domains. Written in a business friendly style with sufficient program planning guidance, this book covers a comprehensive set of topics and advanced strategies centered on the key MDM disciplines of Data Governance, Data Stewardship, Data Quality Management, Metadata Management, and Data Integration. - Provides a logical order toward planning, implementation, and ongoing management of multi-domain MDM from a program manager and data steward perspective. - Provides detailed guidance, examples and illustrations for MDM practitioners to apply these insights to their strategies, plans, and processes. - Covers advanced MDM strategy and instruction aimed at improving data quality management, lowering data maintenance costs, and reducing corporate risks by applying consistent enterprise-wide practices for the management and control of master data.
Publisher: Morgan Kaufmann
ISBN: 0128011475
Category : Computers
Languages : en
Pages : 244
Book Description
Multi-Domain Master Data Management delivers practical guidance and specific instruction to help guide planners and practitioners through the challenges of a multi-domain master data management (MDM) implementation. Authors Mark Allen and Dalton Cervo bring their expertise to you in the only reference you need to help your organization take master data management to the next level by incorporating it across multiple domains. Written in a business friendly style with sufficient program planning guidance, this book covers a comprehensive set of topics and advanced strategies centered on the key MDM disciplines of Data Governance, Data Stewardship, Data Quality Management, Metadata Management, and Data Integration. - Provides a logical order toward planning, implementation, and ongoing management of multi-domain MDM from a program manager and data steward perspective. - Provides detailed guidance, examples and illustrations for MDM practitioners to apply these insights to their strategies, plans, and processes. - Covers advanced MDM strategy and instruction aimed at improving data quality management, lowering data maintenance costs, and reducing corporate risks by applying consistent enterprise-wide practices for the management and control of master data.
Corporate Data Quality
Author: Boris Otto
Publisher:
ISBN: 9783737575928
Category :
Languages : en
Pages :
Book Description
Publisher:
ISBN: 9783737575928
Category :
Languages : en
Pages :
Book Description
Enterprise Data at Huawei
Author: Yun Ma
Publisher: Springer Nature
ISBN: 981166823X
Category : Business & Economics
Languages : en
Pages : 255
Book Description
This book systematically introduces the data governance and digital transformation at Huawei, from the perspectives of technology, process, management, and so on. Huawei is a large global enterprise engaging in multiple types of business in over 170 countries and regions. Its differentiated operation is supported by an enterprise data foundation and corresponding data governance methods. With valuable experience, methodology, standards, solutions, and case studies on data governance and digital transformation, enterprise data at Huawei is ideal for readers to learn and apply, as well as to get an idea of the digital transformation journey at Huawei. This book is organized into four parts and ten chapters. Based on the understanding of βthe cognitive world of machines,β the book proposes the prospects for the future of data governance, as well as the imaginations about AI-based governance, data sovereignty, and building a data ecosystem.
Publisher: Springer Nature
ISBN: 981166823X
Category : Business & Economics
Languages : en
Pages : 255
Book Description
This book systematically introduces the data governance and digital transformation at Huawei, from the perspectives of technology, process, management, and so on. Huawei is a large global enterprise engaging in multiple types of business in over 170 countries and regions. Its differentiated operation is supported by an enterprise data foundation and corresponding data governance methods. With valuable experience, methodology, standards, solutions, and case studies on data governance and digital transformation, enterprise data at Huawei is ideal for readers to learn and apply, as well as to get an idea of the digital transformation journey at Huawei. This book is organized into four parts and ten chapters. Based on the understanding of βthe cognitive world of machines,β the book proposes the prospects for the future of data governance, as well as the imaginations about AI-based governance, data sovereignty, and building a data ecosystem.
Data Quality Assessment
Author: Arkady Maydanchik
Publisher:
ISBN: 9780977140022
Category : Computers
Languages : en
Pages : 0
Book Description
Imagine a group of prehistoric hunters armed with stone-tipped spears. Their primitive weapons made hunting large animals, such as mammoths, dangerous work. Over time, however, a new breed of hunters developed. They would stretch the skin of a previously killed mammoth on the wall and throw their spears, while observing which spear, thrown from which angle and distance, penetrated the skin the best. The data gathered helped them make better spears and develop better hunting strategies. Quality data is the key to any advancement, whether it is from the Stone Age to the Bronze Age. Or from the Information Age to whatever Age comes next. The success of corporations and government institutions largely depends on the efficiency with which they can collect, organise, and utilise data about products, customers, competitors, and employees. Fortunately, improving your data quality does not have to be such a mammoth task. This book is a must read for anyone who needs to understand, correct, or prevent data quality issues in their organisation. Skipping theory and focusing purely on what is practical and what works, this text contains a proven approach to identifying, warehousing, and analysing data errors. Master techniques in data profiling and gathering metadata, designing data quality rules, organising rule and error catalogues, and constructing the dimensional data quality scorecard. David Wells, Director of Education of the Data Warehousing Institute, says "This is one of those books that marks a milestone in the evolution of a discipline. Arkady's insights and techniques fuel the transition of data quality management from art to science -- from crafting to engineering. From deep experience, with thoughtful structure, and with engaging style Arkady brings the discipline of data quality to practitioners."
Publisher:
ISBN: 9780977140022
Category : Computers
Languages : en
Pages : 0
Book Description
Imagine a group of prehistoric hunters armed with stone-tipped spears. Their primitive weapons made hunting large animals, such as mammoths, dangerous work. Over time, however, a new breed of hunters developed. They would stretch the skin of a previously killed mammoth on the wall and throw their spears, while observing which spear, thrown from which angle and distance, penetrated the skin the best. The data gathered helped them make better spears and develop better hunting strategies. Quality data is the key to any advancement, whether it is from the Stone Age to the Bronze Age. Or from the Information Age to whatever Age comes next. The success of corporations and government institutions largely depends on the efficiency with which they can collect, organise, and utilise data about products, customers, competitors, and employees. Fortunately, improving your data quality does not have to be such a mammoth task. This book is a must read for anyone who needs to understand, correct, or prevent data quality issues in their organisation. Skipping theory and focusing purely on what is practical and what works, this text contains a proven approach to identifying, warehousing, and analysing data errors. Master techniques in data profiling and gathering metadata, designing data quality rules, organising rule and error catalogues, and constructing the dimensional data quality scorecard. David Wells, Director of Education of the Data Warehousing Institute, says "This is one of those books that marks a milestone in the evolution of a discipline. Arkady's insights and techniques fuel the transition of data quality management from art to science -- from crafting to engineering. From deep experience, with thoughtful structure, and with engaging style Arkady brings the discipline of data quality to practitioners."
Competing with High Quality Data
Author: Rajesh Jugulum
Publisher: John Wiley & Sons
ISBN: 9781118342329
Category : Business & Economics
Languages : en
Pages : 0
Book Description
Create a competitive advantage with data quality Data is rapidly becoming the powerhouse of industry, but low-quality data can actually put a company at a disadvantage. To be used effectively, data must accurately reflect the real-world scenario it represents, and it must be in a form that is usable and accessible. Quality data involves asking the right questions, targeting the correct parameters, and having an effective internal management, organization, and access system. It must be relevant, complete, and correct, while falling in line with pervasive regulatory oversight programs. Competing with High Quality Data: Concepts, Tools and Techniques for Building a Successful Approach to Data Quality takes a holistic approach to improving data quality, from collection to usage. Author Rajesh Jugulum is globally-recognized as a major voice in the data quality arena, with high-level backgrounds in international corporate finance. In the book, Jugulum provides a roadmap to data quality innovation, covering topics such as: The four-phase approach to data quality control Methodology that produces data sets for different aspects of a business Streamlined data quality assessment and issue resolution A structured, systematic, disciplined approach to effective data gathering The book also contains real-world case studies to illustrate how companies across a broad range of sectors have employed data quality systems, whether or not they succeeded, and what lessons were learned. High-quality data increases value throughout the information supply chain, and the benefits extend to the client, employee, and shareholder. Competing with High Quality Data: Concepts, Tools and Techniques for Building a Successful Approach to Data Quality provides the information and guidance necessary to formulate and activate an effective data quality plan today.
Publisher: John Wiley & Sons
ISBN: 9781118342329
Category : Business & Economics
Languages : en
Pages : 0
Book Description
Create a competitive advantage with data quality Data is rapidly becoming the powerhouse of industry, but low-quality data can actually put a company at a disadvantage. To be used effectively, data must accurately reflect the real-world scenario it represents, and it must be in a form that is usable and accessible. Quality data involves asking the right questions, targeting the correct parameters, and having an effective internal management, organization, and access system. It must be relevant, complete, and correct, while falling in line with pervasive regulatory oversight programs. Competing with High Quality Data: Concepts, Tools and Techniques for Building a Successful Approach to Data Quality takes a holistic approach to improving data quality, from collection to usage. Author Rajesh Jugulum is globally-recognized as a major voice in the data quality arena, with high-level backgrounds in international corporate finance. In the book, Jugulum provides a roadmap to data quality innovation, covering topics such as: The four-phase approach to data quality control Methodology that produces data sets for different aspects of a business Streamlined data quality assessment and issue resolution A structured, systematic, disciplined approach to effective data gathering The book also contains real-world case studies to illustrate how companies across a broad range of sectors have employed data quality systems, whether or not they succeeded, and what lessons were learned. High-quality data increases value throughout the information supply chain, and the benefits extend to the client, employee, and shareholder. Competing with High Quality Data: Concepts, Tools and Techniques for Building a Successful Approach to Data Quality provides the information and guidance necessary to formulate and activate an effective data quality plan today.
Executing Data Quality Projects
Author: Danette McGilvray
Publisher: Academic Press
ISBN: 0128180161
Category : Computers
Languages : en
Pages : 378
Book Description
Executing Data Quality Projects, Second Edition presents a structured yet flexible approach for creating, improving, sustaining and managing the quality of data and information within any organization. Studies show that data quality problems are costing businesses billions of dollars each year, with poor data linked to waste and inefficiency, damaged credibility among customers and suppliers, and an organizational inability to make sound decisions. Help is here! This book describes a proven Ten Step approach that combines a conceptual framework for understanding information quality with techniques, tools, and instructions for practically putting the approach to work β with the end result of high-quality trusted data and information, so critical to today's data-dependent organizations. The Ten Steps approach applies to all types of data and all types of organizations β for-profit in any industry, non-profit, government, education, healthcare, science, research, and medicine. This book includes numerous templates, detailed examples, and practical advice for executing every step. At the same time, readers are advised on how to select relevant steps and apply them in different ways to best address the many situations they will face. The layout allows for quick reference with an easy-to-use format highlighting key concepts and definitions, important checkpoints, communication activities, best practices, and warnings. The experience of actual clients and users of the Ten Steps provide real examples of outputs for the steps plus highlighted, sidebar case studies called Ten Steps in Action. This book uses projects as the vehicle for data quality work and the word broadly to include: 1) focused data quality improvement projects, such as improving data used in supply chain management, 2) data quality activities in other projects such as building new applications and migrating data from legacy systems, integrating data because of mergers and acquisitions, or untangling data due to organizational breakups, and 3) ad hoc use of data quality steps, techniques, or activities in the course of daily work. The Ten Steps approach can also be used to enrich an organization's standard SDLC (whether sequential or Agile) and it complements general improvement methodologies such as six sigma or lean. No two data quality projects are the same but the flexible nature of the Ten Steps means the methodology can be applied to all. The new Second Edition highlights topics such as artificial intelligence and machine learning, Internet of Things, security and privacy, analytics, legal and regulatory requirements, data science, big data, data lakes, and cloud computing, among others, to show their dependence on data and information and why data quality is more relevant and critical now than ever before. - Includes concrete instructions, numerous templates, and practical advice for executing every step of The Ten Steps approach - Contains real examples from around the world, gleaned from the author's consulting practice and from those who implemented based on her training courses and the earlier edition of the book - Allows for quick reference with an easy-to-use format highlighting key concepts and definitions, important checkpoints, communication activities, and best practices - A companion Web site includes links to numerous data quality resources, including many of the templates featured in the text, quick summaries of key ideas from the Ten Steps methodology, and other tools and information that are available online
Publisher: Academic Press
ISBN: 0128180161
Category : Computers
Languages : en
Pages : 378
Book Description
Executing Data Quality Projects, Second Edition presents a structured yet flexible approach for creating, improving, sustaining and managing the quality of data and information within any organization. Studies show that data quality problems are costing businesses billions of dollars each year, with poor data linked to waste and inefficiency, damaged credibility among customers and suppliers, and an organizational inability to make sound decisions. Help is here! This book describes a proven Ten Step approach that combines a conceptual framework for understanding information quality with techniques, tools, and instructions for practically putting the approach to work β with the end result of high-quality trusted data and information, so critical to today's data-dependent organizations. The Ten Steps approach applies to all types of data and all types of organizations β for-profit in any industry, non-profit, government, education, healthcare, science, research, and medicine. This book includes numerous templates, detailed examples, and practical advice for executing every step. At the same time, readers are advised on how to select relevant steps and apply them in different ways to best address the many situations they will face. The layout allows for quick reference with an easy-to-use format highlighting key concepts and definitions, important checkpoints, communication activities, best practices, and warnings. The experience of actual clients and users of the Ten Steps provide real examples of outputs for the steps plus highlighted, sidebar case studies called Ten Steps in Action. This book uses projects as the vehicle for data quality work and the word broadly to include: 1) focused data quality improvement projects, such as improving data used in supply chain management, 2) data quality activities in other projects such as building new applications and migrating data from legacy systems, integrating data because of mergers and acquisitions, or untangling data due to organizational breakups, and 3) ad hoc use of data quality steps, techniques, or activities in the course of daily work. The Ten Steps approach can also be used to enrich an organization's standard SDLC (whether sequential or Agile) and it complements general improvement methodologies such as six sigma or lean. No two data quality projects are the same but the flexible nature of the Ten Steps means the methodology can be applied to all. The new Second Edition highlights topics such as artificial intelligence and machine learning, Internet of Things, security and privacy, analytics, legal and regulatory requirements, data science, big data, data lakes, and cloud computing, among others, to show their dependence on data and information and why data quality is more relevant and critical now than ever before. - Includes concrete instructions, numerous templates, and practical advice for executing every step of The Ten Steps approach - Contains real examples from around the world, gleaned from the author's consulting practice and from those who implemented based on her training courses and the earlier edition of the book - Allows for quick reference with an easy-to-use format highlighting key concepts and definitions, important checkpoints, communication activities, and best practices - A companion Web site includes links to numerous data quality resources, including many of the templates featured in the text, quick summaries of key ideas from the Ten Steps methodology, and other tools and information that are available online