21
Enterprise Modelling and Information Systems Architectures Vol. 8, No. 2, December 2013 4 Martin Ofner, Boris Otto, Hubert Österle Martin Ofner, Boris Otto, Hubert Österle A Maturity Model for Enterprise Data Quality Management Enterprises need high-quality data in order to meet a number of strategic business requirements. Permanent maintenance and sustainable improvement of data quality can be achieved by an enterprise-wide approach only. The paper presents a Maturity Model for Enterprise Data Quality Management (Enterprise DQM), which aims at supporting enterprises in their effort to deliberately design and establish organisation-wide data quality management. The model design process, which covered a period of five years, included several iterations of multiple design and evaluation cycles and intensive collaboration with practitioners. The Maturity Model is a hierarchical model comprising, on its most detailed level, 30 practices and 56 measures that can be used as concrete assessment elements during an appraisal. Besides being used for determining the level of maturity of Enterprise DQM in organisations, the results of the paper contribute to the ongoing discussion in the information systems (IS) community about maturity model design in general. 1 Introduction Data quality management (DQM) as an organisa- tional function comprises all practices, methods, and systems for analyzing, improving and main- taining the quality of data. DQM basically aims at maximizing the value of data (customer data, sup- plier data, or material data, for example) (DAMA 2008). Over the last 15 years DQM has been the subject of analysis in many publications both by researchers (Batini and Scannapieco 2006; Otto et al. 2007; Wang 1998; Wang et al. 1998) and practitioners (English 1999; Loshin 2001; Redman 2000). Although data quality is widely recog- nized as a strategic success factor, the majority of companies consider DQM in their organisa- tion as ‘being in the early phases of maturity’ (Pierce et al. 2008). Particularly certain business requirements, such as effective supply chain man- agement (Kagermann et al. 2010; Tellkamp et al. 2004; Vermeer 2000), improved decision-making (Price and Shanks 2005; Shankaranarayan et al. 2003), compliance with legal or regulatory provi- sions (Friedman 2006; Salchegger and Dewor 2008), or efficient customer relationship management (Reid and Catterall 2005; Zahay and Griffin 2003) demand an enterprise-wide approach to DQM, as such requirements cannot be met by isolated solutions or single business units alone. In order to be able to establish enterprise-wide DQM in the following referred to as Enterprise DQM , changes are needed on a strategic, on an organisational, and on an information systems level (Baskarada et al. 2006; Bitterer 2007; Lee et al. 2002; Ryu et al. 2006). In their effort to bring about these changes companies need support and assistance, particularly with regard to monitoring the progress in establishing Enterprise DQM. Taking this into account, the research question examined in this paper is how companies may deliberately design Enterprise DQM. The word deliberately refers to the need that companies are capable of identifying areas for improvement and deriving appropriate action with regard to Enterprise DQM. The research objective is to design a model that allows assessing the maturity of Enterprise DQM, with the research process following the principles of design science research (Hevner et al. 2004; Österle and Otto 2010). Maturity models support organisational change in- sofar as they represent an instrument for decision-

A Maturity Model for Enterprise Data Quality Management

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013

4 Martin Ofner, Boris Otto, Hubert Österle

Martin Ofner, Boris Otto, Hubert Österle

A Maturity Model for Enterprise Data QualityManagement

Enterprises need high-quality data in order to meet a number of strategic business requirements. Permanentmaintenance and sustainable improvement of data quality can be achieved by an enterprise-wide approachonly. The paper presents a Maturity Model for Enterprise Data Quality Management (Enterprise DQM),which aims at supporting enterprises in their effort to deliberately design and establish organisation-widedata quality management. The model design process, which covered a period of five years, included severaliterations of multiple design and evaluation cycles and intensive collaboration with practitioners. The MaturityModel is a hierarchical model comprising, on its most detailed level, 30 practices and 56 measures that can beused as concrete assessment elements during an appraisal. Besides being used for determining the level ofmaturity of Enterprise DQM in organisations, the results of the paper contribute to the ongoing discussion inthe information systems (IS) community about maturity model design in general.

1 Introduction

Data quality management (DQM) as an organisa-tional function comprises all practices, methods,and systems for analyzing, improving and main-taining the quality of data. DQM basically aims atmaximizing the value of data (customer data, sup-plier data, or material data, for example) (DAMA2008). Over the last 15 years DQM has been thesubject of analysis in many publications both byresearchers (Batini and Scannapieco 2006; Ottoet al. 2007; Wang 1998; Wang et al. 1998) andpractitioners (English 1999; Loshin 2001; Redman2000). Although data quality is widely recog-nized as a strategic success factor, the majorityof companies consider DQM in their organisa-tion as ‘being in the early phases of maturity’(Pierce et al. 2008). Particularly certain businessrequirements, such as effective supply chain man-agement (Kagermann et al. 2010; Tellkamp et al.2004; Vermeer 2000), improved decision-making(Price and Shanks 2005; Shankaranarayan et al.2003), compliance with legal or regulatory provi-sions (Friedman 2006; Salchegger and Dewor 2008),or efficient customer relationship management(Reid and Catterall 2005; Zahay and Griffin 2003)

demand an enterprise-wide approach to DQM,as such requirements cannot be met by isolatedsolutions or single business units alone.

In order to be able to establish enterprise-wideDQM in the following referred to as EnterpriseDQM , changes are needed on a strategic, on anorganisational, and on an information systemslevel (Baskarada et al. 2006; Bitterer 2007; Leeet al. 2002; Ryu et al. 2006). In their effort to bringabout these changes companies need support andassistance, particularly with regard to monitoringthe progress in establishing Enterprise DQM.

Taking this into account, the research questionexamined in this paper is how companies maydeliberately design Enterprise DQM. The worddeliberately refers to the need that companiesare capable of identifying areas for improvementand deriving appropriate action with regard toEnterprise DQM. The research objective is todesign a model that allows assessing the maturityof Enterprise DQM, with the research processfollowing the principles of design science research(Hevner et al. 2004; Österle and Otto 2010).

Maturity models support organisational change in-sofar as they represent an instrument for decision-

Page 2: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013A Maturity Model for Enterprise Data Quality Management 5

makers to assess an organisation‘s actual state,derive actions for improvement, and evaluatethese actions afterwards in terms of their effect-iveness and efficiency (Crosby 1979; Gibson andNolan 1974; Nolan 1973).

The following section of the paper outlines thetheoretical foundations underlying the researchand compares existing maturity models from theDQM domain. After that the research method-ology and the process of designing the MaturityModel for Enterprise DQM are elaborated. Thenthe design rationale of the structural specificationof the Maturity Model (i.e. the conceptual model)is discussed, alongside with procedural guidelinesfor applying this conceptual model. Afterwards,a first evaluation of the Maturity Model is pro-vided, and findings and implications are discussed.The paper concludes with a short summary andrecommendations for further research on thetopic.

2 Theoretical Foundations

2.1 Data and Data Quality

Singular pieces of data specify discrete charac-teristics of objects and processes from the realworld. In this sense, data is free of context (Boisotand Canals 2004; Davenport and Prusak 1998;Spiegler 2000). Business distinguishes betweenmaster data and transaction data. Master dataconsists of attributes describing a company’s corebusiness objects. It constitutes the basis for bothoperative value creation processes and analyticaldecision-making processes (Smith and McKeen2008). Typical classes of master data are suppliermaster data, customer master data, or productmaster data (Mertens 2000). Transaction datadescribes business processes. It relates to masterdata, and therefore its existence is dependenton this master data (Dreibelbis et al. 2008). It ismaster data that is of particular importance toEnterprise DQM, as the quality of such data iscritical for meeting the business requirementsmentioned above. Thus, master data needs to bedefined for the whole of an organisation and mustallow to be identified unambiguously.

When data is used within a certain context orwhen data is processed, it turns into information(Boisot and Canals, 2004; van den Hoven, 2003).Although the terms data and information areclearly distinguished in theory, a clear definitionon what quality means to either aspect does notexist. Both information quality and data qualityis seen as a context dependent, multi-dimensionalconcept, describing the ‘fitness for use’ of inform-ation and data as determined by a user or usergroup (Wang 1998). The fact that informationquality and data quality is considered to be con-text dependent emphasizes the notion that it is upto the user to decide whether certain informationor data is useful (Wang and Strong 1996). Hence,‘fitness for use’ can be perceived in different ways,manifesting itself in so-called data quality dimen-sions. Numerous scientific studies have dealtwith the identification and description of suchdata quality dimensions (Price and Shanks 2005;Wand and Wang 1996; Wang and Strong 1996;Wang et al. 1995). Among the most importantones are accessibility, accuracy, completeness, andconsistency (DAMA 2008).

2.2 Data Quality ManagementData Management Association (DAMA) definesDQM as ‘application of Total Quality Manage-ment (TQM) concepts and practices to improvedata and information quality, including settingdata quality policies and guidelines, data qualitymeasurement (including data quality auditing andcertification), data quality analysis, data cleansingand correction, data quality process improvement,and data quality education’ (DAMA 2008). DQMaims to achieve the following goals: establishDQM as an organisational function, design DQMto cover the organisation as a whole, establish acontinuous improvement process for DQM, qual-ify and authorize staff for executing DQM tasks,provide appropriate techniques and guidelines forDQM (Batini and Scannapieco 2006; English 1999;Wang 1998; Zhang 2000). In order to emphasizethe imperative to establish DQM in an enterprise-wide approach, the paper at hand refers to DQMas Enterprise DQM.

Page 3: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013

6 Martin Ofner, Boris Otto, Hubert Österle

2.3 Maturity Models and OrganisationalChange

Maturity models represent a special class of mod-els, dealing exclusively with organisational andinformation systems related change and develop-ment processes (Becker et al. 2010; Crosby 1979;Gibson and Nolan 1974; Mettler 2010; Nolan 1973).Maturity models consist of an organized set ofconstructs serving to describe certain aspects ofmaturity of a design domain (Fraser et al. 2002).The concept of maturity is often understood ac-cording to the definition of Paulk et al. (1993),who consider maturity to be the ‘extent to which aprocess is explicitly defined, managed, measured,controlled, and effective’. Most maturity modelsexplicitly or implicitly follow this definition, tak-ing a process oriented view when looking at howa design domain can be assessed and optimized.The sole focus on the process perspective hasbeen controversially discussed in literature (Bach1994; Gillies and Howard 2003; Jones 1995; Pfefferand Sutton 1999). What is demanded by critics ofthis approach is an all-encompassing, integratedconcept for measuring levels of maturity, takinginto account technological and cultural aspects aswell (Christensen and Overdorf 2000; Saleh andAlshawi 2005).

Typically, a maturity model consists of a domainmodel and an assessment model. The domainmodel comprises criteria by which the designdomain can be partitioned into discrete unitsto be assessed. The assessment model providesone or multiple assessment dimensions, each ofwhich defining an assessment scale. What is ba-sically assessed is to which extent certain criteriacomply with the scale for each assessment di-mension. In order to structure the assessmentprocess some maturity models also provide ap-praisal methods (e.g. Standard CMMI AppraisalMethod for Process Improvement, SCAMPI) (SEI2006b). Basically, two types of maturity modelscan be distinguished. Staged models build onbest practices to explicitly specify an ideal pathof development of a design domain (Paulk et al.1993). Continuous models are used to review

certain quality features of a design domain atregular intervals, determine the level of maturityfor different features or criteria, and derive ac-tions for improvement. In the case of continuousmodels the path of development is dynamic, i.e. itis not predefined by the model (EFQM 2009).

3 Related Work

3.1 DQM Approaches

In recent years a number of methods have beendeveloped both by the research and the practi-tioners’ community supposed to offer supportand assistance in selecting, adapting and applyingtechniques for improving data quality (Batiniet al. 2009). These methods describe best prac-tices for the DQM domain and can be used toderive criteria for designing a Maturity Model forEnterprise DQM.

The Complete Data Quality Methodology (CDQM)sees DQM as being composed of a series of singu-lar projects for data quality improvement (Batiniand Scannapieco 2006). These projects are resultsoriented, i.e. the data quality to be achieved isput in relation to the costs that are likely to occurin the process. Only those projects are realizedwhich promise to be reasonable and profitablefrom a business perspective.

Redman (2000) developed the Data Quality System(DQS), focusing on the provision of an organisa-tional framework (strategy, training concepts, etc.)and the development of business and technicalcapabilities (data quality planning, data qualitymeasurement, data models, etc.).

Total Data Quality Management (TDQM) is thename of a research program at the MIT. TDQMsees information as a product (known as the in-formation product (IP) approach) that needs tobe produced according to the same principlesphysical goods are produced, including exactspecification of requirements to be met by inform-ation products, control of the production processalong the entire lifecycle of information products,and naming of an information product manager(Wang 1998; Wang and Strong 1996).

Page 4: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013A Maturity Model for Enterprise Data Quality Management 7

Total Quality data Management (TQdM) is amethod that offers support when informationneeds to be optimized for business purposes (Eng-lish 1999). TQdM follows the principles of the IPapproach and focuses even more on the defini-tion of requirements to be met by informationproducts.

To sum up, it can be said that all of these meth-ods refer to results oriented, cultural, processrelated, or technological aspects of data qualitymanagement.

3.2 Maturity Models for DQM

Beside the methods described in the previoussection also maturity models for DQM have beendeveloped. Lee et al. (2002) have proposed amethodology for information quality assessment(AIMQ), which can be used as a basis for informa-tion quality assessment and benchmarking. Thismethodology uses 65 criteria to evaluate resultsto be achieved by DQM.

DataFlux (2007) has come up with a maturitymodel comprising four criteria (people, policies,technology, and risk & reward) by which compan-ies can assess the progress of DQM establishmentin their organisation.

Bitterer (2007) aims at the same objective withtheir maturity model, using quite vague defini-tions of individual levels of maturity instead ofclearly defined criteria.

Ryu et al. (2006) and Baskarada et al. (Baskaradaet al. 2006) have developed maturity models on thebasis of the Capability Maturity Model Integration(CMMI) approach (SEI 2006a). The scope of bothmodels is quite narrow with regard to DQM.While the former defines 16 criteria for specifyingand maintaining metadata (which is seen as aprerequisite for achieving high quality of data),the latter focuses on information systems for themechanical engineering industry, for which itdefines 19 technical criteria.

As Tab. 1 shows, none of the maturity modelsexamined covers all aspects of Enterprise DQM.

Guidelines for designing actions for improvementare offered by two approaches only. Also, allmaturity models examined are characterized bya rigid, predefined path of development. This,however, stands in contrast with the view ofDAMA (2009) that states ‘[. . . ] how each enter-prise implements [DQM] varies widely. Eachorganisation must determine an implementationapproach consistent with its size, goals, resources,and complexity. However, the essential principlesof [DQM] remain the same across the spectrumof enterprises [. . . ]’. Taking this into account, aMaturity Model for Enterprise DQM must pro-vide a dynamic path of development, which eachorganisation may adapt to its individual needsand requirements.

4 Research Approach

4.1 Research Method

The work presented in this paper is an outcomeof design oriented research, following the meth-odological paradigm of Design Science Research(DSR). DSR aims at designing artefacts (constructs,models, methods, or instantiations, for example)in order to solve problems occurring in practice(Hevner et al. 2004; March and Smith 1995). Theartefact to be constructed is a maturity model thatallows to deliberately design Enterprise DQM.

When developing a reliable maturity model acritical factor is the level of maturity of the designdomain itself. The less developed a design domainis, the higher is the uncertainty in terms of havingvalid and reliable knowledge about this designdomain, and the higher is the need for a maturitymodel that is capable of guiding the path of devel-opment for designing the domain. If this is thecase, usually only few cases are available that helpidentify possible criteria and evaluate the model,resulting in maturity models of limited reliabil-ity only. So access to practitioners’ knowledgeis critical for being able to define and evaluaterelevant criteria. Therefore, the overarching re-search method selected for designing a MaturityModel for Enterprise DQM is consortium research,

Page 5: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013

8 Martin Ofner, Boris Otto, Hubert Österle

Enterprise Modelling and Information Systems Architectures

Vol. X, No. X, Month 200X

4 Martin Ofner, Boris Otto, Hubert Österle

Table 1: Existing DQM maturity models in comparison.

As Table 1 shows, none of the maturity models

examined covers all aspects of Enterprise DQM.

Guidelines for designing actions for improvement are

offered by two approaches only. Also, all maturity

models examined are characterized by a rigid,

predefined path of development. This, however,

stands in contrast with the view of DAMA (2009)

that states „[.. ] how each enterprise implements

[DQM] varies widely. Each organisation must

determine an implementation approach consistent

with its size, goals, resources, and complexity.

However, the essential principles of [DQM] remain

the same across the spectrum of enterprises […]”.

Taking this into account, a Maturity Model for

Enterprise DQM must provide a dynamic path of

development, which each organisation may adapt to

its individual needs and requirements.

4 Research approach

4.1 Research method

The work presented in this paper is an outcome of

design oriented research, following the

methodological paradigm of Design Science

Research (DSR). DSR aims at designing artefacts

(constructs, models, methods, or instantiations, for

example) in order to solve problems occurring in

practice (Hevner et al., 2004; March and Smith,

1995). The artefact to be constructed is a maturity

model that allows to deliberately design Enterprise

DQM.

When developing a reliable maturity model a critical

factor is the level of maturity of the design domain

itself. The less developed a design domain is, the

higher is the uncertainty in terms of having valid and

reliable knowledge about this design domain, and

the higher is the need for a maturity model that is

capable of guiding the path of development for

designing the domain. If this is the case, usually

only few cases are available that help identify

possible criteria and evaluate the model, resulting in

maturity models of limited reliability only. So access

to practitioners‟ knowledge is critical for being able

to define and evaluate relevant criteria. Therefore,

the overarching research method selected for

designing a Maturity Model for Enterprise DQM is

consortium research, which represents a

collaborative form of DSR and which is based on

having access to and using practitioners‟ knowledge

(Österle and Otto, 2010). Figure 1 gives an overview

of the research approach, which follows idealized

design research processes (Peffers et al., 2008;

Verschuren and Hartog, 2005). The research process

consists of four activities: analysis, design,

evaluation, and diffusion. The research context is

provided by the Competence Center Corporate Data

Quality (CC CDQ) a consortium research project

consisting of 13 user companies, the Institute of

Information Management of the University of St.

Gallen and the European Foundation for Quality

Management (EFQM).

Furthermore the research methods draws upon

Action Design Research (ADR) as proposed by Sein

et al. (2011). ADR addresses the interaction with

practitioners and the organisational context the

design artefact is supposed to be used for. In

particular, the maturity model design shares the

perception of design and evaluation being an

integrated stage within a design science research

project rather than separated, sequential phases.

The integration of building activities, (organisational)

intervention activities, and evaluation activities (BIE

according to ADR) is depicted in Figure 2 by the

bidirectional arrows connecting analysis, design,

evaluation, and diffusion activities.

Source Results

oriented

criteria

Culture

related

criteria

Process

related

criteria

Techno-

logy-

related

criteria

Guide-

lines

offered

Path of

development

(Lee et al., 2002) 4 0 0 0 No Staged

(DataFlux, 2007) 0 4 4 4 Yes Staged

(Bitterer, 2007) 0 2 2 2 Yes Staged

(Ryu et al., 2006) 0 0 4 0 No Staged

(Baskarada et al., 2006) 0 0 0 4 No Staged

Key: 4 = Criteria formally defined – 2 = Criteria informally defined (embedded in textual descriptions) – 0 = No criteria

Table 1: Existing DQM Maturity Models in Comparison

which represents a collaborative form of DSRand which is based on having access to and usingpractitioners’ knowledge (Österle and Otto 2010).

Fig. 1 gives an overview of the research approach,which follows idealized design research processes(Peffers et al. 2008; Verschuren and Hartog 2005).The research process consists of four activities:analysis, design, evaluation, and diffusion. Theresearch context is provided by the CompetenceCenter Corporate Data Quality (CC CDQ) a con-sortium research project consisting of 13 usercompanies, the Institute of Information Manage-ment of the University of St. Gallen and theEuropean Foundation for Quality Management(EFQM).

Furthermore the research methods draws uponAction Design Research (ADR) as proposed bySein et al. (2011). ADR addresses the interactionwith practitioners and the organisational contextthe design artefact is supposed to be used for.In particular, the maturity model design sharesthe perception of design and evaluation beingan integrated stage within a design science re-search project rather than separated, sequentialphases. The integration of building activities,(organisational) intervention activities, and evalu-ation activities (BIE according to ADR) is depictedin Fig. 2 by the bidirectional arrows connect-ing analysis, design, evaluation, and diffusionactivities.

4.2 Research Process4.2.1 Analysis

Analysis activities began in November 2006, com-prising the identification of the problem and thespecification of requirements to be met by thesolution to be developed. EFQM joined the consor-tium as a strategic partner during this first activityof the research process, after the decision wasmade to use the well-established EFQM Modelfor Excellence as a basis for developing the Ma-turity Model for Enterprise DQM. EFQM is anon-profit aiming at establishing quality orientedmanagement systems in Europe. Among otherthings, EFQM organizes the annual EuropeanQuality Award (EQA), in the course of whichcompanies are assessed by means of the criteriaof the EFQM Model. Relevance of the research tobe undertaken was confirmed by representativesfrom the user companies of the consortium ina focus group interview and a series of expertinterviews as well as by a literature analysis (cf.Related Work). The central outcome of the Ana-lysis activity was a set of functional requirementsto be met by the Maturity Model as specified byboth the user companies of the consortium andEFQM.

4.2.2 Design

The Maturity Model was built in the course ofthree integrated design/evaluate iterations (Fig. 2).

Page 6: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013A Maturity Model for Enterprise Data Quality Management 9

Enterprise Modelling and Information Systems Architectures

Vol. X, No. X, Month 200X

A Maturity Model for Enterprise Data Quality Management 5

Figure 1: Research process.

4.2 Research process

4.2.1 Analysis

Analysis activities began in November 2006,

comprising the identification of the problem and the

specification of requirements to be met by the

solution to be developed. EFQM joined the

consortium as a strategic partner during this first

activity of the research process, after the decision

was made to use the well-established EFQM Model

for Excellence as a basis for developing the Maturity

Model for Enterprise DQM. EFQM is a non-profit

aiming at establishing quality oriented management

systems in Europe. Among other things, EFQM

organizes the annual European Quality Award (EQA),

in the course of which companies are assessed by

means of the criteria of the EFQM Model. Relevance

of the research to be undertaken was confirmed by

representatives from the user companies of the

consortium in a focus group interview and a series of

expert interviews as well as by a literature analysis

(cf. Related work). The central outcome of the

Analysis activity was a set of functional

requirements to be met by the Maturity Model as

specified by both the user companies of the

consortium and EFQM.

4.2.2 Design

The Maturity Model was built in the course of three

integrated design/evaluate iterations (Figure 2). All

three iterations included building activities,

organisational intervention activities (mainly through

action research projects), and evaluation activities.

The concrete model design process was guided by

procedure models for the development of maturity

models (de Bruin et al., 2005; Becker et al., 2010).

Adaptation mechanisms of reference modelling (vom

Brocke, 2007) allowed systematic design of the

Maturity Model on the basis of the EFQM Model,

following the Guidelines of Modeling (GoM) (Schuette

and Rotthowe, 1998). Knowledge about „things that

worked‟ and „things that did not work‟ was used to

draw up a catalog of criteria. This knowledge was

gained from related work and from a number of case

studies conducted in the context of CC CDQ.

4.2.3 Evaluation

Following the BIE principle of ADR, the evaluation of

the Maturity Model was inseparably interwoven with

the design of the Model. Evaluation within the three

design/evaluate iterations was done by focus groups

comprising different stakeholders (organized within

consortium workshops) and in the course of ten

action research projects (cf. Evaluation). Both ex-

ante and ex-post evaluation measures were applied,

i.e. the artefact design theoretical contribution

(interior mode) and its practical use (external mode)

Domain

Design

Evaluation

Diffusion

CC CDQ and EFQM

agreement

State of DQM and maturity

models

Scientificpublications

Managerialpublications

Training material

Focus groupinterviews

Action researchprojects

Survey

Web-based assessment

tool

Maturity modeldesign

Case studies

Orderly referencemodeling

Problem definition

by CC CDQ

Scientificknowledge

• Maturity Modelling

Theory

• Maturity Model

Design

• Capability View on the Firm

Analysis

Practical

knowledge

• Maturity models

• DQM practices and

indicators

Requirements of all

stakeholder

groups

Conferences & seminars

Figure 1: Research Process

All three iterations included building activities,organisational intervention activities (mainlythrough action research projects), and evaluationactivities. The concrete model design process wasguided by procedure models for the developmentof maturity models (Becker et al. 2010; Bruin et al.2005). Adaptation mechanisms of reference mod-elling (Brocke 2007) allowed systematic designof the Maturity Model on the basis of the EFQMModel, following the Guidelines of Modeling(GoM) (Schuette and Rotthowe 1998). Knowledgeabout things that worked and things that did notwork was used to draw up a catalog of criteria.This knowledge was gained from related workand from a number of case studies conducted inthe context of CC CDQ.

4.2.3 Evaluation

Following the BIE principle of ADR, the eval-uation of the Maturity Model was inseparablyinterwoven with the design of the Model. Evalu-ation within the three design/evaluate iterationswas done by focus groups comprising different

stakeholders (organized within consortium work-shops) and in the course of ten action researchprojects (cf. Evaluation). Both ex-ante and ex-postevaluation measures were applied, i.e. the artefactdesign theoretical contribution (interior mode)and its practical use (external mode) were studied(Sonnenberg and Vom Brocke 2012). Evaluationactivities concluded with a survey on the criteriaand maturity levels of the Maturity Model. Aquestionnaire was sent to 128 subject matter ex-perts from the DQM domain, who were selectedfrom the address database of the Institute forInformation Management of the University of St.Gallen. 32 of these experts responded, confirm-ing the criteria previously identified. Twenty ofthem declared to be willing to actively supportthe Maturity Model with their names and thenames of their organisations by means of a jointpublication with EFQM (2011). 49 subject matterexperts from 24 user companies, four consultingcompanies, and EFQM joined to evaluate theModel. Basically, the focus groups and the surveyserved to optimize and verify the componentsand elements of the Maturity Model (in terms

Page 7: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013

10 Martin Ofner, Boris Otto, Hubert Österle

Enterprise Modelling and Information Systems Architectures

Vol. X, No. X, Month 200X

6 Martin Ofner, Boris Otto, Hubert Österle

Figure 2: Design/Evaluate Iterations and Design Decisions.

were studied (Sonnenberg and Vom Brocke, 2012).

Evaluation activities concluded with a survey on the

criteria and maturity levels of the Maturity Model. A

questionnaire was sent to 128 subject matter

experts from the DQM domain, who were selected

from the address database of the Institute for

Information Management of the University of St.

Gallen. 32 of these experts responded, confirming

the criteria previously identified. Twenty of them

declared to be willing to actively support the

Maturity Model with their names and the names of

their organisations by means of a joint publication

with EFQM (EFQM, 2011). 49 subject matter experts

from 24 user companies, four consulting companies,

and EFQM joined to evaluate the Model. Basically,

the focus groups and the survey served to optimize

and verify the components and elements of the

Maturity Model (in terms of optimized wording,

above all), whereas the action research projects

aimed at demonstrating the Model‟s applicability and

benefit (relating to the ability to derive improvement

actions).

4.2.4 Difussion

The Diffusion phase began in the middle of 2008.

The results of the research were disseminated via

various channels. Scientific publications on the topic

deal with the gap in research to be closed,

requirements to be met by a maturity model for

DQM, possible areas of application of such a model,

and the literature research that was conducted

(Ofner et al., 2009; Hüner et al., 2009).The present

paper documents the entire research process, the

design objectives, design decisions, and the process

of evaluating the artefact. Apart from being

documented in writing, the Maturity Model was

presented at various conferences and seminars and

was discussed with participants, among them the

ACM SAC in 2008, the American Conference of

Information Systems (AMCIS) in 2009, the German

Information Quality Management Conference

(GIQMC) in 2010, and the Stammdaten-

Management Forum in 2009 and 2010. Besides, both

the Maturity Model and the appraisal method have

been implemented as a web based assessment tool,

which was made publicly accessible in April 2011

and which allows organisations to conduct self-

assessments regarding Enterprise DQM (cf.

https://benchmarking.iwi.unisg.ch). The assessment

tool also serves as a platform for diffusion of the

Model. Model design

4.3 Scope and requirements

The Maturity Model for Enterprise DQM aims at

enabling companies to deliberately design Enterprise

DQM in their organisation. Requirements to be met

by the artefact were identified by the

representatives from the user companies of the

consortium and by EFQM (cf. Table 2).

2006 2007 2008 2009 2010 2011

Need articulated in

consortium workshop

MM evaluated in

AR project

Requirements specified in

consortium workshop

MM evaluated in

in AR projects

MM evaluated

by EFQM

Cooperation

agreed with EFQM

MM assessed in

consortium workshop

Web-based

assessment tool ready

MM available for

public

DE Iteration 1

DE Iteration 2

DE Iteration 3

DD1

MM evaluated

in AR projects

MM assessed in

consortium workshop

DD2

DD3

DD4

Legend: MM – Maturity Model; DE – Design/Evaluate; DD – Design Decision.

MM evaluated

through survey

Figure 2: Design/Evaluate Iterations and Design Decisions

of optimized wording, above all), whereas theaction research projects aimed at demonstratingthe Model’s applicability and benefit (relating tothe ability to derive improvement actions).

4.2.4 Diffusion

The Diffusion phase began in the middle of 2008.The results of the research were disseminatedvia various channels. Scientific publications onthe topic deal with the gap in research to beclosed, requirements to be met by a maturitymodel for DQM, possible areas of application ofsuch a model, and the literature research thatwas conducted (Hüner et al. 2009; Ofner et al.2009).The present paper documents the entireresearch process, the design objectives, designdecisions, and the process of evaluating the arte-fact. Apart from being documented in writing,the Maturity Model was presented at variousconferences and seminars and was discussed withparticipants, among them the ACM SAC in 2008,the American Conference of Information Systems(AMCIS) in 2009, the German Information QualityManagement Conference (GIQMC) in 2010, andthe Stammdaten-Management Forum in 2009 and

2010. Besides, both the Maturity Model and theappraisal method have been implemented as aweb based assessment tool, which was made pub-licly accessible in April 2011 and which allowsorganisations to conduct self-assessments regard-ing Enterprise DQM. The assessment tool alsoserves as a platform for diffusion of the Model.

5 Model Design

5.1 Scope and Requirements

The Maturity Model for Enterprise DQM aims atenabling companies to deliberately design Enter-prise DQM in their organisation. Requirementsto be met by the artefact were identified by therepresentatives from the user companies of theconsortium and by EFQM (cf. Tab. 2).

5.2 Conceptual Model and DesignDecisions

Fig. 3 illustrates the conceptual elements of theMaturity Model. Model elements adopted fromthe EFQM Excellence Model are indicated with theEFQM namespace prefix. Tab. 3 lists the designdecisions made during different design/evaluate

Page 8: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013A Maturity Model for Enterprise Data Quality Management 11

Enterprise Modelling and Information Systems Architectures

Vol. X, No. X, Month 200X

A Maturity Model for Enterprise Data Quality Management 7

Table 2: Functional requirements to be met by the Model.

4.4 Conceptual model and design decisions

Figure 3 illustrates the conceptual elements of the

Maturity Model. Model elements adopted from the

EFQM Excellence Model are indicated with the EFQM

namespace prefix. Table 3 lists the design decisions

made during different design/evaluate iterations,

leading to more model elements being added

(highlighted with gray background color in Figure 3).

In the following sections the design decisions are

explained in more detail. In order to illustrate every

design decision, each explanation includes a vignette

(Stake, 1995) giving a concrete Enterprise DQM

related example from one of the user companies

taking part in the action research projects or in the

focus groups.

4.4.1 Design decision 1: Use EFQM Excellence

Model as a base model

The first design decision referred to the Maturity

Model for Enterprise DQM to be developed on the

basis of the EFQM Model for Excellence (EFQM,

2009). The EFQM Model is an assessment model that

can be used to identify a dynamic path of

development. What have been adopted in particular

is the overall structure of the EFQM Model and the

content of its assessment model, whereas the

domain model of the Maturity Model to be developed

needs to be „filled‟ with Enterprise DQM specific

content. Adoption of the EFQM Model‟s generic

structure ensures compatibility of the Maturity Model

with existing EFQM methods and techniques for

assessment and analysis. The assessment

dimensions developed by EFQM and its partners

have been used and continuously reviewed for over

twenty years. The content of the domain model of

the Maturity Model is explicated in the following

paragraphs. The Maturity Model is built upon the

logic that an organisation that defines goals for

Enterprise DQM requires certain capabilities in order

to be able to achieve these goals (cf. Figure 3). At

its core, the Maturity Model defines 30 Practices and

56 Measures for Enterprise DQM that can be used as

concrete assessment elements during an appraisal.

Whereas Practices are used to assess if and how well

certain Enterprise DQM capabilities are established in

an organisation already, Measures allow assessing if

and how well the Practices support the achievement

of Enterprise DQM goals.

No. Requirement

R1 Improvement guidelines: The Maturity Model provides practices how to reach the next, higher level of

maturity of Enterprise DQM. It is supposed to be used as a management tool enabling companies to deliberately design Enterprise DQM in their organisation.

R2 Objectivity: The Maturity Model uses a hierarchical model to partition the design domain of Enterprise DQM into smaller entities which can be assessed independently of each other. Also, fuzziness of assessments can be reduced by subdividing the design domain into smaller entities (de Bruin et al., 2005). However, it has to be noted that a maturity model always contains a certain degree of fuzziness.

R3 Dynamic path of development: The Maturity Model is non-prescriptive and allows to identify a dynamic path of development regarding Enterprise DQM. It is important that each company needs to find its own path of development. A maturity model cannot and should not set a predefined path of development to be followed by any company.

R4 Multiple dimensions: The Maturity Model provides multiple dimensions to assess the level of maturity of Enterprise DQM, as progress in organisational change cannot be captured by a single dimension (as progress may refer to the way DQM has been implemented, to business units affected by DQM, etc.).

R5 Assessment methodology: The Maturity Model provides a comprehensive assessment methodology (i.e. a process model, techniques, and tools) for being able to make reliable assessments and to avoid „finger in the wind‟ assessments. The assessment methodology is supposed to allow both self-assessments and assessments by external experts.

R6 Flexibility: The Maturity Model provides configuration mechanisms to reflect specific requirements. It must be applicable for any company, regardless of size or industry. Explicit configuration mechanisms must consistently specify how the Maturity Model may be adapted to company specific requirements.

R7 Conformity with EFQM standard: The Maturity Model complies with EFQM standards and is based on the EFQM Model for Excellence in order to be adopted into the EFQM model family and to be recognized as an official EFQM standard (EFQM, 2003b). Conformity with EFQM standards also ensures connectability with other methods, techniques, and tools.

Table 2: Functional Requirements to be met by the Model

iterations, leading to more model elements beingadded (highlighted with gray background colorin Fig. 3). In the following sections the designdecisions are explained in more detail. In order toillustrate every design decision, each explanationincludes a vignette (Stake 1995) giving a concreteEnterprise DQM related example from one of theuser companies taking part in the action researchprojects or in the focus groups.

5.2.1 Design decision 1: Use EFQMExcellence Model as a base model

The first design decision referred to the MaturityModel for Enterprise DQM to be developed on thebasis of the EFQM Model for Excellence (EFQM2009). The EFQM Model is an assessment modelthat can be used to identify a dynamic path ofdevelopment. What has been adopted in particu-lar is the overall structure of the EFQM Modeland the content of its assessment model, whereas

the domain model of the Maturity Model to bedeveloped needs to be filled with Enterprise DQMspecific content. Adoption of the EFQM Model’sgeneric structure ensures compatibility of theMaturity Model with existing EFQM methodsand techniques for assessment and analysis. Theassessment dimensions developed by EFQM andits partners have been used and continuously re-viewed for over twenty years. The content of thedomain model of the Maturity Model is explicatedin the following paragraphs. The Maturity Modelis built upon the logic that an organisation thatdefines goals for Enterprise DQM requires certaincapabilities in order to be able to achieve thesegoals (cf. Fig. 3). At its core, the Maturity Modeldefines 30 Practices and 56 Measures for EnterpriseDQM that can be used as concrete assessment ele-ments during an appraisal. Whereas Practices areused to assess if and how well certain EnterpriseDQM capabilities are established in an organ-

Page 9: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013

12 Martin Ofner, Boris Otto, Hubert Österle

Enterprise Modelling and Information Systems Architectures

Vol. X, No. X, Month 200X

8 Martin Ofner, Boris Otto, Hubert Österle

Figure 3: Conceptual model of the Maturity Model.

and future needs to manage enterprise data” is a

Practice related to the Enabler criterion 3c (which

itself is part of the Enabler criterion 3), or “Success

rate of enterprise data quality related training and

development of individuals” is a Measure related to

Result criterion “8b. Performance of people results”

(which itself is part of the Result criterion 8) [for a

complete list of Practices and Measures see EFQM

(2011)]. Enabler criteria describe which areas need

to be dealt with in order to establish Enterprise

DQM. “Strategy” addresses leaders to recognize the

importance of high-quality enterprise data as a

prerequisite for being able to respond to business

drivers (compliance with regulatory and legal

directives, integrated customer management,

strategic reporting, or business process integration

and standardization, for example). Leaders are

required to promote a culture of preventive

Enterprise DQM. “Controlling” is about the

quantitative assessment of the quality of enterprise

data. Moreover, the interrelations between

enterprise data quality and business process

performance are identified and monitored.

“Organisation & People” ensures that clearly defined

roles, which are specified by clearly defined tasks

Vignette 1. Use EFQM Excellence Model as a

base model

A German supplier from the auto industry wants to

establish central Enterprise DQM as part of a

program for company-wide process harmonization.

Certain tasks and activities related to Enterprise

DQM are already being done by regional business

units. The company now wants to conduct a

systematic analysis in order to find out who is doing

what already and what needs to be improved. Both

the analysis and the continuous improvement

process is to be assigned to the company‟s quality

management department, which is already using

EFQM methods and models.

Another company (from the chemical industry),

which established Enterprise DQM as a central

management function some years ago, is planning

to integrate DQM oriented objectives into the goal

structure of certain executive employees. A reliable,

standardized methodology is necessary for

determining the achievement of objectives to be

broadly accepted by the employees affected.

Organization

Capability Goal

EFQM::

Practice

EFQM::Maturity

level

Company-

specific practice

Methods and

models

EFQM::

Measure

Company-

specific

measure

Assessment

context

EFQM::

Score level

EFQM::

Assessment

criterion

EFQM::

Assessment

dimension

Context valueContext

category

1..n

1..1

determines

1..n 0..n

enables

achievement

of

0..n

1..1posseses

1..1

0..n

has

has

1..n

1..1

institutionalized in

1..n

1..1

measured by

1..n 1..1has

1..n

1..1

consists

1..n 1..1has

1..1 1..1

assigned

to

1..n

1..1

assessed by

Design result

EFQM domain model

EFQM assessment model

1..1

1..n

possible

outcome of

1..1

1..n

guide creation of

Company-

specific context

category

EFQM::

Enabler criterion

EFQM::

Result criterion

1..1

1..n

grouped

by

1..n

1..1

grouped

by

1..n

1..1

grouped

by

1..1

1..n

grouped

by

Figure 3: Conceptual Model of the Maturity ModelEnterprise Modelling and Information Systems Architectures

Vol. X, No. X, Month 200X

A Maturity Model for Enterprise Data Quality Management 9

Table 3: Overview of Design Decisions (DD).

and decision-making rights, are assigned to

competent people. Appropriate assignment of

Enterprise DQM responsibilities allows to efficiently

and effectively perform DQM related projects and

activities. “Processes and Methods” ensures –

through the use of Enterprise DQM related processes

and services – that expectations are fully satisfied

and that increased value for customers and other

stakeholders is generated. “Data Architecture” refers

to planning and managing the enterprise data

architecture in order to be able to ensure enterprise

data quality in terms of enterprise data storage and

distribution. “Applications” for Enterprise DQM are

supposed to provide functionality that supports DQM

tasks.

Results criteria account for the fact that the way the

Practices are realized has an effect on the people of

a company, its customers (including internal

customers, like e.g. business units or project

teams), the society, and a company‟s overall

business performance, respectively. EFQM provides

an appraisal method for the assessment process,

consisting of a procedure model and techniques for

assessment and analysis (EFQM, 2003a). The

appraisal method uses a series of interviews and

focus groups as well as document analysis for

determining the level of maturity. The most

comprehensive technique offered is “Results,

Approaches, Deploy, Assess and Refine” (RADAR),

which defines seven Assessment dimensions for

Practices and for Measures, respectively (EFQM,

2009, pp. 22-25). The level of maturity is always

determined according to the same principles,

regardless of the assessment technique used. For

each Practice and each Measure a score is

determined for each Assessment dimension using an

Assessment scale. The total result is hierarchically

calculated according to predefined calculation

schemes (EFQM, 2009, p. 27) and then entered on a

1000-point scale and assigned to one of the three

Maturity levels defined by the EFQM (cf. Figure 4).

4.4.2 Design decision 2: Integrate assessment

context

As a second design decision it was agreed that the

idea of an Assessment context needed to be

integrated into the model design, as every single

maturity assessment relates to a certain context

(e.g. management of customer and supplier master

data in regions North America, Europe, and Asia)

that should be predefined prior to the assessment.

What context is specified has an effect on the

selection of experts to be interviewed.

No. Design decision Model elements

DD1 Use EFQM Excellence Model as a base model EFQM::Enabler criterion, EFQM::Result criterion,

EFQM::Practice, EFQM::Measure, EFQM:Assessment dimension, EFQM::Assessment scale, EFQM:Score level, EFQM::Maturity level

DD2 Integrate assessment context Assessment context, Context category, Category value

DD3 Strengthen common understanding of practices Design result, Methods and models

DD4 Allow company specific configuration Company-specific context category, Company-specific practice, Company-specific measure

Vignette 2. Integrate assessment context

A global provider of telecommunications services

aims at establishing Enterprise DQM in order to be

able to meet the need for high-quality master data

for the new business environment. The company

management decided to conduct a maturity

assessment to determine the current level of

maturity of its Enterprise DQM. To do so, 66 persons

from six oranisational functions (finance, IT, sales,

etc.) in five countries were selected for being

interviewed. After a number of interviews had been

conducted the project group wondered about one

interviewee considering data maintenance processes

to be fully optimized and documented, while another

interviewee said these processes were badly

structured and incomplete. The reason for this

discrepancy was that one interviewee referred to

supplier master data for North America, whereas

another interviewee talked about customer master

data for the European market. This was taken as an

indication that experts always relate their individual

assessment to a certain context.

Table 3: Overview of Design Decisions (DD)

Page 10: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013A Maturity Model for Enterprise Data Quality Management 13

isation already, Measures allow assessing if andhow well the Practices support the achievementof Enterprise DQM goals.

Vignette 1. Use EFQM Excellence Model as abase model

A German supplier from the automotive industrywants to establish central Enterprise DQM aspart of a program for company-wide processharmonization. Certain tasks and activities re-lated to Enterprise DQM are already being doneby regional business units. The company nowwants to conduct a systematic analysis in or-der to find out who is doing what already andwhat needs to be improved. Both the analysisand the continuous improvement process is tobe assigned to the company’s quality manage-ment department, which is already using EFQMmethods and models.

Another company (from the chemical industry),which established Enterprise DQM as a centralmanagement function some years ago, is plan-ning to integrate DQM oriented objectives intothe goal structure of certain executive employ-ees. A reliable, standardized methodology isnecessary for determining the achievement ofobjectives to be broadly accepted by the employ-ees affected.

For reasons of clarity, both Measures and Prac-tices are hierarchically grouped on two levelsof detail (as shown in Tab. 4) whereas Measuresare arranged by Result criteria and Practices byEnabler criteria. To give examples, ‘Running anadequate Enterprise DQM training program todevelop people’s knowledge and competenciesregarding their current and future needs to man-age enterprise data’ is a Practice related to theEnabler criterion 3c (which itself is part of theEnabler criterion 3), or ‘Success rate of enterprisedata quality related training and development ofindividuals’ is a Measure related to Result criterion‘8b. Performance of people results’ (which itself ispart of the Result criterion 8) [for a complete list ofPractices and Measures see EFQM (2011)]). Enabler

criteria describe which areas need to be dealt within order to establish Enterprise DQM. ‘Strategy’addresses leaders to recognize the importanceof high-quality enterprise data as a prerequis-ite for being able to respond to business drivers(compliance with regulatory and legal directives,integrated customer management, strategic re-porting, or business process integration and stand-ardization, for example). Leaders are required topromote a culture of preventive Enterprise DQM.‘Controlling’ is about the quantitative assessmentof the quality of enterprise data. Moreover, theinterrelations between enterprise data quality andbusiness process performance are identified andmonitored. ‘Organisation and People’ ensuresthat clearly defined roles, which are specified byclearly defined tasks and decision-making rights,are assigned to competent people. Appropriateassignment of Enterprise DQM responsibilitiesallows to efficiently and effectively perform DQMrelated projects and activities. ‘Processes andMethods’ ensures through the use of EnterpriseDQM related processes and services—that expect-ations are fully satisfied and that increased valuefor customers and other stakeholders is generated.‘Data Architecture’ refers to planning and man-aging the enterprise data architecture in orderto be able to ensure enterprise data quality interms of enterprise data storage and distribution.‘Applications’ for Enterprise DQM are supposedto provide functionality that supports DQM tasks.

Results criteria account for the fact that the waythe Practices are realized has an effect on thepeople of a company, its customers (includinginternal customers, like e.g. business units orproject teams), the society, and a company’s over-all business performance, respectively. EFQMprovides an appraisal method for the assessmentprocess, consisting of a procedure model and tech-niques for assessment and analysis (EFQM 2003).The appraisal method uses a series of interviewsand focus groups as well as document analysisfor determining the level of maturity. The mostcomprehensive technique offered is ‘Results, Ap-proaches, Deploy, Assess and Refine’ (RADAR),

Page 11: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013

14 Martin Ofner, Boris Otto, Hubert Österle

Enterprise Modelling and Information Systems Architectures

Vol. X, No. X, Month 200X

A Maturity Model for Enterprise Data Quality Management 11

Table 4: Enabler and Results criteria.

It is important that all results recorded from each

expert interview or focus group must always be

interpreted in relation to the context specified (e.g.

when an interviewee‟s assessment refers only to

customer master data related practices of Enterprise

DQM in North America).

In order to be able to consolidate the data collected

(from various expert interviews), the context each

interview refers to needs to be annotated

unambiguously. Three generic context categories

plus context values were identified for the Maturity

Criteria Subcriteria

1. Strategy 1a. A strategy for Enterprise DQM is developed, reviewed and updated based on the organisation‟s business strategy

1b. Leaders are personally involved in ensuring that an Enterprise DQM system is developed, shared, implemented, continuously improved, and integrated with the overall organisational management system

2. Controlling 2a. The business impact of data quality is identified and related enterprise data quality measures are defined and managed

2b. The quality of data is permanently monitored and acted upon

2c. Developing, implementing and improving methods of measurement for enterprise data quality metrics

3.Organisation and People

3a. People resources for managing and supporting Enterprise DQM are defined, managed, and improved

3b. People‟s awareness for Enterprise DQM is established and maintained

3c. People are empowered to assume Enterprise DQM responsibilities

4. Processes and Methods

4a. Enterprise DQM processes are systematically designed, managed, and improved

4b. The use and maintenance of enterprise data in core business processes is systematically identified, improved and actively managed

4c. Designing, improving and documenting data creation, use and maintenance (as-is and to-be) for a better understanding of the enterprise data use within the organisation

5. Data Architecture

5a. A common understanding of a data model for the business entities is developed, permanently assessed, and made available to people

5b. Data storage and distribution is systematically designed, implemented and managed

6. Applications 6a. The application landscape is planned, managed, and improved to support Enterprise DQM activities

6b. A rollout plan for closing the gap between the as-is and the to-be application landscape is managed and improved to support Enterprise DQM activities

6c. A roadmap for strategic planning of the application landscape is managed and continuously monitored and improved

7. Customer results

7a. Perception of customer results

7b. Performance of customer results

8. People results

8a. Perception of people results

8b. Performance of people results

9. Society results

9a. Perception of society results

9b. Performance of society results

10. Key results

10a. Strategic outcomes of Enterprise DQM

10b. Performance of Enterprise DQM

Table 4: Enabler and Results Criteria

Page 12: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013A Maturity Model for Enterprise Data Quality Management 15

which defines seven Assessment dimensions forPractices and for Measures, respectively (EFQM2009). The level of maturity is always determinedaccording to the same principles, regardless of theassessment technique used. For each Practice andeach Measure a score is determined for each As-sessment dimension using an Assessment scale. Thetotal result is hierarchically calculated accordingto predefined calculation schemes (EFQM 2009)and then entered on a 1000-point scale and as-signed to one of the three Maturity levels definedby the EFQM (cf. Fig. 4).

5.2.2 Design decision 2: Integrateassessment context

As a second design decision it was agreed thatthe idea of an Assessment context needed to beintegrated into the model design, as every singlematurity assessment relates to a certain context(e.g. management of customer and supplier masterdata in regions North America, Europe, and Asia)that should be predefined prior to the assessment.What context is specified has an effect on the se-lection of experts to be interviewed. If for certainreasons (e.g. limited human resources or budget)certain interview participants cannot be includedin the appraisal (e.g. experts for the Europeanand Asian regions are not available), the specifiedcontext needs to be revised. It is important thatall results recorded from each expert interviewor focus group must always be interpreted inrelation to the context specified (e.g. when aninterviewee’s assessment refers only to customermaster data related practices of Enterprise DQMin North America).

In order to be able to consolidate the data collec-ted (from various expert interviews), the contexteach interview refers to needs to be annotatedunambiguously. Three generic context categoriesplus context values were identified for the Matur-ity Model: data class, geographic affiliation, andIT system (cf. Fig. 5).

Vignette 2. Integrate assessment context

A global provider of telecommunications ser-vices aims at establishing Enterprise DQM inorder to be able to meet the need for high-qualitymaster data for the new business environment.The company management decided to conduct amaturity assessment to determine the currentlevel of maturity of its Enterprise DQM. To doso, 66 persons from six organisational functions(finance, IT, sales, etc.) in five countries were se-lected for being interviewed.But one intervieweereferred to supplier master data for North Amer-ica, whereas another interviewee talked aboutcustomer master data for the European market.This was taken as an indication that expertsalways relate their individual assessment to acertain context.

5.2.3 Design decision 3: Strengthencommon understanding ofpractices

The third design decision relates to each Practicebeing assigned with a set of appropriate Methodsand models (plus Design results) allowing to ex-ecute each Practice in a structured way. SpecifyingDesign results (strategy documents, measurementsystems, etc.) beforehand helps to reduce sub-jectivity of assessments, as interviewees are givenhints as to what type of formal results (documents,templates, reports, systems, etc.) can be expectedto result from each Practice. Fig. 5 illustrates theassessment of a Practice and demonstrates howthe additional information given about possibleDesign results strengthens a common understand-ing. Also, these sets of Methods and models can beused for planning actions for improvement (for acomplete list, see (EFQM 2011)).

Vignette 3. Strengthen common understand-ing of Practices

A leading company from the glass industry isconducting a maturity assessment of its currentEnterprise DQM strategy, organisation, and ar-chitecture, in order to develop an action plan for

Page 13: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013

16 Martin Ofner, Boris Otto, Hubert Österle

Enterprise Modelling and Information Systems Architectures

Vol. X, No. X, Month 200X

A Maturity Model for Enterprise Data Quality Management 11

In order to be able to consolidate the data collected

(from various expert interviews), the context each

interview refers to needs to be annotated

unambiguously. Three generic context categories

plus context values were identified for the Maturity

Model: data class, geographic affiliation, and IT

system (cf. Figure 5).

Figure 4: Score assessment and maturity calculation.

4.4.3 Design decision 3: Strengthen common

understanding of practices

The third design decision relates to each Practice

being assigned with a set of appropriate Methods

and models (plus Design results) allowing to execute

each Practice in a structured way. Specifying Design

results (strategy documents, measurement systems,

etc.) beforehand helps reduce subjectivity of

assessments, as interviewees are given hints as to

what type of formal results (documents, templates,

reports, systems, etc.) can be expected to result

from each Practice. Figure 5 illustrates the

assessment of a Practice and demonstrates how the

additional information given about possible Design

results strengthens a common understanding. Also,

these sets of Methods and models can be used for

planning actions for improvement (for a complete

list, cf. (EFQM, 2011)).

4.4.4 Design decision 4: Allow company specific

configuration

The fourth design decision refers to the Maturity

Model to provide configuration mechanisms, as the

Model is supposed to be applicable to practically any

organisation, regardless of size, industry, or

individual situation regarding Enterprise DQM.

Furthermore, providing configuration mechanisms

emphasizes the idea that each organisation should

be given the opportunity to find its own path of

development with regard to designing Enterprise

DQM. Configuration mechanisms provided by the

Model refer to selection and deselection of elements,

variation with regard to naming of elements, and

Vignette 3. Strengthen common understanding

of Practices

A leading company from the glass industry is

conducting a maturity assessment of its current

Enterprise DQM strategy, organisation, and

architecture, in order to develop an action plan for

improvement. 26 persons from three production

sites in three different countries were selected for

being interviewed by a group of assessors. As there

was poor common understanding of each Practice

among the assessors, the first assessments

conducted were not comparable or summable.

0

100

200

300

400

500

600

700

800

900

1000

t1 t2 t3 t4

Ove

rall

sco

re

Time

EDQM maturity

Key results

Customer results

People results

Society results

Applications

Data Architecture

Processes & Methods

Organisation & People

Controlling

Strategy

Establishingawareness

Creating structures

Becoming effectiveIntegrity

Segmentation

Performance

TrendsTargets

Comparisons

Causes

Measure score

Soundness

Integration

Implementation

SystemMeasurement

Learning andcreativity

Improvement andinnovation

Practice score

Figure 4: Score Assessment and Maturity Calculation

Enterprise Modelling and Information Systems Architectures

Vol. X, No. X, Month 200X

A Maturity Model for Enterprise Data Quality Management 13

company wide level prevents effective Enterprise

DQM. As a consequence, the project team decided to

establish the Practice “Formalize, review and update

scope, strategy, objectives, and processes of

Enterprise DQM that meets stakeholders‟ needs and

expectations and is aligned with the business

strategy” together with the Design result “Strategy

document” following the Methods and Models of the

PROMET-BSD methodology (IMG, 1998). The

“Strategy document” defines the scope, the value

contribution, the mandate and a roadmap for

Enterprise DQM, and is supposed to be verified,

accepted and approved by the leaders of the

company.

Figure 5: Example of a practice assessment form.

6 Evaluation

Generally, evaluating design artefacts must take into

account the dual nature of Design Science Research

aiming at both advancing the scientific knowledge

base and providing results useful in practice.

Sonnenberg and vom Brocke (2012) have identified

four different evaluation types by distinguishing

between ex-ante evaluation in the course of artefact

design activities and ex-post evaluation during

artefact usage activities. Evaluation type 1 is

concerned with problem identification, whereas type

2 mainly addresses the design objectives and the

design approach. Evaluation type 3 can be

understood as a proof of the artefact‟s applicability,

and type 4, finally, as a proof of its usefulness.

Evaluation type 1 was mainly addressed by focus

groups and expert interviews during the first

design/evaluation iteration of the project. The need

for a maturity model was articulated in late 2006,

and specific requirements were revisited in mid

2008. Table 5 lists the results of the evaluation of

the Maturity Model.

P8 Developing, implementing and improving methods of measurement for enterprise data

quality metrics

Assessment context

Data class Supplier data Costumer data Product data

Geographic EMEA NAM APAC

IT System G1 ERP HQ

Possible design result Models and Methods

Measurement

system

Measurement system to assess data quality and

data quality measures by means of metrics.

Generally speaking, metrics provide consolidated

information on complicated phenomena from the real

world on the basis of quantitative measuring. Metric systems are supposed to increase the

meaningfulness of individual metrics by structuring

them and defining relationships between them.

• Method for specifying

business oriented data

quality metrics (Hüner,

2011)

• Methods and models for performance management

(IMG, 1999)

Assessment

Approach 0% 25% 50% 75% 100%

Sound

Integrated

TOTAL for Approach 37, 5%

Deployment 0% 25% 50% 75% 100%

Implemented

Systematic

TOTAL for Deployment 12,5%

Deployment 0% 25% 50% 75% 100%

Measurement

Learning and Creativity

Improvement and Innovation

TOTAL for Deployment 0%

OVERALL TOTAL 25%

Figure 5: Example of a Practice Assessment Form

Page 14: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013A Maturity Model for Enterprise Data Quality Management 17

improvement. 26 persons from three productionsites in three different countries were selectedfor being interviewed by a group of assessors.As there was poor common understanding ofeach Practice among the assessors, the first as-sessments conducted were not comparable orsummable.

5.2.4 Design decision 4: Allow companyspecific configuration

The fourth design decision refers to the MaturityModel to provide configuration mechanisms, asthe Model is supposed to be applicable to practic-ally any organisation, regardless of size, industry,or individual situation regarding Enterprise DQM.Furthermore, providing configuration mechan-isms emphasizes the idea that each organisationshould be given the opportunity to find its ownpath of development with regard to designingEnterprise DQM. Configuration mechanismsprovided by the Model refer to selection anddeselection of elements, variation with regard tonaming of elements, and definition of new ele-ments. Element selection and deselection allowsto limit the scope of an assessment by maskingcertain Practices or Measures. Especially if theModel is used for the first time, it is recommen-ded to work with a reduced scope. Variationwith regard to naming of Practices and Measuresallows to use synonyms, as each organisationprefers its own, individual terms for denotingcertain concepts in order to increase the model’sclarity and raise acceptance on the part of theusers. Definition of new elements allow to fill inplaceholders in order to add further, individualCompany-specific practices, Company-specific mea-sures, or Company-specific context categories.

Vignette 4. Allow company specific configu-ration

A German telecommunications provider is plan-ning to assess the maturity of its EnterpriseDQM related to supplier and customer masterdata maintained by the European ERP system.

An international glass manufacturer focuses onproduct master data in all regional and globalERP systems with a special interest in prac-tices related to data migration projects (due tonegative experiences in the past). A Germanautomotive supplier is planning to improve En-terprise DQM maturity in order to reduce theamount of data related process incidents.

As these examples show, the Maturity Modelis intended to be used by companies from allkinds of industries (chemicals, pharmaceuticals,manufacturing, retail, consumer goods, etc.)and with different experiences made in the past.Each company has its individual assessmentcontext, aims at achieving DQM goals throughindividual practices, and prefers to use differentmeasures to evaluate whether goals have beenachieved. Therefore the Maturity Model needsto be configurable to meet company specificrequirements.

6 Demonstration Case

A company, which is one of the world’s leadingtelecommunications and information technologyservice companies, adapted its business strategyin order to factor in socio-economic develop-ments, such as digitalization of central areas oflife, personalization of products and services, andincreasing mobility of individuals. To validatewhether the strategy is met on a short-term basis,the company defined a number of goals, such asexpanding its leading position in the broadbandsector, entering into the entertainment market, ormeeting its customers’ expectations with regardto rendering certain products and services. Asone measurable objective referring to customersatisfaction it was agreed that customer incid-ents be reduced by 25% within a year. Businessand data management experts of the companysupposed that problems in the management ofcustomer data and product data had produceddata defects which had a negative impact on busi-ness operations, leading to a growing numberof customer incidents. Therefore, the company

Page 15: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013

18 Martin Ofner, Boris Otto, Hubert Österle

initiated a project to assess the as-is maturitylevel of Enterprise DQM, identify interrelationsbetween established practices of Enterprise DQMand the impact on the number of customer incid-ents, and derive improvement actions as deemedappropriate.

The project team, which was made up of businessand data management experts, selected 30 Prac-tices and three Measures from the Maturity Modelfor being used in the assessment. Moreover, twoCompany-Specific practices (e.g. ‘Data integra-tion guidelines are defined, communicated, andapplied in relevant projects’) and one Company-specific measure (‘Number of customer incidents’)were added to take into account the company’sspecific requirements, experiences, and goals. Theassessment context, which also defines the scopeof the assessment, was set to the Context catego-ries ‘Data class’ (‘Customer data’, ‘Product data’),‘Organisational affiliation’ (‘IT Shared Servicedepartment’), and ‘IT System’ (‘Central ERP’) andtheir respective values. Furthermore, the projectteam selected ‘RADAR’ as the assessment method-ology to be applied (EFQM 2003). Twelve businessand IT experts were selected for taking part ininterviews in order to determine the assessmentscores.

The company reached a total score of 305 (out of1000), calculated as the average of the results foreach single criterion (Strategy: 17%; Controlling:40%; Organisation and People: 27%; Processesand Methods: 42%; Data Architecture: 32%; Ap-plications: 72%; Customer Results: 25%; SocietyResults: 25%; People Results: 25%; Key Results: 0%;Overall: 30,5%). Hence, at the time of the assess-ment the company was in the transition processfrom maturity level one (‘Establishing awareness’)to maturity level two (‘Creating structures’). Boththe quantitative results as well as the findingsfrom the interviews identified strategic deficits aspotential root causes of the negative impact ofdata issues on the Key Results (and the increasingnumber of customer incidents). For example, itwas discovered that the lack of an official man-date (allocated to a company’s department) that

allows defining binding rules and guidelines on acompany-wide level prevents effective EnterpriseDQM. As a consequence, the project team decidedto establish the Practice ‘Formalize, review andupdate scope, strategy, objectives, and processesof Enterprise DQM that meets stakeholders’ needsand expectations and is aligned with the businessstrategy’ together with the Design result ‘Strategydocument’ following the Methods and Modelsof the PROMET-BSD methodology (IMG 1998).The ‘Strategy document’ defines the scope, thevalue contribution, the mandate and a roadmapfor Enterprise DQM, and is supposed to be veri-fied, accepted and approved by the leaders of thecompany.

7 Evaluation

Generally, evaluating design artefacts must takeinto account the dual nature of Design ScienceResearch aiming at both advancing the scientificknowledge base and providing results useful inpractice. Sonnenberg and Vom Brocke (2012)have identified four different evaluation typesby distinguishing between ex-ante evaluationin the course of artefact design activities andex-post evaluation during artefact usage activities.Evaluation type 1 is concerned with problemidentification, whereas type 2 mainly addressesthe design objectives and the design approach.Evaluation type 3 can be understood as a proof ofthe artefact’s applicability, and type 4, finally, as aproof of its usefulness.

Evaluation type 1 was mainly addressed by focusgroups and expert interviews during the firstdesign/evaluation iteration of the project. Theneed for a maturity model was articulated in late2006, and specific requirements were revisited inmid 2008. Tab. 5 lists the results of the evaluationof the Maturity Model.

The design decisions mentioned above were theresult of different evaluation types at differentstages of the research process. Fig. 2 shows thatDD1 (Use of the EFQM Excellence model) resultedfrom an evaluation of the design approach (type

Page 16: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013A Maturity Model for Enterprise Data Quality Management 19

Enterprise Modelling and Information Systems Architectures

Vol. X, No. X, Month 200X

14 Martin Ofner, Boris Otto, Hubert Österle

The design decisions mentioned above were the

result of different evaluation types at different

stages of the research process. Figure 2 shows that

DD1 (Use of the EFQM Excellence model) resulted

from an evaluation of the design approach (type 2)

in the course of the first design/evaluation iteration,

and that DD2, DD3, and DD4 resulted from

evaluation activities taking place in the action

research projects (type 3 and 4) in the second and

third design/evaluate iteration.

Table 5: Evaluation of the Model with regard to functional requirements.

Evaluation type 4, i.e. proof of the artefact‟s

usefulness, was analyzed in greater detail. In

particular, the question as to whether the demand

for economic efficiency of the Maturity Model is met

is difficult to answer. Depending on the scope of the

assessment context that was defined, for a project

team to apply the Maturity Model in an organisation

takes five to thirty days if it is to comprise all phases

of the appraisal method (from project preparation to

training of staff to deriving actions for improvement)

(cf. Table 6). Obviously, the effort required for

training staff is higher if the Model is used for the

first time, and gets lower after repeated use.

Applying a maturity model, in general, is a

continuous process, for which appropriate

organisational structures need to be created.

Companies already using EFQM methods and models

should be able to quickly understand the Maturity

Model and use it regularly, and the staff of

companies which have already established quality

management should require training with regard to

the principles and structures of the EFQM Model

only. If there is neither quality management in place

nor any knowledge about the EFQM Model at hand,

companies need to create adequate organisational

structures and build up certain knowledge – which

may generate substantial costs – before they can

apply the Maturity Model. From applying the Model

some of the companies taking part in the action

research projects have derived actions for

improvement (ranging from five to twenty), of which

some were actually implemented (depending on

priorities, budget, or availability of resources).

No. Evaluation result Model element(s)

R1 Improvement guidelines: The Maturity Model provides methods and models for executing each practice properly.

Methods and models

R2 Objectivity: Specifying an assessment context helps assessors and interviewees determine the score for a certain assessment element. Additional information, such as typical design results to be expected, ensures a common understanding of the criteria among all parties involved in the process.

Assessment context, Context category, Context value, Design result

R3 Dynamic path of development: The Maturity Model is based on the EFQM Model for Excellence. This model is a continuous model, which allows a dynamic path of development.

All EFQM model elements

R4 Multiple dimensions: Depending on the assessment technique used, the Maturity Model provides as many as 14 or just one single assessment dimensions. The assessment techniques and dimensions have been developed by the EFQM and its members and have been used for assessing organisations for over twenty years.

-

R5 Appraisal method: The assessment process is supported by a comprehensive assessment methodology provided by the EFQM. The methodology contains a procedure model as well as techniques for analysis, configuration, and assessment. The methodology and the Maturity Model itself have been implemented in a web based prototype.

-

R6 Flexibility: The Maturity Model provides placeholders for company specific adaptation of the Model. Techniques for configuration support the process of company specific adaptation and ensure semantic and syntactic consistency of the Model.

Company-specific context dimension, Company-specific practice, Company-specific measure

R7 Conformity with EFQM standard: The EFQM uses a standardization process („EFQM branding‟) which ensures compliance of potential EFQM models with EFQM principles. The Maturity Model has passed this process successfully. It is now the official standard of EFQM for assessing the maturity of Enterprise DQM in organisations.

-

Table 5: Evaluation of the Model with regard to Functional Requirements

2) in the course of the first design/evaluationiteration, and that DD2, DD3, and DD4 resultedfrom evaluation activities taking place in theaction research projects (type 3 and 4) in thesecond and third design/evaluate iteration.

Evaluation type 4, i.e. proof of the artefact’susefulness, was analyzed in greater detail. Inparticular, the question as to whether the demandfor economic efficiency of the Maturity Modelis met is difficult to answer. Depending on thescope of the assessment context that was defined,for a project team to apply the Maturity Model inan organisation takes five to thirty days if it is tocomprise all phases of the appraisal method (fromproject preparation to training of staff to derivingactions for improvement) (cf. Tab. 6). Obviously,the effort required for training staff is higher ifthe Model is used for the first time, and gets lower

after repeated use. Applying a maturity model,in general, is a continuous process, for whichappropriate organisational structures need to becreated. Companies already using EFQM methodsand models should be able to quickly understandthe Maturity Model and use it regularly, and thestaff of companies which have already establishedquality management should require training withregard to the principles and structures of theEFQM Model only. If there is neither qualitymanagement in place nor any knowledge aboutthe EFQM Model at hand, companies need tocreate adequate organisational structures andbuild up certain knowledge which may generatesubstantial costs before they can apply the Ma-turity Model. From applying the Model some ofthe companies taking part in the action researchprojects have derived actions for improvement

Page 17: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013

20 Martin Ofner, Boris Otto, Hubert Österle

Enterprise Modelling and Information Systems Architectures

Vol. X, No. X, Month 200X

16 Martin Ofner, Boris Otto, Hubert Österle

7 Conclusions

7.1 Contribution of the paper

The paper presents a Maturity Model for Enterprise

DQM, which aims at supporting enterprises in their

effort to deliberately design and establish

organisation wide data quality management. The

elements of the Maturity Model are based on

principles of quality management in general and

existing DQM approaches in particular. The Model‟s

structure and assessment dimensions have been

adopted from the EFQM Model for Excellence. The

Model has been approved by EFQM as the official

framework for quality oriented management of

Table 6: Action research projects.

enterprise data. It comprises, on its most detailed

level, 30 practices and 56 measures that can be

used as concrete assessment elements during an

appraisal. Although the design domain and the

purpose of the Maturity Model are specific, findings

gained during the artefact design process can be

generalized in order to derive further patterns for

designing maturity models (e.g. integrating an

assessment context). Moreover, through explication

of the design process the results can be taken up by

other researchers for verification and extension.

Furthermore, due to the explication of the design

process the model is open to be extended, adapted

and reused by future design science research

endeavours in related fields.

Companies may use the Maturity Model for

Enterprise DQM to conduct maturity assessments

and derive actions for improvement. Specifying

design results to be expected together with taking

advantage of appropriate methods and techniques

from research and practice is highly useful to

support the planning of such actions. The Model‟s

hierarchical structure allows detailed analysis of the

results of a maturity assessment and presentation of

these results to different stakeholder groups in an

organisation.

7.2 Limitations

The Maturity Model for Enterprise DQM has been

used and tested only by large companies so far.

Hence, the findings presented in the paper basically

apply to the structure and requirements of large

companies and cannot be considered to be equally

valid for small companies or single company units.

Another aspect of limitation refers to the fact that

the actions for improvement which were

implemented by the companies in the course of

action research projects could not be verified (in

terms of whether they have actually led to increased

DQM maturity). As most of these actions started

only recently and are expected to take some time

until they start to become effective, the paper does

not include any findings on this aspect.

7.3 Need for further research

Further research is expected to refer to continuous

maintenance and optimization of the Maturity Model

for Enterprise DQM. As the Model is a ”living“

Company Date of

assessment

Project duration

[days]

Model coverage Improvement

actions derived (implemented)

Beiersdorf 05/10 8 Enabler criteria Assessment only

Corning Cable Systems 02/11 25 Enabler criteria 18 (4)

Elektrizitätswerke des Kantons Zürich (EKZ)

06/10 5 Enabler criteria Assessment only

Deutsche Telekom 03/11 20 Enabler and Results criteria

20 (1)

Partner Automotive 07/08 5 Enabler criteria Assessment only

Siemens Enterprise Com. 07/10 30 Enabler criteria 15 (4)

Swisscom IT Services 11/10 10 Enabler criteria 13 (1)

Syngenta 11/11 8 Enabler criteria Assessment only

Stadtwerke München (SWM)

05/10 24 Enabler criteria 10 (2)

ZF Friedrichshafen 08/08 5 Enabler criteria 5 (1)

Table 6: Action Research Projects

(ranging from five to twenty), of which some wereactually implemented (depending on priorities,budget, or availability of resources).

8 Conclusions

8.1 Contribution of the paper

The paper presents a Maturity Model for Enter-prise DQM, which aims at supporting enterprisesin their effort to deliberately design and estab-lish organisation-wide data quality management.The elements of the Maturity Model are basedon principles of quality management in generaland existing DQM approaches in particular. TheModel’s structure and assessment dimensionshave been adopted from the EFQM Model for Ex-cellence. The Model has been approved by EFQMas the official framework for quality oriented man-agement of enterprise data. It comprises, on itsmost detailed level, 30 practices and 56 measuresthat can be used as concrete assessment elementsduring an appraisal. Although the design domainand the purpose of the Maturity Model are spe-cific, findings gained during the artefact designprocess can be generalized in order to derivefurther patterns for designing maturity models(e.g. integrating an assessment context).

Moreover, through explication of the design pro-cess the results can be taken up by other research-ers for verification and extension. Furthermore,due to the explication of the design process themodel is open to be extended, adapted and reusedby future design science research endeavours inrelated fields. Companies may use the MaturityModel for Enterprise DQM to conduct maturityassessments and derive actions for improvement.Specifying design results to be expected togetherwith taking advantage of appropriate methodsand techniques from research and practice ishighly useful to support the planning of suchactions. The Model’s hierarchical structure al-lows detailed analysis of the results of a maturityassessment and presentation of these results todifferent stakeholder groups in an organisation.

8.2 Limitations

The Maturity Model for Enterprise DQM hasbeen used and tested only by large companies sofar. Hence, the findings presented in the paperbasically apply to the structure and requirementsof large companies and cannot be considered tobe equally valid for small companies or singlecompany units. Another aspect of limitationrefers to the fact that the actions for improvementwhich were implemented by the companies in the

Page 18: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013A Maturity Model for Enterprise Data Quality Management 21

course of action research projects could not beverified (in terms of whether they have actuallyled to increased DQM maturity). As most of theseactions started only recently and are expected totake some time until they start to become effective,the paper does not include any findings on thisaspect.

8.3 Need for further researchFurther research is expected to refer to continuousmaintenance and optimization of the MaturityModel for Enterprise DQM. As the Model is a‘living’ artefact, it must be reviewed and revisedfrom time to time in order to keep meeting therequirements of different groups (i.e. the scientificcommunity and the practitioners’ community). Aweb based assessment tool is supposed to facilit-ate the collection of reference values for levelsof maturity regarding Enterprise DQM (best-in-class, industry average, etc.) in order to supportthe benchmarking process in the future. In thisrespect, a central challenge lies in finding a bal-ance between the Model’s flexibility and ensuringcomparability of results across company boundar-ies. Furthermore, future research should examinewhether the findings presented in the paper canbe transferred to other organisational domainsand to smaller companies.

References

Bach J. (1994) The Immaturity of CMM. In:American Programmer 7(9), pp. 13–18

Baskarada S., Gao J., Koronios A. (2006) AgileMaturity Model Approach to Assessing andEnhancing the Quality of Asset Information inEngineering Asset Management InformationSystems. In: Abramowicz W. (ed.) BusinessInformation Systems. Proceedings of BusinessInfomation Systems, pp. 4–17

Batini C., Scannapieco M. (2006) Data Quality.Concepts, Methodologies and Techniques.Springer, Berlin

Batini C., Cappiello C., Francalanci C., MaurinoA. (2009) Methodologies for Data QualityAssessment and Improvement. In: ACM Com-puting Surveys 41(3), 16:1–52

Becker J., Knackstedt R., Pöppelbus̈s J. (2010)Developing Maturity Models for IT Manage-ment? A Procedure Model and its Application.In: Business & Information Systems Engineer-ing 1(3), pp. 213–22

Bitterer A. (2007) Gartner’s Data Quality Matur-ity Model. Gartner Research

Boisot M., Canals A. (2004) Data, informa-tion and knowledge: have we got it right?In: Journal of Evolutionary Economics 14(1),pp. 43–67

vom Brocke J. (2007) Design Principles for Refer-ence Modeling: Reusing Information Modelsby Means of Aggregation, Specialisation, In-stantiation, and Analogy In: Reference Mod-eling for Business Systems Analysis Fettke P.,Loos P. (eds.) Idea Group Publishing, Hershey,pp. 47–75

de Bruin T., Freeze R., Kulkarni U., RosemannM. (2005) Understanding the Main Phases ofDeveloping a Maturity Assessment Model. In:16th Australasian Conference on InformationSystems. Proceedings of the 16th AustralasianConference on Information Systems, pp. 1–11

Christensen C. M., Overdorf M. (2000) Meet-ing the Challenge of Disruptive Change. In:Harvard Business Review 78(2), pp. 67–75

Crosby P. B. (1979) Quality Is Free - The Art ofMaking Quality Certain. McGraw-Hill, NewYork

DAMA (2008) The DAMA Dictionary of DataManagement. Technics Publications, Denver,Colorado

DAMA (2009) The DAMA Guide to the Data Man-agement of Knowledge, First Edition. TechnicsPublications, Bradley Beach

DataFlux (2007) The Data Governance MaturityModel. DataFlux

Davenport T. H., Prusak L. (1998) WorkingKnowledge - How Organizations ManageWhat They Know. Harvard Business SchoolPress, Boston

Dreibelbis A., Hechler E., Milman I., OberhoferM., van Run P., Wolfson D. (2008) EnterpriseMaster Data Management: An SOA Approachto Managing Core Information, 1st ed. IBM

Page 19: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013

22 Martin Ofner, Boris Otto, Hubert Österle

Press, Upper Saddle RiverEFQM (2003) The Fundamental Concepts of Ex-

cellence. EFQM PressEFQM (2009) EFQM Excellence Model 2010.

EFQM PressEFQM (2011) Framework for Corporate Data

Quality Management: Assessing the Organiz-ations Data Quality Management Capabilities.EFQM Press, Brussels

English L. P. (1999) Improving Data Warehouseand Business Information Quality. John Wiley& Sons, New York

Fraser P., Moultrie J., Gregory M. (2002) Theuse of maturity models / grids as a tool inassessing product development capability. In:IEEE International Engineering ManagementConference. Proceedings of the 2002 IEEEInternational Engineering Management Con-ference, pp. 244–249

Friedman T. (2006) Gartner Study on Data Qual-ity Shows That IT Still Bears the Burden. In:

Gibson C. F., Nolan R. L. (1974) Managing thefour stages of EDP growth. In: Harvard Busi-ness Review 52(1), pp. 76–88

Gillies A., Howard J. A. (2003) Managing Changein Process and People: Combining a MaturityModel with a Competency-based Approach.In: Total Quality Management & BusinessExcellence 14(7), pp. 779–787

Hevner A. R., March S. T., Park J., Ram S. (2004)Design Science in Information System Re-search. In: Management Information SystemsQuarterly 28(1), pp. 75–105

Hüner K., Ofner M., Otto B. (2009) Towards aMaturity Model for Corporate Data QualityManagement. In: Shin D. (ed.) ACM Sym-posium on Applied Computing. Proceedingsof the 2009 ACM Symposium on AppliedComputing, pp. 231–238

IMG (1998) PROMET BSD - Business Strategy De-velopment. Information Management Group,St. Gallen

Jones C. (1995) Gaps in SEI Programs. In: Soft-ware Development 3(3), pp. 41–48

Kagermann H., Österle H., Jordan J. M. (2010)IT-Driven Business Models. John Wiley &

Sons, HobokenLee Y. W., Strong D. M., Kahn B. K., Wang R. Y.

(2002) AIMQ: a methodology for informa-tion quality assessment. In: Information &Management 40, pp. 133–146

Loshin D. (2001) Enterprise knowledge man-agement: the data quality approach. MorganKaufmann, San Diego

March S. T., Smith G. F. (1995) Design and naturalscience research on information technology.In: Decision Support Systems 15(4), pp. 251–266

Mertens P. (2000) Integrierte Informationsver-arbeitung (in German: Integrated informationprocessing). Gabler, Wiesbaden

Mettler T. (2010) Maturity Assessment Models:A Design Science Research Approach. In: In-ternational Journal of Society Systems Science3(1), pp. 81–98

Nolan R. L. (1973) Managing the ComputerResource: A Stage Hypothesis. In: Commu-nications of the ACM 16(7), pp. 399–405

Ofner M., Hüner K., Otto B. (2009) Dealingwith Complexity: A Method to adapt andimplement a Maturity Model for CorporateData Quality Management. In: 15th AmericasConference on Information Systems. Pro-ceedings of the 15th Americas Conference onInformation Systems, p. 491

Otto B., Wende K., Schmidt A., Osl P. (2007)Towards a Framework for Corporate DataQuality Management. In: Toleman M., Cater-Steel A., Roberts D. (eds.) Proceedings of 18thAustralasian Conference on Information Sys-tems. The University of Southern Queensland,pp. 916–926

Paulk M. C., Weber C. W., Garcia S. M., ChrissisM. B., Bush M. (1993) Key Practices of theCapability Maturity Model, Version 1.1. In:

Peffers K., Tuunanen T., Rothenberger, A. M.,Chatterjee S. (2008) A Design Science Re-search Methodology for Information SystemsResearch. In: Journal of Management Inform-ation Systems 24(3), pp. 45–77

Pfeffer J., Sutton R. (1999) Knowing "What" Todo is not enough: Turning knowledge into

Page 20: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013A Maturity Model for Enterprise Data Quality Management 23

action. In: California Management Review42(1), pp. 83–108

Pierce E., Dismute W. S., Yonke C. L. (2008) TheState of Information and Data Governance -Understanding How Organizations GovernTheir Information and Data Assets. Interna-tional Association for Information and DataQuality

Price R., Shanks G. G. (2005) A semiotic inform-ation quality framework: development andcomparative analysis. In: Journal of Informa-tion Technology (20) 2005, pp. 88–102

Redman T. C. (2000) Data Quality. The FieldGuide. Digital Press, Boston

Reid A., Catterall M. (2005) Invisible data qualityissues in a CRM implementation. In: Journalof Database Marketing & Customer StrategyManagement 12(4), pp. 305–314

Ryu K.-S., Park J.-S., Park J.-H. (2006) A DataQuality Management Maturity Model. In:Electronics and Telecommunications ResearchInstitute Journal 28(2) 2006, pp. 191–204

Salchegger M., Dewor E. (2008) Höchste Zeitfür ein Solvency II Data Management. In:Versicherungswirtschaft 20, pp. 1714–1718

Saleh Y., Alshawi M. (2005) An AlternativeModel for Measuring the Success of IS Pro-jects: The GPIS Model. In: Journal of Enter-prise Information Management 18(1), pp. 47–63

Schuette R., Rotthowe T. (1998) The Guidelinesof Modeling - An Approach to Enhance theQuality in Information Models. In: Ling T. W.,Lee S. L. (eds.) 17th International Conferenceon Conceptual Modeling. Proceedings of the17th International Conference on ConceptualModeling. Springer, pp. 240–254

SEI (2006a) CMMI for Development, Version 1.2.Carnegie Mellon University, Pittsburgh

SEI (2006b) Standard CMMI Appraisal Methodfor Process Improvement, Version 1.2. Carne-gie Mellon University, Pittsburgh

Sein M. K., Henfridsson O., Purao S., Rossi M.,Lindgren R. (2011) Action Design Research.In: MIS Quarterly 31(1), pp. 37–56

Shankaranarayan G., Ziad M., Wang R. Y. (2003)

Managing Data Quality in Dynamic DecisionEnvironments: An Information Product Ap-proach. In: Journal of Database Management14(4) 2003, pp. 14–32

Smith H. A., McKeen J. D. (2008) Developmentsin Practice XXX: Master Data Management:Salvation Or Snake Oil? In: Communicationsof the Association for Information Systems23(4), pp. 63–72

Sonnenberg C., Vom Brocke J. (2012) Evaluationsin the Science of the Artificial - Reconsideringthe Build-Evaluate Pattern in Design ScienceResearch. In: Peffers K. R., A. M., Kuechler B.(eds.) 7th International Conference on DesignScience Research in Information Systems andTechnology (DESRIST 2012), Las Vegas, NV

Spiegler I. (2000) Knowledge management: anew idea or a recycled concept? In: Com-munications of the AIS 3(4) 374521, pp. 1–24

Stake R. E. (1995) The Art Of Case Study Re-search. SAGE Publications, Thousand Oaks

Österle H., Otto B. (2010) Consortium Research.A Method for Researcher-Practitioner Collab-oration in Design-Oriented IS Research. In:Business & Information Systems Engineering52(2), pp. 273–285

Tellkamp C., Angerer A., Fleisch E., Corsten D.(2004) From Pallet to Shelf: Improving DataQuality in Retail Supply Chains Using RFID.In: Cutter IT Journal 17(9), pp. 19–24

Vermeer B. H. P. J. (2000) How Important isData Quality for Evaluating the Impact of EDIon Global Supply Chains? In: 33rd AnnualHawaii International Conference on SystemSciences. Proceedings of the 33rd AnnualHawaii International Conference on SystemSciences, pp. 1–10

Verschuren P., Hartog R. (2005) Evaluation inDesign-Oriented Research. In: Quality andQuantity 39(6), pp. 733–762

Wand Y., Wang R. Y. (1996) Anchoring DataQuality Dimensions in Ontological Founda-tions. In: Communications of the ACM 39(11),pp. 86–95

Wang R. Y. (1998) A Product Perspective on

Page 21: A Maturity Model for Enterprise Data Quality Management

Enterprise Modelling and Information Systems ArchitecturesVol. 8, No. 2, December 2013

24 Martin Ofner, Boris Otto, Hubert Österle

Total Data Quality Management. In: Commu-nications of the AIS 41(2), pp. 58–65

Wang R. Y., Strong D. M. (1996) Beyond Ac-curacy: What Data Quality Means to DataConsumers. In: Journal of Management In-formation Systems 12(4), pp. 5–34

Wang R. Y., Storey V. C., Firth C. P. (1995) AFramework for Analysis of Data Quality Re-search. In: IEEE Transactions on Knowledgeand Data Engineering 7(4), pp. 623–640

Wang R. Y., Lee Y. W., Pipino L. L., Strong D. M.(1998) Manage Your Information as a Product.In: Sloan Management Review 39(4), pp. 95–105

Zahay D., Griffin A. (2003) Information ante-cedents of personalisation and customisationin business-to-business service markets. In:Journal of Database Marketing 10(3), pp. 255–271

Zhang Z. (2000) Implementation of Total QualityManagement. An Empirical Study of ChineseManufacturing Firms. Dissertation

Martin Ofner

Institute of Information ManagementUniversity of St. GallenMüller-Friedberg-Str. 8CH-9000 St. [email protected]

Prof. Dr. Boris Otto

ProfessorAudi-Endowed Chair of Supply Net OrderManagementTU Dortmund UniversityLogistikCampusJoseph-von-Fraunhofer-Str. 2-4D-44227 [email protected]

Prof. Dr. Hubert Österle

ProfessorInstitute of Information ManagementUniversity of St. GallenMüller-Friedberg-Str. 8CH-9000 St. [email protected]