data warehouse) and use it to support the decision
making process. In spite of the popularity gained by
DSSs in the last decade, a methodology for software
development has not been agreed. System
development involves (roughly speaking) three
clearly defined phases: design, implementation and
maintenance. However, in the development cycle of
traditional software systems, activities are carried
out sequentially, while in a DSS they follow a
heuristic process (Cippico, 1997). Thus,
methodologies for developing operational and DSS
systems are different. Most contributions on
requirements analysis for DSS came from consulting
companies and software vendors. On the academic
side, Winter and Strauch (2003, 2004) introduced a
demand-driven methodology for data warehousing
requirement analysis. They define four-steps where
they identify users and application type, assign
priorities, and match information requirements with
actual information supply (i.e. data in the data
sources). There are several differences with the
methodology we present here. The main one resides
in that our approach is based on data quality, which
is not considered in the mentioned paper. Moreover,
although the authors mention the problem of
matching required and supplied information, they do
not provide a way of quantifying the difference
between them. On the contrary, we give a method
for determining which is the data source that better
matches the information needs for each query
defined by the user. Paim and Castro (2003)
introduced DWARF, a methodology that, like DSS-
METRIQ, deals with functional and non-functional
requirements. They adapt requirements engineering
techniques and propose a methodology for
requirements definition for data warehouses. For
non-functional requirements, they use the Extended-
Data Warehousing NFR Framework (Paim &
Castro, 2002). Although DWARF and the extended
NFR framework are close to the rationale of DSS-
METRIQ, the main differences are: (a) we give a
more detailed and concrete set of tools for non-
functional requirements elicitation; (b) we provide a
QFD-based method for data source ranking; (c) we
give a comprehensive detail of all the processes and
documents involved. Prakash and Gosain (2003)
also emphasize the need for a requirements
engineering phase in data warehousing development,
and propose the GDI (Goal-Decision-Information)
model. The methodology is not described at a level
of detail that may allow a more in-depth analysis.
3 QUALITY CONCEPTS
Many techniques have been developed in order to
measure quality, each one of them associated to a
specific metric. In what follows, we comment on the
ones we are going to use in our proposal.
GQM (Goal Question Metric) is a framework for
metric definition (Basili, Caldiera & Rombach,
1992). It describes a top-down procedure allowing to
specify what is going to be measured, and to trace
how measuring is being performed, providing a
framework for result interpretation. The outcome of
the process is the specification of a system of
measurements that consists of a set of results and a
set of rules for the interpretation of the collected
data. The model defines three levels of analysis: (a)
conceptual (Goal), where a goal for a product,
process or resource is defined; (b) operational
(Question): at this level, a set of questions is used
for describing the way an specific goal will be
reached; (c) quantitative (Metric): the metric
associated with each question.
Quality Function Deployment (QFD) (Akao,
1997), proposed in the 60's by Yoji Akao, was first
conceived as a method for the development of new
products under the framework of Total Quality
Control. QFD aims at assuring design quality while
the product is still in its design stage. It defines an
organizational behavior based on the conception of a
multifunctional team that intends to reach consensus
on the needs of the users and what they expect from
the product. The central instrument of the
methodology is the "house of quality" matrix.
Data Quality. Efforts made in order to improve data
quality are generally focused on data accuracy,
ignoring many other attributes and important quality
dimensions. Wang et al identified four data quality
categories after evaluating 118 variables (Wang &
Strong, 1996): (1) intrinsic data quality; (2)
contextual data quality; (3) data quality for data
representation; (4) accessible data quality. There is
a substantial amount of academic research on the
multiple dimensions applicable to quality of
information. For the sake of space we do not
comment on them in this work. The interested reader
should take a look to the work of Hoxmeier
(Hoxmeier, 2000), Jarke et al (Jarke & Vassiliou,
1997), and many other ones.
4 DSS-METRIQ OVERVIEW
We now introduce DSS-METRIQ, a methodology
specifically devised for requirements elicitation for
DSSs. The methodology consists of five phases:
scenario, information gathering, requirements
integration, data source selection, and document
generation. The rationale of the methodology is
REQUIREMENTS ELICITATION FOR DECISION SUPPORT SYSTEMS: A DATA QUALITY APPROACH
317