GI-Edition
Gesellschaft für Informatik e.V. (GI)
publishes this series in order to make available to a broad public
recent findings in informatics (i.e. computer science and information systems), to document conferences that are organized in cooperation with GI and to publish the annual GI Award dissertation.
The volumes are published in German or English.
Information: http://www.gi.de/service/publikationen/lni/
Markus Nüttgens, Oliver Thomas,
Barbara Weber (Eds.)
M. Nüttgens, O.Thomas, B. Weber (Eds.): EMISA 2011
Broken down into
• seminars
• proceedings
• dissertations
• thematics
current topics are dealt with from the vantage point of research and
development, teaching and further training in theory and practice.
The Editorial Committee uses an intensive review process in order
to ensure high quality contributions.
Lecture Notes
in Informatics
Enterprise Modelling and
Information Systems
Architectures (EMISA 2011)
Hamburg, Germany
September 22–23, 2011
ISSN 1617-5468
ISBN 978-88579-284-0
This volume contains the proceedings of the 4th International Workshop on Enterprise
Modelling and Information Systems Architectures (EMISA 2011) which is jointly organized by the GI Special Interest Group on Modelling Business Information Systems
(GI-SIG MoBIS) and the GI Special Interest Group on Design Methods for Information
Systems (GI-SIG EMISA). It features a selection of 12 full papers and 12 research-inprogress papers from academia and practice on Information Systems, Business Informatics and Computer Science. Leading authors from academia and industry present
state-of-the-art developments, experiences, and best practices in this volume.
3018001 GI P_190 Cover.indd 1
190
Proceedings
24.08.11 15:45
Markus Nüttgens, Oliver Thomas, Barbara Weber (Eds.)
Enterprise Modelling and
Information Systems Architectures
(EMISA 2011)
Hamburg, Germany
September 22-23, 2011
Gesellschaft für Informatik e.V. (GI)
Lecture Notes in Informatics (LNI) - Proceedings
Series of the Gesellschaft für Informatik (GI)
Volume P-190
ISBN 978-3-88579-284-0
ISSN 1617-5468
Volume Editors
Prof. Dr. Markus Nüttgens
Universität Hamburg
22765 Hamburg, Germany
Email: markus@nuettgens.de
Prof. Dr. Oliver Thomas
Universität Osnabrück
49074 Osnabrück, Germany
Email: oliver.thomas@uni-osnabrueck.de
Assoc. Prof. Dr. Barbara Weber
Universität Innsbruck
6020 Innsbruck, Austria
Email: barbara.weber@uibk.ac.at
Series Editorial Board
Heinrich C. Mayr, Alpen-Adria-Universität Klagenfurt, Austria
(Chairman, mayr@ifit.uni-klu.ac.at)
Hinrich Bonin, Leuphana Universität Lüneburg, Germany
Dieter Fellner, Technische Universität Darmstadt, Germany
Ulrich Flegel, Hochschule Offenburg, Germany
Ulrich Frank, Universität Duisburg-Essen, Germany
Johann-Christoph Freytag, Humboldt-Universität zu Berlin, Germany
Thomas Roth-Berghofer, DFKI, Germany
Michael Goedicke, Universität Duisburg-Essen, Germany
Ralf Hofestädt, Universität Bielefeld, Germany
Michael Koch, Universität der Bundeswehr München, Germany
Axel Lehmann, Universität der Bundeswehr München, Germany
Ernst W. Mayr, Technische Universität München, Germany
Sigrid Schubert, Universität Siegen, Germany
Martin Warnke, Leuphana Universität Lüneburg, Germany
Dissertations
Steffen Hölldobler, Technische Universität Dresden, Germany
Seminars
Reinhard Wilhelm, Universität des Saarlandes, Germany
Thematics
Andreas Oberweis, Karlsruher Institut für Technologie (KIT), Germany
! Gesellschaft für Informatik, Bonn 2011
printed by Köllen Druck+Verlag GmbH, Bonn
Preface
The strategic importance of modelling is recognised by an increasing number of
companies and public agencies. Enterprise modelling delivers the blueprints
for co-designing organizations and their information systems, such that they
complement each other in an optimal way. Achieving this interplay requires a
multi-perspective approach that takes into account technical, organizational and
economic aspects. It also recommends the cooperation of researchers from different fields such as Information Systems, Business Informatics and Computer
Science.
The 4th International Workshop on Enterprise Modelling and Information Systems Architectures (EMISA 2011) addresses all aspects relevant for enterprise
modelling as well as for designing enterprise architectures in general and information systems architectures in particular. It is jointly organized by the GI Special Interest Group on Modelling Business Information Systems (GI-SIG MoBIS) and the GI Special Interest Group on Design Methods for Information
Systems (GI-SIG EMISA).
These proceedings feature a selection of 24 papers from academia and practice
on enterprise modelling, enterprise architectures, business process management
and measurements, analysis and explorative studies. We received 39 submissions which were thoroughly reviewed by at least three experts of the program
committee. Twelve contributions were selected as full papers for presentation at
the workshop and for publication in these proceedings. Additional, twelve papers were accepted as research-in-progress papers.
We would like to thank the members of the program committee and the reviewers for their efforts in selecting the papers and helping us to compile a highquality program. We would also like to thank the local organization, in particular
Nadine Blinn, Niels Müller-Wickop and Martin Schultz as well as Michael
Fellmann and Novica Zarvic for their assistance with organizing EMISA 2011.
We also thank Deborah Bunker and Ulrich Frank as keynote speakers. We hope
you will find the papers in these proceedings interesting and stimulating.
August 2011
Markus Nüttgens, Oliver Thomas, and Barbara Weber
Program Co-Chairs
Markus Nüttgens, University of Hamburg
Oliver Thomas, University of Osnabrück
Barbara Weber, University of Innsbruck
Program Committee
Witold Abramowicz, Poznan University of Economics
Antonia Albani, University of St. Gallen
Thomas Allweyer, FH Kaiserslautern
Colin Atkinson, University of Mannheim
Lars Bækgaard, Aarhus University
Jörg Becker, University of Münster
Khalid Benali, Nancy University
Martin Bertram, Commerzbank Frankfurt
Patrick Delfmann, University of Münster
Jörg Desel, FU Hagen
Werner Esswein, TU Dresden
Fernand Feltz, Centre de Recherche Public-Gabriel Lippmann Luxembourg
Ulrich Frank, University of Duisburg-Essen
Andreas Gadatsch, Hochschule Bonn-Rhein Sieg
Norbert Gronau, University of Potsdam
Wilhelm Hasselbring, University of Kiel
Heinrich Jasper, Freiberg University of Technology
Reinhard Jung, University of St. Gallen
Dimitris Karagiannis, University of Vienna
Stefan Klink, University of Karlsruhe
Ralf Klischewski, German University Kairo
Horst Kremers, CODATA-Germany
John Krogstie, University of Trondheim
Susanne Leist, University of Regensburg
Peter Loos, Universität des Saarlandes
Heinrich C. Mayer, University of Klagenfurt
Jan Mendling, Humboldt-University Berlin
Mirjam Minor, University of Trier
Bernd Müller, FH Wolfenbüttel
Bela Mutschler, Hochschule Ravensburg-Weingarten
Volker Nissen, TU Ilmenau
Andreas Oberweis, University of Karlsruhe
Sven Overhage, University of Augsburg
Hansjürgen Paul, Institut Arbeit und Technik
Erik Proper, Henri Tudor, Luxembourg and Radboud Univ. Nijmegen
Michael Rebstock, University of Applied Sciences Darmstadt
Manfred Reichert, Universität Ulm
Stefanie Rinderle-Ma, University of Vienna
Peter Rittgen, University of Borås
Michael Rosemann, Queensland University of Technology
Matti Rossi, Aalto University School of Economics
Frank Rump, Fachhochschule Emden/Leer
Gunter Saake, University of Magdeburg
Eike Schallehn, University of Magdeburg
Carlo Simon, Provadis Hochschule
Marten van Sinderen, University of Twente
Elmar J. Sinz, University of Bamberg
Susanne Strahringer, TU Dresden
Stefan Strecker, University of Duisburg-Essen
Klaus Turowski, University of Augsburg
Gottfried Vossen, University of Münster
Hans Weigand, Tilburg University
Mathias Weske, HPI Potsdam
Robert Winter, University of St. Gallen
External Reviewers
Camlon Asuncion
Nadine Blinn
Marc Buergi
Hans-Georg Fill
Sandro Georgi
Sandra Hintringer
Patrick Hitzelberger
Florian Johannsen
Azeem Lodhi
Benedikt Martens
Gunnar Martin
Syed Saif Ur Rahman
Sebastian Schlauderer
Boris Shishkov
Niksa Visic
Novica Zarvic
Gregor Zellner
Local Organisation
Nadine Blinn, University of Hamburg
Niels Müller-Wickop, University of Hamburg
Martin Schultz, University of Hamburg
GI-SIG MoBIS: Conceptual Modelling is pivotal
for analysing and designing information systems
that are in line with a company's long term strategy
and that efficiently support its core business processes. The Special Interest Group on Modelling
Business Information Systems (SIG MoBIS) within
the German Informatics Society (GI) is aimed at
providing a forum for exchanging ideas and solutions on modelling research within Information
Systems - both for researchers at universities and
experts in industry.
GI-SIG EMISA: The GI Special Interest Group on
Design Methods for Information Systems provides
a forum for researchers from various disciplines
who develop and apply methods to support the
analysis and design of information systems.
Contents
Full Papers
Beate Hartmann
Enterprise Architecture as an Instrument of Strategic Control ............................ 9
Stephan Aier, Sabine Buckl, Bettina Gleichauf, Florian Matthes,
Christian Schweda, Robert Winter
Towards a More Integrated EA Planning: Linking Transformation
Planning with Evolutionary Change .................................................................. 23
Agnes Nakakawa, Patrick Van Bommel, Erik Proper
Applying Soft Systems Methodology in Enterprise Architecture
Creation Workshops........................................................................................... 37
Stefanie Rinderle-Ma, Juergen Mangler
Integration of Process Constraints from Heterogeneous Sources
in Process-Aware Information Systems............................................................. 51
Robin Bergenthum, Jörg Desel, Sebastian Mauser
Workflow Nets with Roles................................................................................. 65
Hanns-Alexander Dietrich, Matthias Steinhorst, Jörg Becker,
Patrick Delfmann
Fast Pattern Matching in Conceptual Models Evaluating and
Extending a Generic Approach.......................................................................... 79
Ulrich Frank
Some Guidelines for the Conception of Domain-Specific
Modelling Languages......................................................................................... 93
Bernd Michelberger, Bela Mutschler, Manfred Reichert
Towards Process-oriented Information Logistics: Why Quality
Dimensions of Process Information Matter ..................................................... 107
Niels Mueller-Wickop, Martin Schultz, Nick Gehrke, Markus Nüttgens
Towards Automated Financial Process Auditing: Aggregation
and Visualization of Process Models ............................................................... 121
Fabian Christ, Benjamin Nagel
A Reference Architecture for Semantic Content Management
Systems ............................................................................................................ 135
Mohammed Abujarour, Ahmed Awad
Enriched Service Descriptions Using Business Process Configurations ......... 149
Jonas Repschläger, Stefan Wind, Rüdiger Zarnekow,
Klaus Turowski
Developing a Cloud Provider Selection Model ............................................... 163
Research-in-Progress Papers
Stefan Zugal, Jakob Pinggera, Barbara Weber
Assessing Process Models with Cognitive Psychology................................... 177
Frank Hogrebe, Nick Gehrke, Markus Nüttgens
Eye Tracking Experiments in Business Process Modeling:
Agenda Setting and Proof of Concept.............................................................. 183
Kathrin Figl, Barbara Weber
Creative Personality and Business Process Redesign ...................................... 189
Matthias Wester-Ebbinghaus, Daniel Moldt
Abstractions in Actor and Activity Modeling.................................................. 195
Sebastian Herwig, !ukasz Lis, Matthias Steinhorst,
Jörg Becker, Patrick Delfmann
A Generic Multi-purpose Conceptual Model Analysis Approach
Conceptual Specification and Application....................................................... 201
André Pflüger, Wolfgang Golubski, Stefan Queins
System Architecture Validation with UML ..................................................... 207
Robert Heinrich, Alexander Kappe, Barbara Paech
Tool Support for the Comprehensive Modeling of Quality
Information within Business Process Models.................................................. 213
Matthias Boehm, Carl Stolze, Oliver Thomas
Understanding IT-Management and IT-Consulting Teaching as
Product Service System: Application of an Engineering Model ..................... 219
Oliver Kopp, Frank Leymann, Sebastian Wagner
Modeling Choreographies: BPMN 2.0 versus BPEL-based Approaches........ 225
Michael Fellmann, Oliver Thomas
Process Model Verification with SemQuu....................................................... 231
Benedikt Martens, Novica Zarvic, Frank Teuteberg, Oliver Thomas
Designing a Risk-based Partner Selection Process for Collaborative
Cloud Computing Environments ..................................................................... 237
Andreas Zolnowski, Martin Semmann, Tilo Böhmann
Introducing a Co-Creation Perspective to Service Business Models............... 243
Enterprise Architecture as an Instrument of
Strategic Control
Beate Hartmann
Department of Information Systems Systems Engineering
University of Bamberg
Feldkirchenstr. 21
96045 Bamberg
beate.hartmann@uni-bamberg.de
Abstract: Enterprises compete in high dynamic and fast changing markets.
Reacting quickly and flexibly to market changes is a crucial competitive advantage
for an enterprise. Therefor premises of strategic planning and achievement of
strategic goals have to be controlled permanently and in a systematic way.
Strategic control which is part of strategic management monitors premises of the
corporate strategy and the environment of the enterprise in order to evaluate
whether the strategic positioning is still appropriate or whether the strategy has to
be adopted due to changes in constraints. This paper shows how the enterprise
architecture of an enterprise can be used as an instrument of strategic control.
Strategic control is characterized by a high demand on information. Hence one
needs to know where needed information is to gather. Enterprise architecture
which is enriched by results and information of strategic management process is a
helpful instrument in delivering information respectively the source of gathering
information used in strategic control. In particular for each enterprise architecture
layer a proposal is given how strategic control can be supported.
1 Introduction
Enterprise Architecture (EA) is used in enterprises in many different ways [ARW08].
Regardless of concrete use the goal of EA is to create an IT environment which is
focused on the corporate strategy [Mi08]. Hence it is mandatory that the strategic
positioning of an enterprise is reflected in the EA in order to be of any value to the
enterprise [Bo10]. The goals of an EA and in particular the goals of business information
systems have to match with the overall strategic goals of the whole enterprise. Therefore
it is essential to design and if necessary to redesign the EA within the strategic
management process [Go06]. Thus information about the process of strategic analysis
and choice are implicit contained in the EA. This information, in turn, can be used in
strategic control.
Premises which are defined within the process of strategic planning are the fundament of
a corporate strategy [BS95]. The strategic positioning of an enterprise is established on
9
those premises. Therefore it is of strategic importance to first document these premises
carefully and second to monitor them permanently. Should there be changes in premises,
strategy might be adopted respectively modified. SAAT ET AL. state, that external forces
such as changes in premises are beyond the scope of EA models [SAG09, p. 4].
Nevertheless the EA provides hints, where those changes in premises are noticed first in
daily business. Another task of strategic control (see chapter 2.1) is to measure the
degree of target achievement, i.e. to control whether the strategic goals of an enterprise
are achieved and the intended strategy is implemented successfully. In order to measure
the degree of target achievement one needs to know which processes respectively which
business application systems or business units were responsible for achieving certain
goals. Also in that case EA, where processes and application systems are mapped in a
model, is able to provide valuable information.
The research question is: How can EA support the strategic control of an enterprise?
This paper introduces a research framework for answering this question. A proposal is
submitted, how information from strategic management process which is needed within
strategic control can be included into EA respectively into an EA model.
The remainder of this paper is structured as follows: In chapter 2 conceptual
fundamentals are given. The understanding of the terms strategic control and enterprise
architecture is stated and related work is given. In chapters 3 - 5 the proposals for
supporting strategic management in general and strategic control in particular are stated
for each model layer of an EA. Chapter 6 concludes with a short summary and an
outlook on further research needs.
2 Conceptual Fundamentals
This chapter provides an overview of conceptual fundamentals which are basic for
understanding the topic and of research work which is related to that topic. The terms
strategic control and enterprise architecture are explained. Based on this understanding
related approaches are stated.
2.1 Understanding of Strategic Control
Strategic Control in Strategic Management Process
Most of the authors in the fields of strategic management and corporate governance
consider strategic control to be part of strategic management [Fr10, p. 171-176; PR94, p.
16; Ha06; SS00, p. 157]. HAHN characterizes control as an obligatory addition of any
planning processes [Ha06]. Strategic control therefore contains the control of strategic
planning respectively strategic maps. Strategic control does not start at the end of the
planning and implementation process but at the very beginning of the strategic
management process. As soon as assumptions within the strategic planning process are
made, those so-called premises have to be overseen [SS87]. Strategic control is
concerned with tracking a strategy as it is being implemented, detecting problems or
changes in its underlying premises, and making necessary adjustments. [PR94, p. 381].
10
Hence strategic control has to be forward-looking since the end result of a strategic
process is still several years off [PR94, p.381] and realized parallel to planning and
implementation processes [Ha06]. Figure 1 clarifies this meaning of strategic control.
Figure 1: Strategic Control in strategic management process
Types of Strategic Control
STEINMANN and SCHREYÖGG distinguish between three types of strategic control:
premise control, implementation control and strategic surveillance [SS87]. Premise
control is designed to check systematically and continuously whether the premises on
which the strategy is based are still valid. [PR94, p. 382]. Thus premise control is
focused on noticing changes in premises as early as possible in order to allow an early
shift in strategy if necessary. Implementation control checks milestones during strategy
implementation in order to recognize deviations from the chosen strategic course [SS87].
Strategic surveillance finally is defined by STEINMANN and SCHREYÖGG as an unfocused
control. The environment of an enterprise and the enterprise´s resources are to be
monitored for information which could be threats but also opportunities for the
enterprise´s continuity.
Problems of Strategic Control
One thing all types of strategic control have in common is a high demand on
information. The kind of desired information is known whereas the source of
information gathering is partly unknown or undefined within an enterprise. Once it is
known who or what (e.g. a business application system) is responsible for collecting the
desired information and where they occur first, the enterprise´s management can respond
quickly to threats or opportunities. A quick respond to market movement is an important
strategic competitive advantage. Besides for instance predefined milestones in the
strategic implementation process it is hardly possible to determine ahead at which time
which information has to be gathered [SS00, p. 248f]. Nevertheless the strategic course
has to be monitored in a structured manner. The EA of a company can make a valuable
contribution to this. Before making proposals for that, the understanding of enterprise
architecture in this paper is given.
2.2 Understanding of Enterprise Architecture
A common definition for EA is not given in the business information systems
community so far. In this paper the understanding of architecture as the fundamental
organization of a system embodied in its components, their relationships to each other,
and to the environment, and the principles guiding its design and evolution [IEEE00]
11
follows the definition by the IEEE. Hence EA is understood as the fundamental
organization of an enterprise.
An EA can only be established successfully in an enterprise if it is considered being a
means to an end. Architecture is a means, and should not become a goal itself [Op09,
p. 45]. OP´T LAND ET AL. state EA amongst others as a mean within strategic
Business/IT-Planning and to align strategic objectives and IT [Op09, p. 40]. Every
intended use should create value to an enterprise. The benefit of an EA will be the
greater the stronger EA is positioned in accordance to goals and objectives of an
enterprise.
An EA model maps the components of an enterprise, relationships within an enterprise
and to the environment into a model. There are various approaches for EA modelling.
Most frameworks for EA modelling disperse the core artefacts of EA to five layers in
order to reduce complexity [WF07]: Business architecture, Process architecture,
Integration architecture, Software architecture and Technology architecture. The
methodological foundation of this approach is the Semantic Object Model (SOM) by
FERSTL and SINZ [FS95; FS06], which is an object- and business process-oriented
methodology for modelling business systems. In the SOM EA the core artefacts
identified by WINTER and FISCHER [WF07] are comprised in three model layers: (1)
enterprise plan, (2) business process model and (3) resource model (see figure 2).
Figure 2: SOM enterprise architecture [FS06, p. 349]
Business objects at the business process model layer have to fulfil goals and objectives
which are deduced from the enterprise plan layer. Furthermore the SOM methodology
provides a concept for explicitly linking business application systems to business process
models [FS06]. Hence information from strategic management process can be
considered continuously from enterprise plan to business application systems.
Strategies respectively strategic goals are initially specified at the first layer of the SOM
EA. Premises of a corporate strategy are not modelled at all in any EA model so far.
12
2.3 Related Work
The design of an EA should be a task of the enterprise´s strategic management. ROSS ET
AL. even claim, that enterprise architecture should be a strategy itself [RWR06]. Most
researchers position the design of EA in the strategic planning process [Op09, p. 40;
RA09; Go06]. In accordance to that BOUCHARAS ET AL. identify in a broad literature
review as benefits of EA better strategic planning and Decision Making [Bo10].
Thus EA is considered to be of value in the strategic planning process. SMOLANDER ET
AL. suggest the metaphor that architecture represents decisions [SRP08]. Following this
metaphor EA represents management´s decisions, esp. strategic planning decisions. A
benefit of EA in strategic control in contrast is barely mentioned. An extensive literature
review of publications in journals of the AIS1 Senior Scholars´ Basket of journals
brought no hits searching for articles linking EA and strategic control. There have been
some research works, especially within the TEAR2-workshops, concerning the benefits
of EA. The main focus of articles reviewed by BOUCHARAS ET AL. in the paper The
Contribution of Enterprise Architecture to the Achievement of Organizational Goals
lies on empirical studies analysing the benefit of EA in practice [Bo10]. However
strategic control cannot be identified as a context, for which EA is of value. Possibly it is
because, to the best of my knowledge, the potential benefit of EA to strategic control has
not been analysed so far.
3 Supporting Strategic Control by the Enterprise Plan
In this chapter a proposal is introduced how information generated and used in strategic
management process (in the following called strategic information) can be integrated in
EA models and in turn can be used in strategic control. The following focuses on the
enterprise plan layer of the EA model (see fig. 2). The proposals for supporting strategic
control by other layers of the EA are given in chapters 4 and 5.
Studies show that premises, the underlying assumptions of every strategy, are not
documented well [BP03]. But documentation is required in order to check the premises
within strategic control whether they are still valid or changes have occurred.
Furthermore goals of business information systems should coincide with the strategic
positioning of an enterprise. Therefore premises as well as strategic goals should be
documented in the enterprise plan layer.
The enterprise plan of the SOM methodology takes an outside perspective onto the
business system [FS06]. In the system of objectives and goals, which is the behavioural
view of the enterprise plan, the planned strategies, goals and objectives as well as
constraints are modelled. FERSTL and SINZ do not provide a metamodel for the design
objects of the enterprise plan layer, since the modelling of the SOM methodology
focuses on the second and third layer of the SOM EA. But in order to document
information needed in strategic control in a structured manner a formal description of
1
2
Association for Information Systems
Trends in Enterprise Architecture Research
13
premises and strategic goals is needed. Therefore following classification is introduced
which leads to the metamodel in figure 3. The relationship between strategies, strategic
goals, measured values and actions bases on the concept of the Balanced Scorecard by
KAPLAN and NORTON [KN97] which is a management tool for translating strategy into
action.
Premises are assumptions on which a strategy is based on (see chapter 2.1). A premise
can be basis of several strategies and a strategy usually bases on more than one premise.
A premise is a specific constraint. There exist further constraints, e.g. regulation on
taxation or accounting. Those constraints are out of focus here because they affect all
competitors in the same way and are not assumptions of strategic positioning. Premises
can be further differentiated into critical and non-critical premises or internal and
external premises. This can be modelled either as another meta object or as attributes. A
strategy aims one or more strategic goals. Conversely a strategic goal can be assigned to
several strategies. A strategic goal is a specific goal. Besides strategic goals there are
also operational goals, for example. Since in this paper information used in strategic
management is considered, other than strategic goals are out of focus. A goal can be split
into sub goals. Goals respectively their achievement are restricted by constraints. For
every goal at least one measured value has to be defined which measures the degree of
goal achievement. A measured value belongs to exactly one goal. A target value is given
to each measured value as an attribute of this meta object. In order to operationalize a
strategy for every goal at least one action is defined. An action in turn can contribute to
achieve several goals.
Figure 3: Metamodell for the considered part of the enterprise plan layer
This metamodel is the basis for supporting strategic control by EA, since strategic
information is initially attached to the top layer of the EA model. Further layers are
modelled under consideration of this strategic information.
Next an example for illustration purposes is introduced. e-Car plc. is a company that
deals with electric cars. The company entered the German automobile market recently.
Table 1 shows an excerpt from strategies and underlying premises which were defined
by the management of e-Car plc. This example is a modification and extension of the
example given in [LWF10].
14
Premise
A new product can only be launched on the
market at a moderate price level
Growing demand for electric cars
Government aid for electric cars remains
until 2014
Strategy
Cost leadership
Growth strategy
Cost leadership
Table 1: Example premises and strategies
Figure 4 shows the relationship between strategic goals, actions, measured values and
target values as an extension of the metamodel defined above: A top-level strategic goal
is to be a reliable partner. In consideration of the strategies (see tab. 1) a subgoal is e.g.
pay on time. The achievement of this goal is evaluated by the measured value date of
payment. Target value for goal achievement is one day before the contractual date of
payment. In order to achieve this goal the action diligent billing management should
be implemented.
Figure 4: Modelling of strategic goals at enterprise plan layer
4 Supporting Strategic Control by Business Process Models
At business process model layer solution procedures for realizing the enterprise plan are
specified [PS10]. That means that the implementation of the corporate strategy is
modelled. Hence the defined actions in the enterprise plan have to be reflected in the
business process model. Premises and strategic goals defined in the enterprise plan
should be mapped to certain business objects in order to clarify which business object is
responsible for putting defined actions into practice and for delivering which strategic
15
information. Before this mapping is explained in detail a short introduction into the
modelling method of the SOM methodology is given.
In the SOM methodology there exist two views of a business process model. The
interaction schema is the view on structure whereas the task-event schema is the view
on behavior [FS06]. Strategic goals and constraints are only modelled in the structural
view. In the interaction schema business objects and business transactions which
transmit goods, services or messages between these business objects are modelled. A
short example illustrates the modelling method. The negotiation between a company and
a customer consists of an initiating transaction (i), a contracting transaction (c) and an
enforcing transaction (e). Fig. 5 shows the structure of this business process.
Figure 5: Interaction schema
4.1 Modelling of strategic goals
In strategic control the achievement of strategic (interim) goals is monitored.
Furthermore it is checked which actions put into practice successfully and which actions
missed their targets. Hence strategic goals and their sub goals have to be mapped to
business objects. These business objects are then responsible for translating these goals
and for implementing assigned actions.
In the SOM methodology every business object has to fulfil goals and objectives which
are given by a management object. Goals and objectives in general and strategic goals in
particular are modelled as follows. A management object (e.g. Management in fig. 6)
sets goals and objectives of an operational object (e.g. Production in fig. 6). This goal
relation is modelled as g-transaction. The operational object in return informs the
management object about the degree of goal achievement [FS06].
Example e-Car plc.: The two top-level goals Be a reliable partner and Profit
maximization are mapped to the business object Management (see fig. 6). The
subgoals are assigned adequately by this management object to the operational objects
Production and Sales. For example, the subgoal Pay on time has to be fulfilled by
the business object Production.
16
g:
P
g: ay o
D
g: eliv n tim
Go ery
e
g:
Lo od q on t
w C ua
im
l
os ity e
ts
me
olu
s v ime
t
ale
h s on
Hig ery
g: eliv
D
g:
Figure 6: Modelling of strategic goals at business process model layer
Every action can be mapped to a transaction of the business process model [Me96, p.
153]. The result of an action is a service which is performed either by the sender or by
the receiver of a transaction.
Example e-Car plc.:
Action
Quality aspects are
part of contracts with
delivers
Diligent billing
management
Service
Working out adequate
contracts
Transaction
c: PO preproducts
Registration of invoices
immediately and check
by a second person
c: invoice
Table 2: Actions and transactions
The action Quality aspects are part of contracts with delivers (goal: Good quality) has
to be implemented by working out adequate contracts, which is prerequisite for the
transaction PO pre-products. The business sub object Procurement sends this
transaction. The action Diligent billing management (goal: Pay on time) is applied
through registration of invoices immediately and checking by a second person whenever
the transaction invoice reaches the business sub object Production finance (see tab. 2
and fig. 7). For the sake of clarity, modelling of measured values and target values in the
business process model are omitted.
For modelling objectives and therefore for mapping them form enterprise plan layer to
business process model layer rules are defined, e.g.:
"
"
A modelled goal in the business process model has to correspond to a goal
defined in the enterprise plan respectively has to be assigned to a goal as a sub
goal. (Secures continuous goal-orientation and consistency)
A strategic goal and corresponding actions have to be linked in the same
business object respectively in a controlled business object. (Secures that a
business object is able to achieve the strategic goal)
17
g: Good Quality
g:
g: Pay
g: Deli on
g: Goo very time
Lo d
w q u o n ti
Co ali m
e
sts ty
g
g :P
pu : F a ay o
rc vo n
ha ra tim
se ble e
g:
g: Hig
De h s
liv ale
er
y o s vo
n t lum
im
e e
g:
g: Goo
De d
liv qua
er
y o lity
nt
im
e
Figure 7: Modelling of actions at business process model layer
(excerpt, transactions within Production are omitted)
Modelling strategic goals and assigned actions in this detailed way ensures that business
processes are aligned to corporate strategy.
4.2 Modelling of Premises
Another important task of strategic control is to monitor the premises of strategies. The
question raises how and where those premises are to be monitored and who is
responsible for that. SCHREYÖGG and STEINMANN suggest the business unit which is
factual competent to monitor premises. [SS00, p. 249]. Thus premises are attached to
business objects in the business process model following the rules:
"
"
Which business object notices a change of a premise first?
In which transaction between business objects can a changing premise be
noticed first?
For modelling premises the notation of modelling context of business systems is
adopted. WAGNER and FERSTL define the context of business objects and transactions as
influencing factors of their environment which are not already modelled in
transactions [WF09, p. 4]. Hence context is understood as specific constraints of
entrepreneurial activities. Since premises are also understood as constraints the use of
this notation is justified.
18
Figure 8: Modelling of premises at business process model layer
Example e-Car plc.: The premise Growing demand for electric cars is characterised by
a high number of requests and orders. These transactions reach the business object
Sales (see fig. 8). The premise A new product can only be launched at a moderate
price level is overseen by the same business object through analysing the realized sales
numbers and prices. In the decomposition of the business object Sales premises can be
attached to the business sub objects Sales administration and Sales management (see
fig. 9). The premise Government aid remains until 2014 has to be overseen by the
management of e-Car plc.
g
g
g
g
g
Figure 9: Modelling of premises at business process model layer - refinement
(excerpt, transactions within Sales are omitted)
For modelling premises and therefore for mapping them from enterprise plan layer to
business process model layer rules are defined, e.g.:
19
"
"
Premises which are not modelled in the business process model layer are to be
monitored by the management (e.g. business object Management in figure 8).
A premise can also be attached to more than one business object.
Once it is known which business objects have to monitor which premises the business
objects are then responsible for the way they monitor premises. Business application
systems can support this monitoring (see chapter 5). A procedure for dealing with
changed premises has to be defined by corporate governance, which is out of focus here.
5 Supporting Strategic Control by Business Application Systems
Software development projects, thus also development projects of business application
systems supporting business processes, fail frequently because of lacking goal
synchronization between software development project goals and goals of the whole
enterprise [BM11]. Therefore it is necessary to take strategic goals during specification
of the business application system into account.
The third layer of the SOM EA (see fig. 2) comprises the resource model. It describes
the actors for carrying out the business tasks of the business system [PS10, p. 59].
Those resources are needed for execution of business processes. Besides the
specification of business application systems also the organization of personnel is
specified. The latter is not considered in the following since this is a classical field of
organization theory. Nevertheless considering the strategic positioning of an enterprise is
also very important in personnel decisions. The better the first and second layer of the
SOM EA is modelled and the more information about strategic positioning is included,
the stronger business application systems will be aligned to strategic goals, which is an
important contribution to Business-IT-Alignment.
In a business process model goals are attached to single business objects. One or more
business objects are linked to one business application system. I.e. the linked business
application system supports the execution of the business processes. For more
information how derivation is done in detail see [FS06]. Therefore it is explicitly
specified that a business application system has to support the goals respectively the
actions which are attached to the business objects linked to this application system. If the
measured value of these goals can be measured automatically then checking the degree
of goal achievement is easy. E.g. if the payment process is IT-supported then it can be
automatically checked, whether the outstanding accounts of the business object
Production finance are paid one day before the date agreed on in the contracts (see fig.
7). The same is true for the desired information in strategic premise control. Premises
are also attached to business objects und therefore - if changes in premises can be
determined automatically - also to one or more business application systems. An
example would be the number of queries respectively conclusions of contract which are
directed to the business object Sales administration via a web shop (see fig. 9).
20
6 Summary and Outlook
Detailed information about the parts of an enterprise and their relationship to each other
are included in the EA. Enriched by information from strategic management process the
EA can be a valuable instrument of strategic control. The paper on hand introduces a
proposal how information respectively the source of gathering information needed in
strategic control like changes in premises or responsibility for achieving strategic goals
can be taken from the EA. The other way round putting information from strategic
management process into all layers of an EA model ensures that business process models
and business application systems are going along with the strategic positioning of an
enterprise.
An essential prerequisite for EA being an instrument of strategic control can be deduced
from that: The EA has to be developed within the strategic management process and has
to be adopted when changes in strategy are made.
Further need for research is necessary in defining this involvement in the process of
strategic management. Furthermore tool support in modelling strategic information at all
layers of the SOM enterprise architecture is needed. Tool-supported modelling of EA
especially documenting premises and strategic goals in an enterprise architecture tool is
mandatorily required for understanding a strategic positioning of an enterprise in order to
follow this positioning in further modelling of the EA.
References
[ARW08]Aier, S.; Riege, C.; Winter, R.: Unternehmensarchitektur - Literaturüberblick und Stand
der Praxis. In Wirtschaftsinformatik, 2008, 50; pp. 292304.
[BM11] Buhl, H. U.; Meler, M. C.: Die Verantwortung der Wirtschaftsinformatik bei ITGroßprojekten. Symptome, Diagnose und Therapie. In Wirtschaftsinformatik, 2011; pp.
5962.
[Bo10] Boucharas, V. et al.: The Contribution of Enterprise Architecture to the Achievement of
Organizational Goals: A Review of the Evidence. In (Proper, E. et al. Eds.): Trends in
Enterprise Architecture Research. 5th International Workshop, TEAR 2010, Delft, The
Netherlands, November 12, 2010. Proceedings. Springer, Berlin, Heidelberg, 2010; pp.
1-15.
[BP03] Becker, W.; Piser, M.: Strategische Kontrolle. Ergebnisse einer empirischen
Untersuchung. Otto-Friedrich-Universität, Bamberg, 2003.
[BS95] Band, D. C.; Scanlan, G.: Strategic control through core competencies. In Long Range
Planning, 1995, 28; pp. 102114.
[Fr10]
Freeman, R. E.: Strategic management. Univ. Press, Cambridge, 2010.
[FS95] Ferstl, O. K.; Sinz, E. J.: Der Ansatz des Semantischen Objektmodells (SOM) zur
Modellierung von Geschäftsprozessen. In Wirtschaftsinformatik, 1995, 37; pp. 209220.
[FS06] Ferstl, O. K.; Sinz, E. J.: Modeling of Business Systems Using SOM. In (Bernus, P.
Ed.): Handbook on architectures of information systems. Springer, Berlin, 2006; pp.
347367.
[Go06] Goethals, F. G. et al.: Management and enterprise architecture click: The FAD(E)E
framework. In Information Systems Frontiers, 2006, 8; pp. 6779.
21
[Ha06]
Hahn, D.: Strategische Kontrolle. In (Hahn, D.; Taylor, B. Eds.): Strategische
Unternehmungsplanung - Strategische Unternehmungsführung. Springer-Verlag Berlin,
Heidelberg, 2006; pp. 451-464.
[IEEE00] IEEE Computer Society: IEEE Recommended practice for architectural description of
software-intensive systems, New York, 2000.
[KN97] Kaplan, R. S.; Norton, D. P.: The balanced scorecard. Harvard Business School Press,
Boston, Mass, 1997.
[LWF10] Leuning, B.; Wagner, D.; Ferstl, O. K.: Hochflexible Geschäftsprozesse in der Logistik ein Integrationsszenario für den Forschungsverbund forFLEX. Bayerischer
Forschungsverbund Dienstorientierte IT-Systeme für hochflexible Geschäftsprozesse
(forFLEX), Bericht-Nr. forFLEX-2010-001, 2010.
[Me96] Mehler-Bicher, A.: Ein objekt- und geschäftsprozeßorientiertes Architektur-modell für
Management-Support-Systeme. Tectum-Verl., Marburg, 1996.
[Mi08] Minoli, D.: Enterprise architecture A to Z. CRC Press/Taylor & Francis, Boca Raton,
Fla., 2008.
[Op09] Op't Land, M. et al.: Enterprise architecture. Creating Value by Informed Governance.
Springer, Berlin Heidelberg, 2009.
[PR94] Pearce, J. A.; Robinson, R. B.: Strategic management. Formulation, implementation, and
control. Irwin, Burr Ridge, Ill., 1994.
[PS10] Pütz, C.; Sinz, E. J.: Model-driven Derivation of BPMN Workflow Schemata from SOM
Business Process Models. In Enterprise Modelling and Information Systems
Architectures, 2010, 5; pp. 5772.
[RA09] Riege, C.; Aier, S.: A Contingency Approach to Enterprise Architecture Method
Engineering. In (Lamersdorf, W.; Feuerlicht, G. Eds.): Service-oriented computing ICSOC 2008 workshops. ICSOC 2008 international workshops, Sydney, Australia,
December 1, 2008 ; revised selected papers. Springer, Berlin, Heidelberg, 2009; pp.
388399.
[RWR09] Ross, J. W.; Weill, P.; Robertson, D. C.: Enterprise architecture as strategy. Creating a
foundation for business execution. Harvard Business School Press, Boston, Mass., 2009.
[SAG09] Saat, J.; Aier, S.; Gleichauf, B.: Assessing the Complexity of Dynamics in Enterprise
Architecture Planning - Lessons from Chaos Theory. AMCIS 2009 - Proceedings of the
15th Americas Conference on Information Systems. San Francisco, California, 6-9
August 2009.
[SRP08] Smolander, K.; Rossi, M.; Purao, S.: Software architectures: Blueprints, Literature,
Language or Decision? In European Journal of Information Systems, 2008, 17, pp 575588.
[SS87] Schreyögg, G.; Steinmann, H.: Strategic Control: A New perspective. In Academy of
Management Review, 1987, 12; pp. 91103.
[SS00] Steinmann, H.; Schreyögg, G.: Management. Grundlagen der Unternehmensführung;
Konzepte, Funktionen, Fallstudien. Gabler, Wiesbaden, 2000.
[WF07] Winter, R.; Fischer, R.: Essential Layers, Artifacts, and Dependencies of Enterprise
Architecture. In Journal of Enterprise Architecture, 2007; pp. 112.
[WF09] Wagner, D.; Ferstl, O. K.: Enhancing Flexibility and Precision of Enterprise
Architectures by Modeling Context-Aware Business Processes. Bayerischer
Forschungsverbund Dienstorientierte IT-Systeme für hochflexible Geschäftsprozesse
(forFLEX), Bericht-Nr.: forFLEX-2009-003, 2009.
22
Towards a More Integrated EA Planning: Linking Transformation Planning with Evolutionary Change
Stephan Aier1, Sabine Buckl2, Bettina Gleichauf1, Florian Matthes2,
Christian M. Schweda2, Robert Winter1
1
Institute of Information Management, University of St. Gallen
Müller-Friedberg-Strasse 8, 9000 St. Gallen, Switzerland
{stephan.aier | bettina.gleichauf | robert.winter}@unisg.ch
2
Chair for Software Engineering of Business Information Systems (sebis),
Technische Universität München
Boltzmannstr. 3, 85748 Garching, Germany
{sabine.buckl | matthes | christian.m.schweda}@mytum.de
Abstract: Enterprises are subject to continuous change driven by top-down
planned transformation projects as well as by bottom-up initiatives realizing what
is called the evolution of the enterprise. Enterprise architecture (EA) planning presents itself as a means for facilitating and controlling this change. Nevertheless, the
methods and models of EA planning developed in recent years either have a strong
focus on planned (proactive) transformations or on guided (reactive) evolution. In
this paper, we outline an EA planning method that accounts for both types of enterprise change by illustrating the interplay of EA planning, requirements, release,
and synchronization management. Specifically we focus on the coordination of design activities modeled as intermeshed closed-loop control systems and on an integrated information model describing EA transformation planning.
Keywords: Enterprise architecture planning, methods, models, techniques
1 Introduction
Todays enterprises find themselves confronted with ever changing economic, regulatory, and technical environments that they are forced to continuously adapt to [RWR06],
[Wa05]. Performing changes that are necessary or help to leverage idle opportunities is a
complex task, aggravated by the intricate and highly interwoven architecture of the overall organization. Thereby, architecture or enterprise architecture (EA), respectively, is
understood as the fundamental organization of a system [the enterprise] embodied in its
components, their relationships to each other, and to the environment, and the principles
guiding its design and evolution [IOS07].
A commonly accepted instrument to guide enterprise transformations is EA planning,
whose main goal is to enhance and maintain the mutual alignment of business and
IT [HV93], [LLO93]. Effectively established, EA planning as part of an EA management (EAM) function provides an organization-wide perspective on change, making the
impact of projects on EA elements transparent. While business is commonly seen as the
driver of organizational change, projects are the implementers and by their nature a
23
cross-cutting aspect. Changes to an enterprise can be differentiated into evolution (incremental change, usually bottom-up driven) and transformation (fundamental change,
usually top-down driven). Whereas support for the former type of change is typically
provided by functional methods of business administration, e.g. human resources, distribution, or marketing, the latter requires a holistic management function to systematically
support organizational transformation [Wi08]. Although EA planning is commonly regarded as a core capability of EAM [Pu06], only a few authors propose well-engineered
methods and models for EA planning. Especially, the need to integrate EA planning with
existing management functions like enterprise-wide requirements management or project
portfolio management to derive consistent transformation plans as well as synchronization management to implement changes is not considered in existing approaches. Therefore, this paper contributes to the field of EA planning by answering the following research questions:
!
How can top-down driven, planned transformation and bottom-up driven guided
evolution of the enterprise be consistently integrated in EA planning?
!
Which methods or conceptualizations from related disciplines (requirements, project portfolio, and synchronization management) are suitable for consistently planning transformation and evolution aspects of enterprise change?
In order to answer the above questions, we follow the identify problem/motivate define
objectives design demonstrate [evaluate] process by [Pe07]. Section 2 motivates
the need for an integrated EA planning method and model and derives relevant requirements. Prefabrics from the areas of EA planning, change, and requirements management
as well as project portfolio, and synchronization management are revisited in Section 3.
Our artifact a method for integrated EA planning is presented in Section 4. The theoretic discussion on the artifact is complemented in Section 5 by a case study, which
serves as a demonstration and initial evaluation 1. Finally, Section 6 critically reflects our
contribution and provides an outlook on further research.
2 Deriving requirements for integrated planning
Matthes et al. conducted a survey on the state-of-the-art in EAM tools [Ma08]. Central
means of analyzing these tools is a set of nine scenarios that describe typical activities
of EAM. The scenario descriptions provide core concerns on an abstract level which are
further concretized with questions that are to be answered in performing the activity.
While initially developed as a set of requirements for analyzing EAM tools, the scenarios also provide a basis for deriving requirements for an integrated EA planning method
and underlying model. In short the methodology for scenario elicitation can be summarized in five steps as follows [Ma08]:
!
Derive initial scenario description from project experiences, scientific literature,
and end-user input
1
Although our demonstration is based on a real industry case, it is not hold a full evaluation. However, this
single case provides support for the proposed artifact to justify its discussion in the scientific community by
this publication.
24
!
Conduct on-site workshops with the surveys industry partners
!
Revise scenario descriptions to on-site feedback provided by reviewers and add or
reshape scenarios appropriately
!
Conduct remote review round with industry partners of the survey
!
Revise scenario descriptions to feedback
Following this methodology, a comprehensive set of scenarios based on existing literature and further leveraging the knowledge of 30 user companies has been compiled.
Three of these scenarios are essential for EA change processes, namely enterprise-wide
requirements management, project portfolio management, and synchronization management. In the following the concerns of these scenarios are briefly summarized and the
corresponding questions are reformulated as requirements for an integrated EA planning
method and underlying model.
Enterprise-wide requirements management is concerned with gathering and documenting
the requirements raised at different times and in different parts of an organization. A
requirement describes a desired state of an organizations EA or part thereof. Affected
EA elements can reside on all EA layers from business to infrastructure. In the management process these requirements are classified in respect to their urgency and are
grouped by affected EA elements. Thus, typical requirements of enterprise-wide requirements management read as follows:
(R1)
Collect and document requirements in a uniform way
(R2)
Categorize requirements in respect to their urgency (must vs. may)
(R3)
Link requirements to affected EA elements
(R4)
Derive project proposals from requirements
Project portfolio management decides which project proposals are executed as projects
in the next planning period. The process plans a short term future state of the EA based
on various measures and takes a transformation-centric perspective by analyzing which
EA elements are changed by which project. By doing so, the management process diagnoses potential for alignment between change activities. Such analyses can uncover that
a project targets optimizations to an element which is conversely retired by another project. Project portfolio management raises the following requirements:
(R5)
Estimate project proposals costs and economic impact
(R6)
Analyze and resolve dependencies between transformation and evolution project proposals targeting the same EA element
(R7)
Ensure that urgent project proposals are prioritized high and scheduled on short
time initiation
(R8)
Assign and distribute project budgets
Synchronization management establishes a common time-schedule for approved projects
respecting their interdependencies. In particular, projects that aim at affecting the same
EA elements at the same time but may not be integrated into a single project are re-
25
scheduled to minimize project interplay. Synchronization management further is no onetime activity but has to continually track project states such that actions for re-scheduling
can be taken early, if necessary. Thus, the following requirements can be derived:
(R9)
Schedule projects to minimize pertaining EA-induced interdependencies
(R10)
Control project progress and re-schedule in case of project delays
Reflecting our research questions (see Section 1) requirements (R1), (R3), (R4), (R6),
and (R9) are of interest and are addressed by our method and model.
3 Preliminary and related work
The question of engineering enterprise transformation using EA models has been addressed in detail by [AG10b]. However, [AG10b] do not consider evolutionary bottomup change initiatives that may occur on a local project level. We aim at enhancing the
existing method towards an integrated approach that reflects methods and models from
enterprise-wide requirements, project portfolio, and synchronization management (R1)
and thus integrates centrally derived transformation plans as well as locally triggered
evolutionary change (R6). Figure 1 shows how EA planning depends on other processes.
Requirements Management
Product Portfolio
Management
Business Process
Management
Application Portfolio
Management
IT Infrastructure
Management
requirements
EA Management
Business and IT
Strategy
Development
as-is models
strategic
requirements
EA Planning
EA As-Is
Documentation
potentials
EA Analysis
EA principles,
to-be models,
project
descriptions
as-is state
as-is EA
Implementation
Release
Management
Project Portfolio
Management
Synchronization
Management
Figure 1: Contextual diagram of the EA planning process [AG10a]
The EA planning process is one of several EAM processes, complementing EA as-is
documentation as well as EA analysis processes. EA planning is driven by requirements
from the different sub-architectures. As the planning process is based on EA models,
information about as-is models and condensed information from EA analysis is needed.
Besides that, strategic principles from business and IT strategy processes guide the EA
planning process. Main outputs of the planning process are to-be models of future architecture states as well as project descriptions about development projects that will implement the planned enhancements. Those project proposals are accompanied by EA principles that are developed within the EA management processes and distributed into im-
26
plementation processes via EA planning. The actual implementation is then controlled
by release, project portfolio, and synchronization management processes. From there,
information about the as-is state of EA is fed back to EA management and to subarchitectures management processes. In Subsection 3.1 we will review approaches for
top-down transformation planning. In Subsection 3.2 we will complement these by approaches for bottom-up planned evolutionary change.
3.1 Top-down transformation planning
[AG10a] present a method for EA transformation that puts emphasis on the methods
outputs as well as on their interdependencies.
The result of the first step of the method is a list of EA elements that need to be changed
in order to reach the targeted to-be state (R3). Those EA elements are the core subjects
of transformation activities. Additionally, as EA elements must not be considered as
stand-alone artifacts, information about their relationships to other EA elements needs to
be captured as well. Thus, the method produces a list of predecessor/successor relationships linking the as-is with the to-be state, as a basis for plan development activities.
These relationships are covered in a transformation model.
The second step plans transformation in detail. The steps result is a description of project proposals (R4) containing a list of preconditions, a list of affected EA elements and
a relative time line that respects individual life cycles of and temporal interdependencies
between EA elements (R9). Based on these proposals actual projects can be defined in a
third step, taking available budgets and resources into account.
The analysis of differences in the as-is and the to-be state identifies successor relationships between model elements which represent successor relationships of enterprise
artifacts, more precisely of their lifecycle phase in production. From there, the general
dependencies of pre-production and decommissioning lifecycle phases can be derived
for each segment. The actual sub-division of such pre-production phases varies depending on the type of artifact considered and may comprise various development activities,
including specification or testing. Preceding analyses should be grounded in an according information model describing the concepts of the enterprise and the transformation
model. [Bu09] have presented such a model introducing the notion of ProjectAffectable
and applying it to the EA model concepts business application and business support,
respectively. Figure 2 shows an adapted version of this model, which also accounts for
temporality aspects of EA planning. Thereto, the start and end dates of the projects
work packages are temporal attributes in the sense of the corresponding pattern of
[CEF99]. This means that the transformation model can change over time entailing an
adaptation of the duration of the work packages (R10).
27
Figure 2: Information model adapted from [Bu09]
The next steps of the EA planning method aim at scheduling an effective sequence of
development activities upon the affected artifacts (R9). Therefore, the change-induced
relationships between the model elements must additionally be taken into account. A
temporal restriction, for example, may arise, if a new business process is to be introduced in the to-be state and the supporting applications must be developed before the
business process can be activated. Additionally, individual model element lifecycles that
might be affected, for example, by vendor support must be respected and integrated.
Scheduling the development activities and establishing the transformation model can be
supported by planning techniques from the field of project management. Gantt or milestone charts are commonly used to represent schedules. However, these techniques lack
in showing interdependencies between activities and events, which is provided by network techniques [Ke09]. Among such techniques to construct project network diagrams,
the Project Management Institute mentions the Precedence Diagramming Method and
the Arrow Diagramming Method that are able to display temporal dependencies between
individual activities [PMI00]. Based on project network models, formal analyses can be
applied to determine actual dates for project plans. A common technique is the Critical
Path Method that calculates deterministic early and late start and end dates for each activity, based on an activity-on-node network [AW90].
3.2 Bottom-up planned evolutionary change
In classic software development requirements engineering is understood as a process to
create and maintain a high-level requirements document for a system (R1). According to
Sommerville [So01] any requirements engineering process is comprised of four activities. Feasibility studies are conducted against the background of an initial set of requirements to answer whether the system under consideration contributes to the overall objective of the organization and whether a system can be implemented within the current
technological and economic constraints. Based on the answers to the preceding questions
system requirements are subsequently (requirements elicitation and analysis) discovered
in cooperation with the stakeholders. These requirements are classified and organized
into groups of related and mutually coherent requirements. Preceding their documentation in a requirements specification document, the requirements are prioritized and negotiated upon. During requirements validation different kinds of checks are performed
using multiple techniques as systematic or formal review. Thereby, checks for internal
validity (consistency), external validity (completeness and realism), and quality (verifiability) are performed on the requirements specification.
28
Proceeding from one-time requirements engineering in a software project to requirements-related processes for a long-living software solution, one comes to the discipline
of requirements management [So01]. In the single-system environment of classical software engineering the corresponding change management process boils down to three
essential activities of identifying a change proposal, analyzing the proposed change with
respect to costs and feasibility, and implementing the change. The IT Infrastructure Library (ITIL) [OGC00] furthers this basic understanding of change management by introducing additional steps for authorizing change, which form a counterpart of the feasibility analysis discussed by [So01]. Further, ITIL adds a post implementation review,
which in line with classic software engineering may be understood as revising the requirements specification.
Release management is primarily understood as the process of making software available to its users. What at first sight seems to be an easy task is aggravated by the complex
interdependencies of modern component-based software. [HW03] discuss five requirements that release management has to satisfy: 1) make component dependencies explicit,
2) keep releases consistent with respect to these dependencies, 3) control user access to a
release, 4) support development team transparently, and 5) store history of releases and
their usages. Key concept of release management is component metadata, i.e. specifications on dependencies between components, but also on dates of release and of withdrawal.
After an internally and externally validated set of requirements has been converted to an
agreed project portfolio (R4), synchronization management aims at the interdependencies between the changed projects (R9). [ES10] describes a method for identifying such
dependencies exemplified along the following three types of dependencies:
1) organizational dependencies, resulting from multiple projects targeting the same part
of the organization at the same time, 2) interface dependencies, resulting from the need
to migrate applications interfaces simultaneously with changing the related applications,
and 3) technical dependencies, resulting from infrastructure dependencies of applications, e.g. the compatibility with only distinct operating systems.
4 An integrated method for EA planning
The artifact of the paper at hand is an integrated method [MS95] for EA planning taking
global top-down driven transformation plans as well as local bottom-up driven evolution
of partial architectures into account.
A method consists of a sequence of design activities which are related to a procedure
model. The produced design results of the design activities are represented by an information model. Additionally, a method consists of roles which describe the participants
involved in the design activities. At the same time the inclusion of roles introduces various perspectives different stakeholders may have on design activities. Instructions on
how the design results are produced and documented are provided by adequate techniques [Br05], [Gu94].
29
Our analysis of related work shows that all of the relevant requirements (Section 2) but
(R6) are addressed by preliminary or related work described in Section 3. Our analysis
also shows that general planning cycles for global EA transformation plans on the one
hand side and more local requirements, release, and synchronization management on the
other hand side are similar. While the granularity of their planning objects differs (and
thus the approaches formality) their activities and techniques are similar. The main
challenges for an integrated planning approach are its organization and coordination as
well as its consistent description in an information model (R6). Therefore, this contribution focuses on the coordination of design activities (Subsection 4.1) and on an integrated information model (Subsection 4.2).
4.1 Coordination mechanisms
Deriving consistent and integrated plans for EA transformation and evolution is a complex task because
!
a number of subsystems (represented e.g. by architectural layers) need to be coordinated and the subsystems are again complex,
!
a number of stakeholders responsible for one or more subsystems need to be coordinated, thus the task can only be solved by division of labor,
!
neither the (sub) system(s) nor its environment(s) can be expected to remain unchanged during a planning cycle and a plans implementation.
It is unlikely to develop a consistent and integrated plan for EA transformation and evolution in a single iteration by central planning. Instead several iterations of centralized
and distributed planning will be necessary.
In cybernetics this situation is modeled as a closed-loop control system comprised of
goal setting, (meta-)planning, controlling, running, and measuring. The measurement
results are fed back to the controller who then adjusts the systems input. If engineered
properly closed-loop control systems achieve defined goals even in a dynamic environment. In case a system consists of several other systems it can be modeled as a system of
intermeshed closed-loop control systems (Figure 3) [HG94]. In such a situation running
system A is equivalent to (meta-)planning and controlling of system B.
goals
(meta) planning and
controlling system A
input of to-be values
running and
measuring system A = input of to-be values
(meta) planning and
feedback of as-is values
controlling system B
running and
measuring system B
feedback of as-is values
input
Figure 3: Closed-loop control systems based on [HG94]
30
output
Systems of intermeshed closed-loop control systems are guided and controlled by
1) overall goals and 2) contracts describing the interfaces between each closed-loop
control system. Overall goals are defined by business strategy, IT strategy, and EA strategy in particular. The structure of interfaces is defined by the governance structure relevant for the participants of the planning cycle. Legitimacy of each partial plan as well as
the integrated solution can be evaluated based on existing EA principles. Such systems
can be vertically or horizontally intermeshed. Vertically intermeshed systems can be
found in hierarchical relations, e.g. overall application planning is vertically intermeshed
with application planning for divisions, business units, regions, products etc. Horizontally intermeshed systems can be found with peer-level planning, e.g. application planning
is horizontally intermeshed with IT infrastructure planning.
As the examples above illustrate, the general approach to design an integrated EA planning function as a system of intermeshed closed-loop control systems allows for a variety of configurations. According to the Conant-Ashby theorem every good regulator of a
system must be a model of that system [CA70] which means that the effectiveness of
the individual configuration depends on how well it represents the actual interdependencies of the planning objects. The following information model ensures that the theorem
is fulfilled. In Section 5 we will discuss an example configuration.
4.2 Information model
Figure 4 shows an information model that is capable of describing the transformation
model for EA planning in an embracing manner.
Figure 4: Information model for describing the EA planning transformation model
Central concept of the information model is the mixin PROJECTAFFECTABLE which
marks a concept from an EA information model as subject to EA planning, thereby allowing a using organization to decide which concepts, e.g. business applications, are
considered in EA planning.
31
Beyond the basic linkage between the concept REQUIREMENT and the concept
WORKPACKAGE in Figure 2, the comprehensive information model introduces a different conception. Firstly, a requirement reflects an intended change in contrast to a
work package that describes a scheduled change. Secondly, complementing the relationships introduces and retires that target changes on an EA-relevant level, the requirement
provides the relationship optimizes. This relationship allows describing maintenance
requirements that do not reshape the enterprise on an EA-relevant level. Together, the
three types of relationships and the relationship precedence cover all types of dependencies between architecture elements in a uniform manner (see requirements (R3) and
(R6)). In particular, they supply a basis for a technique for comparing an as-is and a tobe state (T1). Further, each requirement bears a target date, i.e. a date as of which the
change is to be effective, as well as an expected duration for implementation. Based on
this information, the technique of a critical path analysis (T5) can be implemented.
Above information not only supplies concepts for documenting requirements, but also
for documenting decisions on requirements. A requirement can on the one hand switch
to the role APPROVEDREQUIREMENT, where it is further linked to a work package
for implementation. On the other hand, a requirement may also be converted to a POSTPONEDREQUIREMENT to reflect that currently no plan for resolving the identified
conflicts can be scheduled. Thereby, techniques for requirements management (T2) as
well as techniques for synchronization management (T3) are supported via the temporal
and explicit dependencies between the requirements. Table 1 summarizes design results,
information model, design activities, and techniques of our method for an integrated EA
planning.
Table 1: Method summary
Method for planning enterprise transformation
Design Results
DR1. Meta result [or input]: System of local and global partial plans (R1)
DR2. Meta result [or input]: Coordination mechanisms (R6)
DR3. Meta result [or input]: Principles and goal structures (R6)
DR4. List of model elements that need to be changed in each partial plan
(ProjectAffectables) (R3)
DR5. Information about relationships to other model elements and other
partial plans (requirement relations) (F)
DR6. Description of project proposals containing work packages realizing
approved requirements (R4, R9)
Information
Model
Design Activities
I1. Transformation Model (R3, R6; provides basis for T1-3, T5)
A1. Analyze differences between as-is and to-be (requirements) for each
local and global planning iteration (DR4)
A2. Synchronize partial plans and identify inconsistencies (DR5)
A3. Separate segments, i.e. identify work packages and projects (DR6)
For each segment:
A4. Identify general temporal interdependencies (DR6)
A5. Schedule an effective sequence of development activities, by refining
temporal interdependencies (DR6)
A6. Draw network plan displaying all interdependencies (DR6)
32
Techniques
T1. Graph comparison (for A1)
T2. Techniques of requirements management (for A1)
T3. Techniques of synchronization management (for A2, A3)
T4. Establishing project network model/graph using Precedence Diagramming Method and the Arrow Diagramming Method (for A5)
T5. Deriving a project timeline using Critical Path Method (for A5)
5 Demonstration
In this section, we demonstrate (and initially evaluate) our approach employing an industry case illustrating one possible configuration of a system of intermeshed closed-loop
control systems as proposed in this paper. The using company provides banking solutions, offering a banking platform to private and universal banks. The companys EAM
focuses on three main fields: Application development, application management, and
operations. The application development division is responsible for the development of
the banking platform which is described by an application architecture model, an IT
architecture model, and an operating model.
Challenges within the architectural development plan are the coordination of activities of
development teams (R6), assurance that all dependencies are addressed, and that milestones of integration and development activities are met simultaneously (R9). If, for
example, a component of an application needs an interface to a component of another
application (R3) at a certain time for a certain milestone (e.g. test or release), it has to be
assured that both components are available at that very point in time. This simple example grows very complex as the banking platform comprises over 200 applications, each
consisting of a multitude of components that each has its own lifecycle.
The planning activities at the company aim at the further development of their core
product, a banking platform, which consists of various applications, interfaces and middleware components. As the company also provides the hosting of the platform, infrastructure is an issue, too. For this perspective, the companys information model covers
the relationships between business applications, their constituting components and the
underlying hardware devices (see Figure 5). The three conceptual classes representing
architecture elements are thereby considered project affectable, i.e. are subject to the
dependence planning mechanisms described in our method.
Figure 5: Information model of the demonstration case
In such a scenario both, top-down transformation plans and the bottom-up evolution of
partial architectures occur and need to be integrated (R6). Further development of the
banking platform (high-level transformation plan) is driven by a superordinate planning
team that defines the general development strategy of the platform and the relevant principles (DR1, DR2, DR3). At the same time requirements for the development of the
individual components and sub-architectures are collected separately. The configuration
33
of the underlying intermeshed closed-loop control system is a matrix structure defined
by lead architects (business, applications, IT, operations, and security) and domain architects responsible for business domains structured by business products (DR1, DR2).
Partial architecture roadmaps are defined following the general EA guidelines. The
roadmaps consist of models representing snapshots of the desired architecture for up to
three future points in time, taking into account existing vendor specific constraints if
applicable (DR4). Individual component development plans are then integrated into
these roadmaps. The coordination between high-level transformation plans and local
roadmaps of evolution is ensured by the matrix structure. While domain architects control the development and realization of high-level transformation plans, lead architects
drive and ensure the development of components taking detailed restrictions into account. In joint workshops all parties consolidate their plans based on EA goals and principles. Possible alternatives are thereby discussed and evaluated (DR5). This particular
matrix configuration uses vertical and horizontal coordination of the meshed systems.
Vertical coordination occurs e.g. by aligning a domain specific transformation plan with
the detailed plans for applications of that domain. Horizontal coordination occurs e.g. by
aligning the partial architectures of the lead architects.
To determine a sequence of development activities general milestones and components
to be developed are defined. Afterwards, the development phases of the elements
lifecycles are planned in detail, i.e. in terms of specification and testing phases. Upon
those specifications a rough project program schedule can be defined (DR6). Finally, on
the basis of the project outlines actual project planning is enabled. For that purpose, the
project proposals can be enriched with information about costs, quality metrics, staff,
risks, and resources.
The case description illustrates that an appropriate configuration of a system of intermeshed closed-loop control systems that takes the concerns of relevant stakeholders
(domain and lead architects) into account is a efficient solution for coordination of topdown transformation planning and bottom-up evolutionary change. Whether the results
of such a planning approach justify the respective efforts cannot be shown here. Therefore, the case study does not serve as a full evaluation. However, discussions with the
case study company showed that apart from a well defined plan such an approach has
a number of positive side effects most notably stakeholders have a common high level
of knowledge about the organization and are therefore able to effectively coordinate each
other before or during actual projects.
6 Reflection and outlook
In this paper, we analyzed top-down driven methods for EA transformation planning as
well as bottom-up driven methods for requirements, release, and synchronization management. We integrated design activities and techniques used in these domains into a
method for an integrated EA planning. Our method description focuses on the coordination of distributed planning activities and an underlying information model, which provides the modeling concepts for addressing requirements (R1), (R3), and (R6).
34
Building on the contained information about the transformation and evolution plans, the
delineated method and techniques particularly aim at requirements (R4) and (R9). The
applicability of method and information models has finally been demonstrated with an
industry case.
However, there still are a number of questions that we have not answered yet but given
an indication in the case only. For example our approach does not generally describe
how exactly configuration mechanisms for systems of intermeshed closed-loop control
systems work and whether there may be a (limited) number of reference configurations
for an integrated EA planning
Also criteria for determining the number of iterations for global and local planning cycles are unknown. It is further unknown, which exit conditions are valid, i.e. when to
stop the EA planning phase and proceed with refinements and implementations.
Finally, it has to be shown that there is a business case for such a coordinated approach of an integrated EA planning opposed to an even more decentralized and less
coordinated development plan. However, in practice we observe a growing demand and
acceptance for such a coordinated planning. This is because alternatively a coordination
of projects applying for funding is often performed by prioritization and projects are
postponed after available budgets are assigned. The prioritization process, taking only
the financial dimension into account, however, often results in dysfunctional plans.
Bibliography
[AG10a] Aier, S.; Gleichauf, B.: Applying design research artifacts for building design research
artifacts: A process model for enterprise architecture planning. In (Winter, R.; Zhao, J.L.;
Aier, S.): (eds.) Global Perspectives on Design Science Research, 5th International Conference, DESRIST 2010, Springer, 2010; pp. 333-348.
[AG10b] Aier, S.; Gleichauf, B: Towards a systematic approach for capturing dynamic transformation in enterprise models. In HICSS, 2010; pp. 1-10.
[AW90] Antill, J.; Woodhead, R.: Critical path methods in construction practice. WileyInterscience, 1990.
[Br05] Braun, C.; Wortmann, F.; Hafner, M.; Winter, R.: Method construction - a core approach
to organizational engineering. In SAC, 2005; pp. 1295-1299.
[Bu09] Buckl, S.; Ernst, A.M.; Matthes, F.; Schweda, C.M.: An information model for managed
application landscape evolution. Journal of Enterprise Architecture (JEA) 5(1), 2009; pp.
12-26.
[CA70] Conant, R.C.; Ashby, W.R.: Every good regulator of a system must be a model of that
system. Intl. J. Systems Science, 1970; pp. 89-97.
[CEF99] Carlson, A.; Estepp, S.; Fowler, M.: Temporal patterns. In Pattern Languages of Program
Design. Addison Wesley, Boston, MA, USA, 1999.
[ES10] Ernst, A.M.; Schneider, A.W.: Roadmaps for enterprise architecture evolution. In Engels, G.; Luckey, M.; Pretschner, A.; Reussner, R.: (eds.) 2nd EuropeanWorkshop on
Patterns for Enterprise Architecture Management (PEAM2010) Paderborn, Germany,
2010; pp. 253-266.
[Gu94] Gutzwiller, T.A.: Das CC RIM-Referenzmodell für den Entwurf von betrieblichen,
transaktionsorientierten Informationssystemen. Ph.D. thesis, Universität St.Gallen, 1994.
35
[HG94] Hahn, D.; Grünewald, H.: PuK, Controllingkonzepte: Planung und Kontrolle, Planungsund Kontrollsysteme, Planungs-und Kontrollrechnung. Gabler, 1994.
[HV93] Henderson, J.C.; Venkatraman, N.: Strategic alignment: Leveraging information technology for transforming organizations. IBM Systems Journal 32(1), 1993; pp. 472-484.
[HW03] van der Hoek, A.; Wolf, A.L.: Software release management for component-based software. SOFTWARE-PRACTICE AND EXPERIENCE 33(1), 2003; pp. 77-98.
[IOS07] International Organization for Standardization.: ISO/IEC 42010:2007 Systems and software engineering - Recommended practice for architectural description of softwareintensive systems, 2007.
[Ke09] Kerzner, H.: Project management: a systems approach to planning, scheduling, and
controlling. Wiley, 2009.
[LLO93] Luftman, J.N.; Lewis, P.R.; Oldach, S.H.: Transforming the enterprise: The alignment of
business and information technology strategies. IBM Systems Journal 32(1), 1993; pp.
198-221.
[MS95] March, S.; Smith, G.: Design and natural science research on information technology.
Decision Support Systems 15(4), 1995; pp. 251-266.
[Ma08] Matthes, F.; Buckl, S.; Leitel, J.; Schweda, C.M.: Enterprise Architecture Management
Tool Survey 2008. Chair for Informatics 19 (sebis), Technische Universität München,
Munich, Germany, 2008.
[OGC00] Office of Government Commerce (OGC).: ITIL - Service Delivery. IT Infrastructure
Library (ITIL), The Stationery Office, Norwich, UK, 2000.
[PMI00] Project Management Institute.: A Guide to the Project Management Body of Knowledge,
2000.
[Pe07] Peffers, K.; Tuunanen, T.; Rothenberger, M.; Chatterjee, S.: A design science research
methodology for information systems research. Journal of Management Information Systems 24(3), 2007; pp. 45-77.
[Pu06] Pulkkinen, M.: Systemic management of architectural decisions in enterprise architecture
planning. four dimensions and three abstraction levels. In: Proceedings of the 39th Annual Hawaii International Conference on System Sciences, 2006; pp. 179c.
[RWR06]Ross, J.W.; Weill, P.; Robertson, D.C.: Enterprise Architecture as Strategy. Harvard
Business School Press, Boston, MA, USA, 2006.
[So01] Sommerville, I.: Software Engineering. Pearson Studium, Munich, Germany, 6th edn.,
2001.
[Wa05] Wagter, R.; van den Berg, M.; Luijpers, J.; van Steenbergen, M.: Dynamic Enterprise
Architecture: How to Make IT Work. John Wiley, 2005.
[Wi08] Winter, R.: Business engineering - betriebswirtschaftliche konstruktionslehre und ihre
anwendung in der informationslogistik. In: Integrierte Informationslogistik. Springer,
Berlin, Heidelberg, Germany, 2008; pp. 17-38.
36
Applying Soft Systems Methodology in Enterprise
Architecture Creation Workshops
Agnes Nakakawa1 , Patrick van Bommel 1 , H.A. Erik Proper1,2
1
ICIS, Radboud University Nijmegen
P.O. BOX 9010, 6500 GL Nijmegen, The Netherlands.
2
CRP Henri Tudor, L-1855 Luxembourg-Kirchberg, Luxembourg.
A.Nakakawa@science.ru.nl, pvb@cs.ru.nl, Erik.Proper@tudor.lu
Abstract: Lack of effective involvement of stakeholders is one of the main drawbacks
of enterprise architecture initiatives. Ongoing attempts to overcome this involve using
Collaboration Engineering to develop a collaboration process that enterprise architects
can execute to facilitate collaborative sessions with stakeholders during architecture
creation. However, a field study evaluation of this process revealed that it offered inadequate support for stirring vigorous and rigorous discussions during activities that
required organizing and assessing problem or solution aspects that resulted from brainstorming activities. Since Soft Systems Methodology (SSM) helps to structure rational
thinking about messy situations, its techniques can be adapted to supplement the design of the collaboration process with support for triggering discussions and creating
a shared understanding and vision among stakeholders. This paper therefore presents
a script that shows how this can be done, and discusses its evaluation in a real case.
1 Introduction
The common drawbacks of enterprise architecture development include the lack of: effective involvement of stakeholders; support and commitment from top management; effective communication; and a mutual understanding of the purpose of the architecture effort
[Ben05, Gar09, Sta95]. Enterprise architects have reported various challenges that hinder
effective stakeholder involvement during architecture creation (see [NBP10]). The challenges are mainly caused by two issues. First, the success of collaborative sessions that
involve enterprise architects and stakeholders mainly depends on the presence of a professional (or skilled) facilitator. Second, there is lack of a clear, predictable, and repeatable
way of managing tasks that require effective stakeholder involvement. Ongoing attempts
to overcome these issues involve using Collaboration Engineering to develop a collaborative process that enterprise architects can execute (by themselves) so as to manage
tasks that require effective collaboration with stakeholders during enterprise architecture
creation. Collaboration Engineering was chosen because it offers affordable facilitation
to practitioners (in this case enterprise architects) of recurring high-value tasks (like enterprise architecture creation), by enabling the development of repeatable processes that
37
practitioners can execute without hiring a professional facilitator [BVN03, VB05].
According to [KV07, VB05], a collaborative process for a given task is designed using
the following procedure: specifying the goal and deliverables of the process; defining the
activities that participants must execute so as to achieve the goal; specifying the reasoning
phases participants must undergo in order to achieve the goal; and describing detailed facilitation support for each activity. Facilitation support is specified by articulating: (a) the
Group Support System (GSS) tools that should be used (or alternative techniques) during
the collaborative sessions; (b) how the tools should be configured; and (c) the message
prompts that should be followed [BVN03]. This design approach was adapted when formulating the collaboration process for effectively involving stakeholders during enterprise
architecture creation. This process is herein referred to as Collaborative Evaluation of Enterprise Architecture Design Alternatives (CEADA). The earlier version of CEADA was
evaluated in a field study (of five organizations) where it was effective in supporting activities that required stakeholders to brainstorm, prioritize or rank or rate concerns and
requirements for the architecture; and perform multi-criteria evaluation of possible enterprise architecture design alternatives. However, CEADA was still lacking adequate support for stirring vigorous and rigorous discussions when executing activities that required
stakeholders and architects to reduce and organize aspects from brainstorming activities;
and assess possible interrelationships and implications. This was reflected in the feedback
(see section 2) from stakeholders who participated in the sessions supported by CEADA;
the facilitator; and the observer of the sessions.
Since the main focus of this research is to offer effective stakeholder involvement in architecture creation, in this paper we propose to address the above weakness by supplementing
CEADA with techniques for enhancing the creation of a shared understanding and vision
during execution of activities that involve organizing and discussing brainstormed aspects.
We focus on adapting Soft Systems Methodology (SSM) because of its reputation for
managing complex and ill-structured organizational problems through structuring rational
thinking about them [Che98]. We also adapt the cause-effect analysis diagram (or Fishbone or Ishikawa) technique because of its support for thorough problem analysis [Ish86].
Since SSM offers implicit facilitation support for collaborative workshops or discussion
debates among problem owners and solver(s), Collaboration Engineering is further used in
designing the facilitation script that shows how SSM and Ishikawa techniques can be used
in enterprise architecture creation. For the remaining part of this paper, Section 2 presents
the research methodology; Section 3 discusses the script that extends CEADA to address
its weaknesses; Section 4 presents the evaluation results of CEADA’s performance in a
real organization; and Section 5 concludes the paper.
2 Research Methodology
This research adapts the Design Science methodology. In Design Science, significant organizational problems are solved by: designing suitable artifacts using knowledge derived
from scientific theories and industrial experiences; evaluating those artifacts using experiment and real life environments; and refining the artifacts to address their inefficiencies
38
[BVN03]. In this research CEADA is the ultimate artifact, which was designed based
on literature from the enterprise architecture development realm, collaborative problem
solving and decision making realm, and expert advice from enterprise architects and professional facilitators. In our earlier work[NBP11], a field study evaluation of CEADA
indicated that it is already successful in effectively supporting communication during the
execution of tasks that involve brainstorming and evaluating problem and solution aspects.
However, evaluation feedback also indicated the need to enhance CEADA’s support for
tasks that require stakeholders to filter, clarify, organize, and discuss problem and solution
aspects that arise from a brainstorming activity. This is implied by the following three
issues highlighted in the feedback from the field study.
First, more time in the sessions was spent seeking a shared understanding and shared vision of problem and solution aspects, and yet some stakeholders complained (at the end of
the sessions) that they had less discussion time. Second, stakeholders suggested that using
(simple) visual representations of their ideas (prior to discussing and evaluating the architecture models) would have helped them to better analyze and understand the problem
and solution aspects and the architecture models. Third, the participation of stakeholders
decreased during activities that involved organizing and discussing problem and solution
aspects. This could have been caused by the less hands-on nature of the way the organize tasks were executed. These three issues indicate the need to devise a way that will
improve the execution of CEADA activities that require explicit articulation, proper organization, and thorough discussion of the problem and solution aspects. Such activities
require stakeholders and architects to undergo the reduce, clarify, and organize patterns
of reasoning 1 . However, Collaboration Engineering research has widely addressed issues
encountered in the other patterns of reasoning (i.e. generate, evaluate, and build consensus), but more research is needed in addressing issues encountered in the reduce, clarify,
and organize patterns of reasoning such that the effectiveness of collaboration processes
can be improved [VB05].
Therefore, this paper reports work on an extended version of CEADA that attempts to
address the above issues by triggering purposeful discussions that help stakeholders to
acquire a shared understanding of problem and solution aspects during enterprise architecture creation workshops. The extended CEADA comprises of a script designed by
adapting knowledge from SSM, the Ishikawa diagram technique, and Collaboration Engineering. Thereby, the weaknesses of any of these methods in effectively supporting
execution of activities in architecture creation workshops, are overcome by the strengths
of the other methods. The extended CEADA has been evaluated in a real organization and
findings are discussed in section 4.
1 According to [BVN03, BVN+ 06], there are six patterns of reasoning that a group may undergo, i.e.: generate, moving from a state of having few to more concepts through brainstorming; reduce, moving from a state
of having many to few concepts that are worthy of further attention; clarify, moving from a state of having less
to more shared understanding and meaning of concepts; organize, moving from a state of having less to more
understanding of the relationships among concepts under consideration; evaluate, moving from a state of having
less to more understanding of the value of concepts under consideration; and build consensus, moving from a
state of having fewer to more group members agree and commit to proposed concepts.
39
3 Soft Systems Methodology in Architecture Creation
This section discusses the adaptation of SSM and Ishikawa diagram techniques to improve
CEADA. In addition, the usability of the CEADA script has been demonstrated using scenarios from one of the organizations in which CEADA was earlier evaluated, i.e. Wakiso
District Local Government (WDLG). WDLG is an administrative and service delivery organization that operates under the Ministry of Local Government in Uganda; and consists
of 11 departments, each having a number of subsections [Wak08].
Baseline (as-is)
enterprise
architecture
models
Use to formulate
- Rich Picture
- Analysis One Two Three
Represent
in form of
- Root definitions
- CATWOE analysis
- Conceptual models
Use to formulate
Represent in
form of
1. Study the
organizations
problem situation
4. Take action to
improve problem
situation
2. Formulate purposeful
activity models describing
the desired situation
Target (to-be)
enterprise
architecture models
3. Compare models
with real world views
Debate about desirable
& feasible changes
Use to refine
Find accommodations
between conflicting interests
Figure 1: Adaptation of SSM in Enterprise Architecture Creation (Based on [Che98])
According to [Che98], SSM mainly comprises of the following four stages. Stage one
involves studying the problem situation and representing it using: (a) Rich Picture – a
holistic expression of aspects that describe the problem situation e.g. processes, actors
and their thoughts or concerns; and (b) Analysis One Two Three – a description of all
problem owners, cultural or social aspects (i.e. roles or norms or values), and political aspects of the problem situation. Stage two involves describing the desired situation by: (a)
Formulating Root Definitions – explicit phrases that briefly describe the desired transformation in form of “what should be done”, “how it should be done”, and “why it should be
done”; (b) Performing a CATWOE analysis of each Root Definition – assessing Customers
who will be affected by the transformation, Actors who will implement the transformation, Transformation process(es) that will be affected, World views regarding the transformation, Owner(s) of the transformation, and Environmental and external issues that
will affect the transformation; (c) Using the Root Definitions and their CATWOE analysis
to assemble all proposed transformation processes into conceptual or purposeful activity
models that describe the desired situation. Stage three involves comparing the conceptual
models with real world views through holding debates which aim at: (a) defining desirable
and feasible changes for improving the problem situation; (b) finding accommodations between conflicting interests regarding the problem situation and the desired changes. Stage
four involves taking appropriate action to address the problem situation.
The adaptation of the SSM techniques defined in stages one to three into enterprise architecture creation workshops is shown in Figure 1. From figure 1, Rich Picture and Analysis
One Two Three can be adapted as techniques for gathering information for formulating
baseline architecture models (see section 3.1), whereas Root Definitions, CATWOE anal-
40
1. Communicate purpose of session 1
Tool Name: Manual Step Name: [Activity 1.0]
Facilitator Notes: Purpose of session is to gather information on the organizations problem & solution aspects
2. Define services offered by, & processes executed in, the organization
Tool Name: Generate Step Name: [Activity 1.1.1] Outline File Name: GenOrgnProcesses.mw
Step Specific Information: Allow unlimited No. of items per topic; viewing other participant input; & participants
to control topic selection. This configuration applies to all activities supported by the generate tool in this script.
Topic List: (1) What are the services offered by, & the processes executed in, this organization? (2) What are
the organizations information inflows? (3) What are the organizations information outflows? (4) Who are the
external partners of this organization (e.g. unions, consortiums, sponsors, etc), & which departments do they
collaborate with? (5) What are the completed & ongoing projects in the organization?
Facilitator Notes: Explain the data capturing functions (or answering formats) for each question to participants,
i.e.: (1) {Department}-{services it offers}-{Processes it executes to offer its services}; (2) {type of information
inflow}-{information sender}; (3) {type of information outflow}-{information recipient}; (4) {Department}-{External
partner(s)}; (5) {project name}-{project status}-{where to find documentation about the project}
3. Clarify & organize attributes of processes in the organization
Tool Name: Organize Step Name: [Activity 1.1.1.1] Output File Name: OgzdProcesses.mw
Step Specific Information: Input file - [GenOrgnProcesses.mw]; Step will run in Outliner Mode.
Facilitator Notes: Display the generated list of processes & invoke a specialization-driven-division (i.e. divide
stakeholders into small groups based on their departments/specialization). This enables detailed assessment of
the as-is situation & enhances communication, shared understanding, & homogeneity in a subgroup. Guide
each subgroup to do the following:
(1) Clean & clarify items in the processes list under their department
(2) Distribute copies of the Ishikawa diagram template for capturing process attributes to all subgroups.
Guide them to fill in the required prompts in the template, using the cleaned list of processes in their department.
(3) Display key symbols you have chosen for drawing a Rich Picture. Guide each subgroup to draw a Rich
Picture of the as-is situation of its department, showing the actors, their roles, their concerns etc.
(4) Converge subgroups & encourage representatives of each subgroup to discuss their department-specific
Ishikawa diagrams for processes; & department-specific Rich Picture. To help the group to acquire a shared
understanding of the as-is situation, map all the department-specific Ishikawa diagrams for processes onto an
organization-wide Ishikawa diagram for processes; & merge all the department-specific Rich Pictures to form an
organization-wide Rich Picture. Plug these diagrams on to the holistic data capture template.
4. Define the challenges or problems faced in the organization
Tool Name: Generate Step Name: [Activity 1.1.2] Outline File Name: GenOrgnChallenges.mw
Topic List: What are the challenges or problems faced in this organization?
Facilitator Notes: Explain the answering format: {Department}-{challenges faced by the department}
5. Determine the root problem cause
Tool Name: Organize Step Name: [Activity 1.1.3] Output File Name: OrgnRootProblemCauses.mw
Step Specific Information: Input file - [GenOrgnChallenges.mw]; Step will run in Outliner Mode.
Facilitator Notes: Display the generated list of challenges, & invoke a specialization-driven-division (this is
explained in activity 1.1.1.1). Guide each subgroup to do the following:
(1) Clean & clarify items in the challenges list under their department
(2) Distribute copies of the Ishikawa diagram template for problem analysis to all subgroups. Guide them to
perform a cause-effect analysis of problems in their department, by filling the required prompts in the template.
The department-specific Rich Pictures from activity 1.1.1.1 can be used as a source of information.
(3) Converge subgroups & encourage representatives of each to discuss their department-specific Ishikawa
diagrams for problem analysis & the identified root problem cause(s). Clean & organize the challenges list using
output from the subgroups. Map all the department-specific Ishikawa diagrams for problem analysis onto an
organization-wide Ishikawa diagram for problem analysis, & identify the general root problem cause(s).
6. Identify un-captured problem aspects & any anomalies in the output of problem analysis
Tool Name: Generate Step Name: [Activity 1.1.4] Outline File Name: CommentsOnProblemAnalysis.mw
Topic List: (1) What concerns do you have about any of the Ishikawa diagrams for problem analysis? (2) Which
effects & causes of problems (or organizational weaknesses & threats) have not been represented?
Facilitator Notes: Explain answering format: {Department}-{cause/effect/weakness/threat}
Figure 2: Facilitation Script for adapting SSM in Architecture Creation
ysis, and activity models can be adapted as techniques for gathering information for formulating target architecture models (see section 3.2).
3.1
Creating Baseline Enterprise Architecture Models
SSM is a structured inquiry process which uses models as a source of questions about a
problem situation [Che98]. This section presents five structured and question-triggering
techniques that can be used in interviews and group sessions (or when reviewing organization documentation) to gather information for creating baseline architecture models.
These include: (a) the Ishikawa diagram, a technique useful for assessing problems and
their causes [Ish86]; (b) Rich Picture, and Analysis One Two Three [Che98] (these tech-
41
7. Assess Amendments to the problem analysis outcome
Tool Name: Organize Step Name: [Activity 1.1.5] Output File Name: AdditionsProblemAnalysis.mw
Step Specific Information: Input file - [CommentsOnProblemAnalysis.mw]; Step will run in Outliner Mode.
8. Seek consensus on outcomes of problem analysis
Tool Name: Evaluate Step Name: [Activity 1.1.6] Output File Name: ConsensusProblemAnalysis.mw
Step Specific Information: Input Topic List Items - [Yes, I agree with the problem analysis results; No, I do not
agree with the problem analysis results]; Evaluation Method - [Vote (Yes/No)]; Display Variability - [Yes].
Facilitator Notes: Encourage participants who do not agree to attach comments on their votes. Discuss the
comments & re-vote to enable participants to reach a realistic level of shared understanding of problem aspects
9. Define aspects that describe the organizations problem scope
Tool Name: Generate Step Name: [Activity 1.1.7] Outline File Name: GenAspectsProblemScope.mw
Topic List: (1) Which processes or stakeholders (i.e. current problem owners) are affected by the problems? (2)
Which processes or stakeholders (i.e. possible problem owners) are likely to be affected by the problems?
Facilitator Notes: The filled Ishikawa diagram for problem analysis can be used as a source of information.
10. Clarify & organize the generated aspects on the problem scope
Tool Name: Organize Step Name: [Activity 1.1.8] Output File Name: OgzdAspectsProblemScope.mw
Step Specific Information: Input file - [GenAspectsProblemScope.mw]; Step will run in Outliner Mode.
11. Evaluate & agree on aspects that define the organizations problem scope
Tool Name: Evaluate Step Name: [Activity 1.1.9] Output File Name: AgreedProblemScope.mw
Step Specific Information: Input topic file - [OgzdAspectsProblemScope.mw]; Evaluation Method - [Select
(Mark all that apply)]; Display Variability - [Yes]
12. Reaffirm key organization principles associated with problem aspects
Tool Name: Generate Step Name: [Activity 1.2.0] Outline File Name: AssociatedOrgnPrinciples.mw
Topic List: (1) Which business (& architecture principles, if they exist) are associated with the problems the
organization is facing or with any possible solutions to the problems? (2) What are the accepted practices in the
organisation? (3) What are the existing management or governance frameworks in this organisation?
Facilitator Notes: answering format is: {name of principle/practice/framework} - {source of information about it}
13. Specify business strategy & goals
Tool Name: Generate Step Name: [Activity 1.3.0] Outline File Name: GenBusinessStrategyGoals.mw
Topic List: Which business strategy & goals does the organization have so as to overcome its problems?
Facilitator Notes: Explain the answering format: {strategy: goals}. If the business strategy & goals are already
documented, provide a seed file listing them & prompt participants to give their views about the strategy & goals.
14. Clarify & validate the business strategy & goals
Tool Name: Organize Step Name: [Activity 1.3.1] Output File Name: OgzdBusinessStrategyGoals.mw
Step Specific Information: Input file - [GenBusinessStrategyGoals.mw]; Step will run in Outliner Mode.
Facilitator Notes: Categorize business goals according to the business strategy that are intended to achieve.
Plug results into the right section in the holistic data capture pyramid or template.
15. Evaluate & agree on the business strategy & goals
Tool Name: Evaluate Step Name: [Activity 1.3.2] Output File Name: RankBusinesStrategyGoals.mw
Step Specific Information: Input topic file - [OgzdBusinessStrategyGoals.mw]; Evaluation Method:[Rank from
1 to N (Use each value only once)]; Display Variability: [Yes]
Facilitator Notes: Prompt stakeholders to prioritize or rank the business goals.
16. Determine possible business solution alternatives for achieving the strategy & goals
Tool Name: Generate Step Name: [Activity 1.4.0] Outline File Name: GenBusinessSolnAlternatives.mw
Topic List: What can the organization do in order achieve its business strategy & goals?
17. Clarify & organize the generated business solution alternatives
Tool Name: Organize Step Name: [Activity 1.4.1] Output File Name: OgzdBusinessSolnAlternatives.mw
Step Specific Information: Input file - [GenBusinessSolutionAlternatives.mw]; Step will run in Outliner Mode.
Facilitator Notes: Guide stakeholders to: identify which business solution alternatives can be treated as
specifications of a given solution alternative; & assess each solution alternative by filling in the required prompts
in the Ishikawa diagram template for capturing constraints of a given business solution alternative.
Figure 3: Facilitation Script for adapting SSM in Architecture Creation (Contd.)
niques are defined in the preceding section); (c) short data capture statements or functions;
and (d) a holistic data capture pyramid template, which has been formulated based on
literature in [Lan05, OPW+ 08] to give an overview of all aspects to be addressed during enterprise architecture creation. The CEADA script in figures 2, 3, 10, and 11 shows
when and how to apply these techniques during group sessions with client stakeholders.
It has been formulated based on the procedure for designing collaborative processes (see
section 1), and it has been documented using MeetingworksTM GSS. The detailed procedure of how the activities in CEADA were derived was discussed in our earlier work (see
[NBP11]), the script herein extends CEADA by offering execution details of its activities.
In figure 2 the facilitation notes of activity 1.0 introduce the holistic data capture pyramid
template (see figure 4), which provides a shared holistic view and shared understanding of
data that is to be gathered during architecture creation and all the information templates
that are to be used. In addition, the facilitation notes in figure 2 (activities 1.1.1.1 and
1.1.3), figure 3 (activity 1.4.1), and figure 11 (activity 4.2) show how the Ishikawa dia-
42
Mission
Vision
Values
Strategy Principles
Culture
External
Goals
regulations
Ishikawa for
Constraints
Actions to achieve the goals (Ak)
Enterprise architecture
Ishikawa for
attributes of
processes in
as-is state
Ishikawa for
Requirements
Domain/unit
architectures for
to-be state
Domain/unit
architectures for
as-is state
Ishikawa for
attributes of
processes in
to-be state
Problem
analysis
Ishikawa
Current
Rich Pictures
Operations (processes, products, people,
or as-is
Desired
or to-be
Figure 4: Holistic Data Capture Pyramid Template (Based on [Lan05, OPW+ 08])
Beneficiaries
Outflows(Results)
Inflows (Inputs)
Organization or
department process
Internal Actors
External Actors
Figure 5: Adapted Ishikawa Diagram Template for Capturing Process Attributes
gram technique has been adapted in mainly four ways. These include: Ishikawa diagram
template for capturing attributes of processes executed in an organization (see figure 5);
Ishikawa diagram template for analyzing the problematic aspects in an organization (see
figure 6 for the problem analysis template that was used at WDLG); and Ishikawa diagram
templates for defining constraints and requirements that the architecture pertaining to a
given (business) solution alternative must fulfill (see figure 8). The facilitation notes in the
script also indicate that once these templates are populated with data, they can be plugged
into the holistic data capture pyramid template in figure 4. Furthermore, in figure 2 (activity 1.1.1.1), facilitator notes also introduce the use of a Rich Picture in capturing baseline
operations of departments in an organization. From figure 7, there are six key symbols
that can be used to formulate a Rich Picture. As shown in figure 7, these symbols were
used to formulate a Rich Picture for the as-is situation at WDLG. It shows that WDLG
involves various complexities of people (represented with a face-like symbol and accompanying text) e.g. departments, committees, community. The level of detail in a Rich
Picture depends on the problem solver [Che98]. Thus, to avoid an overcrowded Rich Picture and to enable shared understanding about the major problem(s) faced, the subsections
in each of the 11 departments have not been represented. The dotted lines represent twoway communication, and the storage symbol on a dotted line indicates a decentralized way
of data storage. The un-dotted lines represent responsibilities of a given group of people,
the stars represent the services offered, and the cloud call-out represents problems faced.
43
Cause of the
problem (CkPkSkDi)
Problem faced by the
department subsection (PkSkDi)
Department (Di (for i=1,2,
,11)
Department
subsection (SkDi)
WDLG
Figure 6: Adapted Ishikawa Diagram Template for Problem Analysis in WDLG
As the script indicates in figure 2 (activity 1.1.1.1), these symbols can be used to draw
department-specific Rich Pictures, which can later be merged into an organization-wide
Rich Picture.
3.2
Creating Target Enterprise Architecture Models
The script in figure 3 (activity 1.4.1) and figure 11 (activity 4.2) describe when and how
to use the adapted Ishikawa template for gathering data on constraints and requirements
that the architecture must address (see figure 8). In addition, in figure 10 (activity 1.6.1)
the script describes when and how to articulate the purpose of the architecture creation initiative using the architecture purpose template shown in figure 9. Furthermore, the script
in figure 11 (activity 4.1) shows when and how to use SSM’s Root Definitions and CATWOE analysis to gather information on requirements and solution scenarios that must be
represented in the target architecture models. For example, one of the Root Definitions
for WDLG can be denoted as (R), where: R = {What to do – enable centralized data
storage and IT-supported data processing and sharing in WDLG}; {How to do it – by
purchasing database servers and client computers, and setting up and maintaining a Local Area Network}; {Why do it – in order to have consistent data records, realtime data
retrieval, and improved service delivery in WDLG}. Consequently, the CATWOE analysis of R is as follows: Customers {WDLG staff, Wakiso district community}; Actors
{staff in the IT section of the planning department of WDLG, ICT and Internet service
providers}; Transformation process {input – isolated databases and manual departmentspecific processes, output – centralized database and IT-supported departmental but coherent processes}; World views {centralized database system yields consistent data records
and enhances data sharing}; Owners {Technical Planning Committee, Ministry of Local
Government}; and Environmental or external issues {financial resources from the Ministry}. Furthermore, the script in figure 11 (activity 4.2.2) shows that requirements can
be thoroughly elaborated using the Requirements Elaboration Template (see figure 12).
This template shows how the above CATWOE aspects of SSM can be used to assess the
requirements that the architecture must address. Figure 12 shows that activity models can
be used to represent the transformation processes that must be implemented or executed
44
We lack: IT enabled information
processing, centralized information
storage, real time information
retrieval, & data sharing
Ministry of Local
Government
Technical Planning
Committee
Create
Community
welfare services
e.g. governing
laws, tax
collection,
water, roads,
buildings,
financial
management,
etc
Statutory
bodies
District 3-year
Development
Plan
Prepare &
Oversee
implementation of
Registry
Provide
Planning
Provide
Natural
resources
Finance
Provide
Provide
CAO
Education
Internal
audit
Health
Build
Works
Production
& marketing
Provide
Community
welfare services
e.g. schools,
hospitals,
farms, land use
planning,
environmental
preservation etc
Create
Create
Community based
development
Create
Lower Local
Governments
Make use of
Make use of
Development
programs
Make use of
Community
Community
Figure 7: Rich Picture For WDLG as-is Situation
so as to achieve a given requirement. Figure 12 is an adaptation of the format of activity
models provided by [HWP99].
4 Performance Evaluation of the CEADA Script
The CEADA extension script presented in section 3 was evaluated in the Department of
Radiotherapy at Mulago Hospital in Uganda. This department is concerned with: the
treatment of cancer; training of radiography students of the Makerere University College
of Health Sciences; undertaking research on cancer treatment; and sensitizing the public
about cancer issues. Out of the 20 staff members in the department, about 14 participated
in the CEADA evaluation. The aim of CEADA in this case was to guide the collaborative creation of an enterprise architecture that would improve the execution of operational
processes in the radiotherapy department.
The execution of the CEADA script was done using interviews and a collaborative session.
Output from interviews was first analyzed and it was then used to formulate the preliminary models that were discussed in the collaborative session. Due to the busy schedule of
the department (given its various patients), only one collaborative session was conducted,
in which all activities were executed. Thereafter, members who attended the session were
45
External Constraints
??
Why should
it be done?
??
Internal Constraints
??
??
??
??
??
How should
it be done?
??
??
??
??
Solution
Alternative (SAp)
What should
be done?
??
Architecture Requirements
of the Chosen Solution
Alternative
??
??
High Level Solution
Specifications
Figure 8: Adapted Ishikawa Diagram Template for Constraints and Requirements
To support decision
making regarding
the desired
transformation
To show the impact
of the desired
transformation
Purpose of Enterprise Architecture
To specify business To inform & contract
requirements
service providers
Figure 9: Diagram Template for Purpose of the Architecture Initiative (Based on [OPW+ 08])
given questionnaires so as to capture their evaluation of the performance of the CEADA
script against a number of quality criteria. Table 1 presents the evaluation results. Column
2 of table 1 shows the evaluation criteria that were used. These criteria are discussed in
detail in [NBP11]. The scores used to measure CEADA’s performance were based on a
5-point Likert scale, as shown in the last row of table 1. The mean score of CEADA’s performance under each criterion (or sub criterion) is shown in table 1. The standard deviation
of the scores (as shown in table 1) illustrates the degree of variability among stakeholders
regarding CEADA’s scores under each criterion. The mean scores in table 1 indicate that
CEADA’s performance was good regarding the support for creating a shared understanding and shared vision among stakeholders. In addition, the low standard deviations of
scores (shown in the last column of table 1), implies that there was generally a high level
Table 1: Performance Evaluation Results of the CEADA Script
#
Evaluation Criteria for Measuring Performance of the CEADA Script
Performance Indicator
Mean score Standard deviation
of scores
4.36
0.50
1 Support for creating a shared understanding (among stakeholders) of the problem
and solution aspects of the organization
a Support for enabling stakeholders to understand why some of their
4.15
0.55
concerns/views would not apply in some contexts
b Support for enabling stakeholders to understand the concerns of other
4.36
0.84
stakeholders about the current and future operations in the organization
c Support for enabling stakeholders to understand the results of the architecture
4.25
0.45
process
2 Support for enabling stakeholders to freely express their views about the current
4.43
0.94
operations in the organisation
3 Support for attaining stakeholders satisfaction with the activities done in the
4.50
0.52
collaborative session(s)
4 Support for attaining stakeholders satisfaction with the outcome(s) of the
4.23
0.60
collaborative session(s)
a Support for enabling constructive critiquing of ideas generated in the session(s)
1.50
0.52
b Support for enabling stakeholders to understand the objectives of the session(s)
4.29
0.61
Note: The evaluation scale used in the questionnaire that evaluated CEADA sessions is a 5 point Likert scale with
responses ranging from strongly disagree (point 1) to strongly agree (point 5)
46
18. Choose appropriate business solution alternative
Tool Name: MultiCriteria Step Name: [Activity 1.4.2] Output File Name: ChosenSolnAlternative.mw
Step Specific Information: List of Alternatives - [OgzdBusinessSolnAlternatives.mw]; Input Criteria List Items [Suitability with respect to: (1) satisfying organization principles i.e. internal constraints; (2) achieving business
strategy; (3) achieving business goals; and (4) satisfying external regulations or constraints]
Facilitator Notes: Prompt stakeholders to comment on their scores of the business solution alternatives.
19. Define external constraints that should be considered during enterprise architecture effort
Tool Name: Generate Step Name: [Activity 1.5.0] Outline File Name: GenExternalConstraints.mw
Topic List: (1) What are the policies or laws from regulatory authorities (& what do they imply) that relate to the
defined solution aspects? (2) What are the policies or laws from the organization's consortiums, donors or
sponsors, unions etc, that relate to the defined solution aspects?
Facilitator notes: Explain answering format: {name of body} - {external principle} - {implied external constraint}
20. Clarify & organize the identified external constraints
Tool Name: Organize Step Name: [Activity 1.5.1] Output File Name: OgzdExternalConstraints.mw
Step Specific Information: Input file - [GenExternalConstraints.mw]; Step will run in Outliner Mode.
Facilitator notes: Plug results into the right section in the holistic data capture pyramid or template.
21. Define the purpose of the enterprise architecture effort
Tool Name: Generate Step Name: [Activity 1.6.0] Outline File Name: GenPurposeEAEffort.mw
Topic List: (1) What is the main issue that the enterprise architecture effort must address in order to achieve
the defined solution aspects? (2) In realizing the desired solution aspects, how do you want to use the resultant
enterprise architecture?
22. Clarify & organize aspects regarding the purpose of the architecture effort
Tool Name: Organize Step Name: [Activity 1.6.1] Output File Name: OgzdAspectsPurposeEAEffort.mw
Step Specific Information: Input file - [GenPurposeArchitectureEffort.mw]; Step will run in Outliner Mode.
Facilitator notes: Guide participants to classify their ideas about the purpose of the architecture into 4 major
categories (which are adaptation of the general uses of enterprise architecture defined in Op t Land et al
(2008)), i.e. enterprise architecture will be used for: (a) supporting decision making regarding the chosen
business solution alternative; (b) showing the impact of the business strategy and the chosen business solution
alternative; (c) specifying business requirements; (d) informing & contracting service providers.
23. Agree on the general purpose of the enterprise architecture effort
Tool Name: Evaluate Step Name: [Activity 1.6.2] Output File Name: AgreedPurposeArchEffort.mw
Step Specific Information: Input topic file - [OgzdAspectsPurposeEAEffort.mw]; Evaluation Method - [Select
(Mark all that apply)]; Display Variability - [Yes]
24. Define high level solution specifications & scope of the enterprise architecture effort
Tool Name: Generate Step Name: [Activity 1.7.0] Outline File Name: GenSolnSpecsScope.mw
Topic List: (1) Which business services, functions, departments or areas of the organisation should the
architecture creation effort focus on? (2) Which of the existing business functions should not be considered in
the desired state? (3) Which architecture domains (i.e. business, data, applications, & technology) should be
covered during architecture creation? (4) What is the desired level of detail that the architecture creation effort
should focus on? (5) From the completed & ongoing projects in the organization, which architectural resources
are available in the organisation, and can be considered for use during architecture creation? (6) List any
constraints (i.e. enterprise-wide or project-specific constraints) that the organization principles impose on the
implementation of the defined business strategy or chosen business solution alternative
25. Clarify & assess the defined high level solution specifications or architecture scope dimensions
Tool Name: Organize Step Name: [Activity 1.7.1] Output File Name: OgzdSolnSpecsScope.mw
Step Specific Information: Input file: GenSolnSpecsScope.mw; Step will run in Outliner Mode.
Facilitator notes: assess aspects with respect to problem scope & the purpose of architecture
26. Agree on high level solution specifications & scope of the architecture effort
Tool Name: Evaluate Step Name: [Activity 1.7.2] Output File Name: AgreedSolnSpecsScope.mw
Step Specific Information: Input topic file - [OgzdSolnSpecsScope.mw]; Evaluation Method - [Select (Mark all
that apply)]; Display Variability: [Yes]. Plug results into the holistic data capture pyramid or template.
27. Determine all solution owners & their roles in the architecture effort
Tool Name: Generate Step Name: [Activity 1.8.0] Outline File Name: GenSolutionOwners.mw
Topic List: List all stakeholders who will be affected by, or benefit from, or participate in the implementation of,
the desired solution; & mention what the role of that stakeholder will be in the enterprise architecture effort.
Facilitator Notes: Answering format: {problem/solution owner} - {Role in the architecture development effort}
Figure 10: Facilitation Script for adapting SSM in Architecture Creation (Contd.)
of consensus (among stakeholders) on the performance of CEADA under each criterion.
Moreover, comparing CEADA’s performance results at the department of Radiotherapy
with its performance results in field study one (see [NBP11]), the performance of CEADA
regarding the support for creating a shared understanding and shared vision among stakeholders has improved. For example, in field study one, the average score of CEADA’s
performance under “creating a shared understanding and shared vision among stakeholders” was 3.80. This average score has increased to 4.28 after the adaptation of SSM and
Ishikawa diagram techniques into CEADA. Certainly, the difference in CEADA’s performance may also be caused by other factors such as the variation in the complexity and
scope of the organization’s problem and solution aspects, or in the type or personalities
of stakeholders dealt with. However, stakeholders’ evaluation highlighted that the Rich
Picture model gives a general view of operations in an organization but lacks detailed information about how the operational processes are executed and their attributes. It was also
acknowledged that the use of the Ishikawa model templates and the activity models (that
are represented in form of a requirements elaboration template) helped to freely represent
47
28. Clarify & organize list of solution owners
Tool Name: Organize Step Name: [Activity 1.8.1] Output File Name: OgzdSolutionOwners.mw
Step Specific Information: Input file - [GenSolutionOwners.mw]; Step will run in Outliner Mode.
29. Determine all decision makers in the architecture process & their roles
Tool Name: Generate Step Name: [Activity 1.9.0] Outline File Name: DecisionMakersRoles.mw
Topic List: Who are the key decision makers in the organization?
Facilitation notes: Answering format: {position in the organization} - {decision making responsibility}
End of session 1 & preparations for session 2
30. Communicate purpose of session 2
Tool Name: Manual Step Name: [3.0] Facilitator Notes: Give a recap of results from session 1
31. Define concerns regarding problem & solution aspects from session 1
Tool Name: Generate Step Name: [3.1] Outline File Name: GenConcernsProbSoln.mw
Topic List: (1) What concerns do you have regarding the problem & solution aspects?; (2) What has not been
captured in the problem aspects? (3) What has not been captured in the solution aspects? (4) Which problem &
solution owners have not been included?
32. Clarify & organize concerns about problem & solution aspects
Tool Name: Organize Step Name: [Activity 3.2 and 3.3] Output File Name: OgzedConcernsProbSoln.mw
Step Specific Information: Input file - [GenConcernsProbSoln.mw]; Step will run in Outliner Mode.
Facilitator Notes: Discuss the concerns & refer to the Ishikawa diagram template for problem analysis.
33. Validate & agree on the concerns about problem & solution aspects
Tool Name: Evaluate Step Name: [Activity 3.4] Output File Name: ValidConcernsProbSoln.mw
Step Specific Information: Input topic file - [OgzedConcernsProbSoln.mw]; Evaluation Method - [Select (Mark
all that apply)]; Display Variability - [Yes]
34. Define business requirements that the enterprise architecture must fulfill
Tool Name: Generate Step Name: [Activity 4.1] Outline File Name: GenReqtsForEA.mw
Topic List: To move from the current to the desired state, define WHAT should be done; HOW it should be
done; & WHY it should be done.
Facilitator Notes: Answering format is an adaptation of the Root Definitions of SSM, i.e.: {WHAT should be
done}; {HOW it should be done}; & {WHY it should be done}
35. Clarify & organize requirements for the enterprise architecture
Tool Name: Organize Step Name: [Activity 4.2] Output File Name: OgzdReqtsForEA.mw
Step Specific Information: Input file: GenReqtsForEA.mw; Step will run in Outliner Mode.
Facilitator Notes: Invoke a specialization-driven-division & guide stakeholders to:
(1) Identify and delete all duplicate root definitions; determine which root definitions should be sub root
definitions of other root definitions.
(2) Fill in the prompts in the Ishikawa diagram template for capturing requirements for the architecture.
Converge all subgroups & prompt stakeholders to identify any incomplete requirement (i.e. one which is not in
the format of: {WHAT should be done}; {HOW it should be done}; & {WHY it should be done}. Plug results into
the holistic data capture pyramid or template.
36. Elaborate requirements for the architecture using CATWOE analysis and Activity Models of SSM
Tool Name: Generate Step Name: [Activity 4.2.1] Outline File Name: ElaboratedReqsForEA.mw
Topic List: (1) For each requirement, who are the Customers or beneficiaries? (2) Who are the Actors
responsible for realizing each requirement? (3) What are the required inputs to realize the Transformation
process described in each requirement, and what are the expected outputs from the transformation process?
(4) What are the World views that make each requirement meaningful or rational? (5) Who are the owners or
sponsors responsible for sponsoring the realization of each requirement or stopping it? (6) What are the
Environmental & external issues or constraints that may hinder the realization of each requirement?
Facilitator Notes: guide the stakeholders in doing a CATWOE analysis of each agreed on root definition
37. Clarify & organize CATWOE aspects that describe the requirements & solution scenarios
Tool Name: Organize Step Name: [Activity 4.2.2] Output File Name: OgzdElaboratedReqtsForEA.mw
Step Specific Information: Input file - [ElaboratedReqsForEA.mw]; Step will run in Outliner Mode.
Facilitator Notes: Guide stakeholders to use the Requirements Elaboration Template to clarify, organize, &
verify aspects that describe the requirements & solution scenarios that the enterprise architecture must address.
38. Validate & agree on the requirements for the enterprise architecture
Tool Name: Evaluate
Step Name: [Activity 4.3.0] Output File Name: ValidatedReqtsForEA.mw
Step Specific Information: Input topic file - [OgzdElaboratedReqtsForEA.mw]; Evaluation Method - [Rank
from 1 to N (Use each value only once)]; Display Variability: [Yes]
Facilitator Notes: Encourage participants to comment on the ranks they assign to requirements.
Figure 11: Facilitation Script for adapting SSM in Architecture Creation (Contd.)
and structure various aspects associated with the execution of the processes. Therefore,
although CEADA’s improved performance may not only be attributed to the adaptation of
these new techniques, the new techniques played a key role in supporting the organize and
clarify patterns of reasoning.
5 Conclusions
Creating an enterprise architecture in a collaborative setting (with active participation of
clients) helps to create ownership and commitment among stakeholders. Thus, CEADA
is being developed using Collaboration Engineering so as to provide enterprise architects
with an explicit way of actively involving stakeholders during architecture creation. From
the first field study evaluation of CEADA, some weaknesses were identified in its design
48
Customers or
Beneficiaries or victims
of requirement Xn
External Factors that may
affect the achievement of
requirement Xn
Transformation
processes to be executed
in order to achieve a
given requirement (Xn)
Process
P2
??
Process
P4
Process
P1
Process
P5
Process
P7
Process
P3
Process
P6
Process
Pk
Owners or sponsors
associated with
achieving
requirement Xn
Process
P8
??
Internal Actors that will
achieve requirement Xn
World view that
Indicates the
significance of
achieving
requirement Xn
External Actors
Figure 12: Requirements Elaboration Template
that called for adaptation of SSM techniques so as to address complexities that occur in a
collaborative enterprise architecture development environment. Therefore, herein CEADA
has been extended using a script that adds on our earlier work by providing facilitation
support for using SSM and Ishikawa diagram techniques to execute activities that require
the use of clarify and organize patterns of reasoning.
With the CEADA script, hands-on techniques have been adapted for each activity that requires clarifying, organizing, and discussing aspects from brainstorming tasks; and guidelines are given on how enterprise architects can use SSM and Ishikawa diagram techniques
to gather data (in group sessions with clients) that can be used to create baseline and target
architecture models. Although the Rich Picture has been adapted as a technique for gathering baseline information, it can also be used to gather information on the target situation.
Also, embedded in the script is a questionnaire that can be used during interviews with
stakeholders, or as template for gathering information from organization documentations.
Thus, this research kick-starts the support for enterprise architecture data gathering and
stakeholder involvement with SSM, in order to create a shared understanding and shared
vision of the problem and solution aspects among organizational stakeholders. However,
in line with the Design Science approach, the repeatability and predictability of the script
is yet to be determined through further evaluation of CEADA. Further evaluation and refinement of the CEADA script will result in a very useful method that can be used along
with the existing enterprise architecture frameworks.
Acknowledgements: We are extremely grateful to staff members of the department of
Radiotherapy in Mulago hospital for their participants in this research.
49
References
[Ben05]
Benard, S.A.: An Introduction to Enterprise Architecture - Linking Business and Technology. AuthorHouse, Indiana, 2005.
[BVN03]
Briggs, R.O.; Vreede, G.J. de; Nunamaker, Jr. F.: Collaboration Engineering with
ThinkLets to Pursue Sustained Success with Group Support Systems. Journal of Management Information Systems, 19, 2003; pp. 31–64.
[BVN+ 06] Briggs, R.O. ; Kolfschoten, G.L.; Vreede, G.J. de; Dean, D.L.: Defining Key Concepts
for Collaboration Engineering. AMCIS, Acapulco, Mexico, 12, 2006.
[OPW+ 08] Op ’t Land, M.; Proper, H.A.; Waage, M.; Cloo, J.; Steghuis, C.: Enterprise Architecture:Creating Value by Informed Governance. Springer Berlin, Germany, 2008.
[Che98]
Checkland, P.: Systems Thinking, Systems Practice. John Wiley & Sons Ltd, 1998.
[Gar09]
Gartner, Inc.: Gartner Identifies Ten EnterpriseArchitecture Pitfalls: Gartner Enterprise
Architecture Summit 2009. http://www.gartner.com/it/page.jsp?id=1159617. Accessed
on August 15th 2010.
[BVN03]
Hevner, A.R.; March, S.T.; Park, J.; Ram, S.: Design Science in Information Systems
Research. MIS Quarterly, 28(1), 2004; pp. 75–105.
[HWP99]
Hopkins, T.; Whitford, I.; Pidd, M.; Winter, M.: Handling Strategic Problems: Methodologies. http://www.orsoc.org.uk/about/teaching/strategicproblems/m s 3frs.htm. Accessed on February 15th 2011.
[Ish86]
Ishikawa, K.: Guide to Quality Control.. 2nd Edition. Asia Productivity Organization,
Tokyo. 1986.
[KV07]
Kolfschoten, G.L.; Vreede, G.J. de: The Collaboration Engineering Approach for Designing Collaboration Processes CRIWG, Springer, Heidelberg, LNCS volume 4715,
2007; pp. 95–110.
[Lan05]
Lankhorst, M. et al: Enterprise Architecture at Work: Modelling, Communication, and
Analysis. Springer Berlin, Heidelberg, 2005.
[NBP10]
Nakakawa, A.; Bommel, P. van; Proper, H.A.: Challenges of Involving Stakeholders
When Creating Enterprise Architecture. EIS, Eindhoven, The Netherlands, CEUR-WS
662, 2010; pp. 43–55.
[NBP11]
Nakakawa, A.; Bommel, P. van; Proper, H.A.: Definition and validation of requirements
for collaborative decision making in enterprise architecture creation. International Journal of Cooperative Information Systems (IJCIS), 20(1), 2011; pp. 83–136.
[Sta95]
The Standish Group.:
The Standish Group Report – CHAOS Report, 1995.
http://spinroot.com/spin/Doc/course/Standish Survey.htm. Accessed March 3rd 2011.
[VB05]
Vreede, G.J. de; Briggs, R.O.: Collaboration Engineering: Designing Repeatable Processes for High-Value Collaborative Tasks HICSS, IEEE, 2005.
[Wak08]
Wakiso District Council: Three Year Development Plan for Wakiso District Wakiso
District Local Government offices, 2008.
50
Integration of Process Constraints from Heterogeneous
Sources in Process-Aware Information Systems
Stefanie Rinderle-Ma and Juergen Mangler
University of Vienna, Austria
Faculty of Computer Science
{stefanie.rinderle-ma,juergen.mangler}@univie.ac.at
Abstract: Business Process Compliance (BPC) has gained significant momentum in
research and practice during the last years. Although many approaches address BPC,
they mostly assume the existence of some kind of unified base of process constraints
and focus on their verification over the business processes. However, it remains unclear how such an integrated process constraint base can be built up, particularly at
the presence of process constraints stemming from heterogeneous sources such as internal controls, external contracts, or security and privacy policies. Hence in this paper, we propose a representation framework for the integration of process constraints
stemming from different sources. The representation framework is generic such that
existing formalisms to represent process constraints can be integrated. The framework
is illustrated by means of real-world process constraints from different domains and
heterogeneous sources. Altogether process constraint integration enables consistency
checks and optimizations as well as maintenance and evolution of the constraint base.
1 Introduction
Business Process Compliance (BPC) has become a major challenge for enterprises nowadays. BPC requires that business processes comply with certain relevant rules, regulations,
or norms. Although many approaches address BPC [NS07, LSG07, SGN07, LRD10,
AWW11], they mostly assume the existence of some kind of integrated base of process
constraints and focus on the verification of these constraints over the business processes.
By process constraints we refer to those rules in the domain of interest, that are associated with processes. As literature and practice show, the co-existence of processes and
process constraints is common and desired [HG09, LRD10]. However, it remains unclear
how such an integrated process constraint base can be built up and used for compliance
checks within the Process-Aware Information System (PAIS), even though this constitutes
the essential prerequisite for all further compliance checks.
The particular challenge of providing an integrated process constraint base results from
the heterogeneity of sources process constraints might stem from. First of all, the sources
and purposes of process constraints might be different. Examples comprise internal quality directives such as Six Sigma or ITIL, external controls such as standards (e.g., ISO
001), regulations by authorizing bodies, guidelines [PT09], or regulations based on contracts (business contracts) [SGN07]. For a specific application scenario consider Fig. 1
51
Figure 1: Process Constraints Stemming from Heterogeneous Sources Imposed on Current PAIS
that displays example processes and rules from the medical domain: here, business processes might be subject to medical guidelines [PT09] on the one side and also subject to
authorization and privacy constraints on the other side. Fig. 1 also covers simple resource
synchronization as well as temporal rules. The basic challenge is to integrate all these
different kinds of process constraints within one representation in order to enable consistency checks, integrated compliance verification, and a better support for constraint base
maintenance. The latter becomes important since process constraints are often subject to
change and evolution [LRGD09].
In this paper, we present a framework that enables the integrated representation for process constraints stemming from heterogeneous sources.1 The framework covers process
perspectives such as control flow, data flow, time, and resources that can be subject to
process constraints. The integrated representation also enables the specification of behavior which is relevant for process constraints that require some sort of action (e.g., synchronization between process instances). The design of the representation framework
also enables the integration of existing specification formalisms for process constraints,
for example, SeaFlows [LRD10], logic-based rules [PSvdA07], or BPMN-Q constraints
[AWW11]. The unified representation is validated against a selection of real-world process constraints. We further discuss how such a framework can be implemented within
PAIS. Altogether the proposed approach provides a first step towards the integrated management, verification, and evolution of process constraints in PAIS, tackling the challenge
1 An
extended version is available as technical report [MRM11].
52
Figure 2: Process Constraint Properties
of heterogeneous sources for process constraints.
Sect. 2 discusses constraint properties in PAIS. The representation formalism for integrated process constraints is presented in Sect. 3 followed by an evaluation in Sect. 4.
Sect. 5 compares related approaches and Sect. 6 concludes with summary and outlook.
2 Constraint Properties in Process-Aware Information Systems
In the previous sections, a unified representation for process constraints has been presented. In this section, we discuss its usage within Process-Aware Information Systems
(PAIS) along different constraint properties as set out in Fig. 2.
Usage: Process constraints can be used to verify the compliance of business processes
with relevant regulations and constraints. This is useful for process design and evolution.
Constraints can also be used to alter or affect the behavior of processes (behavioral). Affecting the behavior of processes includes the attribution of process activities or, more
generic, the attribution of structural patterns. Examples include resources allocation of
process activities.
Additionally, constraints can be used to specify meta constraints. Meta constraints are
intended to check the consistency of other constraints, for example, constraint C5 in Fig.
1. Another example is a meta constraint specifying that for each process activity referring
to a resource centrifuge synchronization constraint C10 in Fig. 1 is assigned to.
Application: Application deals with the question at which phase of a process life cycle
constraints may be enforced. Basically, at design-time constraints are checked to verify
that a process complies to certain criteria (compliance) constraints. All constraints that are
checked at design-time are compliance constraints or meta constraints.
All behavioral constraints, and some compliance constraints are checked at run-time. Examples include the checking for data value constraints, or the attribution of structural patterns (the attributes are used at run-time, though can be checked by meta constraints at
design-time). Compliance constraints may affect run-time under certain circumstances.
E.g. although a process may not conform to a specific constraint at design-time, it may
do so at run-time, because only a certain execution path violates the constraint. Hence
corresponding process instances are to be monitored at run-time [LRD10].
Scope: The most important scope perspective is structure as in all constraints structural
53
information has to be present (either explicit or implicit through connections of information to structure), as otherwise they would not be connected to processes and could thus
not be enforced in PAIS. Note that if a constraint holds on only structural information, it
is always a compliance constraint.
Data is also an integral part of a process tightly connected to structure. At design-time,
it is possible to check if data types and data flow conform to a certain schema. Further
it can be checked whether certain data values will lead to compliance violations are runtime. If, for example, a treatment process states that a lab test is to be performed for all
patients beyond 65 years and the corresponding medical guideline states that the lab test
is mandatory for patients beyond 62 year, it can be already checked at design-time, that
for patients between 62 and 64 there will be a violation of the corresponding constraint at
run-time.
Resource and time are attributes that are typically connected to structure (or data). Their
purpose is to describe information that aids the execution of a process. Resource-aware and
time-aware constraints in PAIS are typically checked or enforced at run-time including:
• Resource assignment to process activities, typically specified based on access constraints (e.g., constraint C11 in Fig. 1). Based on role assignment, process activities
are offered to authorized actors in their work lists at run-time (e.g. by a RBAC
component).
• On top of access constraints, authorization constraints can be specified such as dynamic separation of duties. Dynamic authorization constraints are verified during
run-time.
• Assign temporal information to activities (e.g. the normal duration of a certain activity is 2 hours with a standard deviation of 10 minutes).
Origin: All constraints become available through not specified external resources, either
at design time, run-time or change time. They are identified and structured by constraint
designers and then made available through a constraint-base, and have to be enacted from
this point on. One exception is constituted by SLAs for dynamic service selection. These
constraints become available through execution of a process instance, and vanish after the
instance finishes.
3 A Framework for Process Constraint Integration
In this section the framework for integrating process constraints in PAIS is introduced
that should cover all constraint properties set out in Section 2. As fundamental property,
process constraints are associated with one or several processes that are implemented and
executed within the PAIS of interest. As process literature study shows, all formalisms on
process constraints (implicitly) incorporate a structural pattern that contains at least one
activity either executed within a process or stored within the process activity repository.
54
In the SeaFlows project [LRD10], for example, process constraints are tied to one or more
process activities and consist of an antecedent and a consequence part. More precisely,
based on a triggering antecedent pattern, a consequence pattern must hold to fulfill the
constraint. Process constraints as defined in DECLARE also refer to process activities and
the semantics for relations between activities are based on LTL (Linear Temporal Logic)
[PSvdA07]. BPMN-Q offers compliance patterns containing structural patterns within the
triggering part of the constraint [AWW11].
In the following section we will define an extended concept of Linkage for process constraints that subsumes the structural pattern of the constraint, but might further incorporate information on the Context of the process constraint, i.e., process types or process
instances the constraints refer to, and the Trigger position. The latter is required for capturing process constraints that can trigger an action such as synchronization.
3.1
Expressing Process Context by Linkage
The following definition of Linkage integrates the Contextc of a process constraint c
together with its structural pattern SPc and a trigger position TPc (cf. Fig. 3).
Definition 1 (Linkage) Let P be a set of all process types of consideration and IP be the
set of process instances running according to a process type P ∈ P. A process type P ∈
P is described by a process schema SP := (AP , EP , DP )2 where AP denotes the set of
activities, EP the set of control/data edges, and DP the set of data elements SP consists
of. Then the linkage Linkagec to a process type P is defined as follows:
Linkagec ⊆ Contextc × SPc × TPc
Contextc ⊆ P × IP
∃SPc
TPc ∈ ∅, before(an,P ), after(an,P )
where
∧
∧
In Def. 1, Contextc describes in which processes or process instances a constraint may
occur. Possibilities include single instances (e.g. SLAs for services that have been dynamically selected), all instances of a process (attribution of resources to tasks) or even
instances in multiple processes (when a resource is used in multiple processes, and synchronization has to take place). Structural patterns SPc express the association between
process constraint and process. In connection with context, it defines for which parts of
process types and process instances the corresponding process constraint is to be enforced
or verified. Structural patterns may not only spawn single activities but also several activities connected through complex control decisions. Finally, the trigger position TPc is
relevant for synchronization constraints, since synchronization constraints are constraints
that not only set out certain conditions on process execution, but enforce an action. In this
2 In order to stay meta-model independent, we assume a simple generic representation for process schemas
that can be adopted by any (imperative) process meta model.
55
case the trigger position specifies whether the action should take place before the affected
activity is started or after. The Trigger Position is solely present for run-time (behavioral)
constraints. As a structural pattern may not only spawn a single activity but also several
activities connected through complex control decisions, it is necessary to determine when
exactly a constraint has to fire. This is the equivalent of describing the condition under
which an event occurs, in an Event Condition Action (ECA) rule. The trigger position itself is simple: before or after an activity occurs. Of course multiple before / after positions
can be specified.
Both, SPC and TPc are grouped as Connection in Fig. 3 to express their tight integration. A TPc can not exists without a set of activities matched by a structural pattern SPc .
Apart from TPc , linkage information can be statically matched against processes, thus it
is possible to determine enabled and idle constraints (as described before).
From now on we will denote a linkage Linkagec in the compact form:
Linkagec : ((P, IP ), SPc , TPc )
Consider compliance constraint C6 depicted in Fig. 1: “C6:
operation, patient must be examined”.
Before any invasive
The antecedent pattern “existence of activity invasive operation” within
a process triggers the check whether this activity is preceded by an activity “patient
examination”. The linkage for this compliance constraints turns out as
LinkageC6 : ((Invasive Surgery, ALL), SPC6 , ∅)
with
SPC6 : ∃a1 Is(a1 , examine patient)
∃a2 Is(a2 , conduct surgery)
a1 A ∗ a 2
∧
∧
where
A∗ : arbitrary activities between a1 and a2
3.2
Integrating Data, Time, and Resources
Existing approaches mostly deal with control flow constraints, i.e., constraints that are
only referring to structural patterns within a process [AWW11]. The only approaches
that have addressed data flow aspects within process constraints are SeaFlows [KLR+ 10]
and BPMN-Q [AWW11]. In accordance to these approaches, the data flow perspective
within a process constraint can be represented as condition on the structural pattern it
refers to. Consider constraint C8 from Fig. 1: “C8: Beyond age 62, a blood
test must be performed”.
56
Figure 3: Process Constraint Structure (Integrated Representation)
We can see that the data flow condition “Beyond age 62” imposes a restriction on the
structural pattern of the constraint (“blood test”). However, control and data flow are
only two perspectives of a process. As case studies show, process constraints might also
refer to conditions on time and resources, e.g., constraints C3, C4, C5, C7, C9, and C11
within the example depicted in Fig. 1. Hence, constraint integratoin requires the specification of data, time, and resource conditions on top of the linkage part of the constraint.
Definition 2 (Condition) Let c be a process constraint with linkage Linkagec referring
to a process type P described by process schema on linkage SP := (AP , EP , DP ). The
condition of c imposed on Linkagec is defined as
Conditionc ⊆ expr(DP ) × expr(timec ) × expr(resourcec )
where
• expr(DP ) denotes a logical expression over the data elements of P
• expr(timec ) denotes a temporal expression and
• expr(resourcec ) denotes a logical expression over the resources associated with P
(typically modeled within a organizational or resource model)
It is important to mention that a condition cannot exist without referring to a linkage.
Thus, a condition describes, under the premise of a given linkage, either (a) a combination
of data, temporal or resource conditions that have to be present or (b) a combination of
data, temporal or resource conditions that have to be present in order for a given behavior
to be apply.
For constraint C6, for example, no specific condition is imposed on the linkage part.
Hence:
ConditionC6 : ∅
When considering constraint C3 in Fig. 1, things get more interesting, since “After a
blood test wait 4 hours before sonography” obviously imposes a time
57
condition on the linkage part. Less obviously, an additional condition is imposed on the
data flow, since the 4 hours time frame between blood test and sonography are only required for the same patient, formally:
LinkageC3 : (Invasive Surgery, ALL), SPC3 , ∅)
with
SPC3 : ∃a1 Is(a1 , blood test)
∃a2 Is(a2 , sonography)
a1 A∗ a2
∧
∧
and
ConditionC3 : patient(a1 ) = patient(a2 )
min time between(a1 , a2 , 4h)
∧
Similar considerations can be made for data (“C8: beyond age 62”) and resources
(C7: examination by same user as surgery) which can be also represented
by a condition part imposed on the linkage of a constraint.
3.3
Expressing Behavior within Process Constraints
The last missing piece is, given a certain linkage and condition, to allow for a certain
assignment or behavior. Assignment becomes necessary for access constraints such as
C11. Behavior specifies for example the action part of a synchronization constraint such
as C10. Hence, we complete the unified constraint representation by a Behavior part.
Definition 3 (Behavior) Let c again be a process constraint. For a given Linkagec and
Conditionc , a behavior can be either empty, the attribution of resource or time information
or instructions specific for process execution (e.g. exceptions).
3.4
Summary: Linking, Condition, Behavior
The overall representation for process constraints consisting of linkage, condition, and
behavior is summarized in Fig. 3. In Sect. 4 we will evaluate the unified representation
against all case study constraints shown in this paper and existing related approaches.
4 Evaluation
In Sect. 4.1, we show how a selection of real-world process constraints can be expressed
based on the integrated representation framework . In fact, we have analyzed several
58
application domains and related rules within SeaFlows [LRD10] throughout the last 5
years. Additionally, the examples from the medical domain (cf. Fig. 1) are analyzed.
4.1
Process Constraints from Different Domains
Medical Guidelines: A first medical guideline taken from the report “Prevention, Detection, Evaluation, and Treatment of High Blood Pressure” [U.S03] is depicted in Fig.
4-R1) together with a fragment of a corresponding treatment process. An analysis of this
constraints over the process shows that blood pressure ≥ 20/10 mmHg above
goal blood pressure refers to the data perspective (data element blood pressure)
and part initiating therapy refers to structural pattern initiate therapy).
631 .$4!,A< CB!4$<!@$ > ;<''4 8#$==B#$
=<;;:
AKN998KN
*( L<;;: AKN998KN %9 7 654$5 >>3O !L;+N O;!< L<;;: AKN998KN2 );'9%:NK!#%;' 9M;8<: LN
O%+N' #; %'%#%!#%'O #MNK!A@ 1%#M #1; !ON'#92 ;'N ;( 1M%)M 898!<<@ 9M;8<: LN ! #M%!0%:N/#@AN
601
.$4!,A< CB!4$<!@$ > 5A,,!@A&!'@
.MN >%'%>8> !ON (;K #MN (%K9# :;9N %9 9%? 1NNJ9, .MN 9N);': :;9N %9 #; LN !:>%'%9#NKN:
!(#NK ! >%'%>8> ;( I 1NNJ9 %( #MN (%K9# :;9N !:>%'%9#NKN: !# !ON 7 $6 >;'#9 !': )8KKN'# !ON
H 6I >;'#M92 ! >%'%>8> ;( G 1NNJ9 F!9 (%'!< :;9NE !# !ON 7 $6 >;'#M !': )8KKN'# !ON
LN#1NN' 6I !': DC >;'#M9, B;K MN!<#M@ )M%<:KN'2 '; (8K#MNK :;9N9 'NN:N: %( #MN #MN (%K9#
:;9N !:>%'%9#NKN: !# !ON 7 6I >;'#M,
*'%#%!#N
#MNK!A@
,,,
N?!>%'N
"ON
&;9N$
(7!#=& 4'=$20/ -+)
=&A&?%$A<&%9:
-%+N (%K9#
:;9N
*
-%+N
9N);':
:;9N
('&%$#"!=$:
Figure 4: Medical Guidelines
The second medical guideline (cf. Fig. 4-R2) describes the vaccination of children for
Pneumococcal [PT09]. This guideline refers to the data perspective, e.g., data element
age) and structural pattern
SPR2 : ∃a1 Is(a1 , give first dose)
∃a2 Is(a2 , give second dose)
a1 A∗ a 2
∧
∧
In addition, it refers to temporal aspects, i.e., certain time intervals between vaccination
should be maintained.
Financial Sector: Fig. 5-R3 shows a financial regulation taken from regulatory package
BASEL II [Bas06]. Here the basic question is whether “capital ratio” is a process data
element or an application data element, i.e., whether the constraint refers to the data perspective or application perspective. If we assume that capital is solely application data as
depicted for Process P1, the constraint above does not refer to any structural aspect of the
process and hence is not considered as process constraint. By contrast, as for Process P2,
the capital ratio is still calculated by the application connected with activity calculate.
However, the value is explicitly written as process data element ratio. Consequently,
the linkage part of the constraint refers to structural pattern calculate.
Authorization: Separation of duties is one example for authorization constraints [SM10].
Assume the constraint depicted in Fig. 6-R4. It refers to the functional perspective
59
"!6 531/1-3/4 "20.4/,3+1 * )313(.( '/&3,/4 "2%.3$2(21,#
4!3%"<< 4D1
0." %'+?#'& !'#?3 ?< %'&%$&'#") $<?CA #." )">?C?#?3C 3> !"A$&'#3!; %'+?#'& 'C) !?<985"?A.#") '<<"#<*
0." #3#'& %'+?#'& !'#?3 2$<# /" C3 &35"! #.'C -,* 0?"! E %'+?#'& ?< &?2?#") #3 DBB, 3> 0?"! D
%'+?#'&* @***= 03#'& !?<985"?A.#") '<<"#< '!" )"#"!2?C") /; 2$&#?+&;?CA #." %'+?#'& !":$?!"2"C#<
>3! 2'!9"# !?<9 'C) 3+"!'#?3C'& !?<9 /; DE*7 @?*"* #." !"%?+!3%'& 3> #." 2?C?2$2 %'+?#'& !'#?3 3>
-,= 'C) '))?CA #." !"<$&#?CA >?A$!"< #3 #." <$2 3> !?<985"?A.#") '<<"#< >3! %!")?# !?<9*
6'#?3
('&%$&'#"
!?<9
***
4!3%"<< 4E1
('&%$&'#"
!?<9
***
Figure 5: Financial Regulation (BASEL II)
based on the corresponding process activities requisite purchase and approve
purchase. Further it refers to the organizational perspective by imposing a mutual exclusion constraint on process resources.
"!4 31/-+20.,/0+* ) ('&,2,/0+* +% $1/#
,/" 5"340. +/0 3"87646260.4 2/" 5731/-4" 0* )00(4
03 4"3'61"4 4/07&( .02 %" 2/" 5"340. +/0 -5530'"4
2/" 5731/-4"#
!"340.
!"340.
$"8764260.
5731/-4"
95530'"
5731/-4"
Figure 6: Authorization
Service Level Agreements (SLAs): The following constraint (cf. Fig. 7-R5) was taken
from http://aws.amazon.com/ec2-sla/. Its linkage refers to the corresponding process via
structural pattern file claim, to the data perspective for the decision constraint
ServiceYear-Unavailable < 99.5, and to the time perspective via restrictions
over the trailing. Further (SLAs) can be found in regulation packages for internal quality
such as Six Sigma or ITIL.
642
/$#-!+$ )$-$> <9#$$7$5& 3 10&!7$ <9#$$7$5&
! %>;85#&2 %$/ -('& $ %'$(# $/+ 8(#& 8)&(2 !//>$' DB8(#& @&2%&/8$=& 5:&2 8)& 82$('(/= 741
.$+; .25B; ,&'5H FF"F1C" !//>$' DB8(#& @&2%&/8$=& (; %$'%>'$8&. ,+ ;>,82$%8(/= -25#
A??C 8)& B&2%&/8$=& 5- 1 #(/>8& B&2(5.; .>2(/= 8)& <&2:(%& 9&$2 (/ H)(%) !#$65/ 30J
H$; (/ 8)& ;8$8& 5- I&=(5/ D/$:$('$,'&" G- +5> )$:& ,&&/ >;(/= !#$65/ 30J -52 '&;; 8)$/
741 .$+;E +5>2 <&2:(%& 9&$2 (; ;8('' 8)& B2&%&.(/= 741 .$+; ,>8 $/+ .$+; B2(52 85 +5>2 >;& 58)& ;&2:(%& H('' ,& .&&#&. 85 )$:& )$. A??C I&=(5/ !:$('$,('(8+"
5- 8)& ;&2:(%& H('' ,& .&&#&. 85 )$:& )$. A??C I&=(5/ !:$('$,('(8+"
/$#-!+$.$,#
15,-,!>@ ? ==@4;
"""
*('& %'$(#
*
*
('&%$#"!:$8
Figure 7: Uptime Agreement
4.2 Evaluation
The feasibility of the unified representation is evaluated by means of the overall 16 process
constraints and along the constraint properties described in Section 2. For this we classify
the 16 process constraint into different constraint types that embody a combination of
properties as represented by the unified representation.
As depicted in Tab. 1 we identified 14 different process constraint types. The first two
constraint types deal with Resource Attribution (e.g., roles, actors, nodes) and Timing Information Attribution (e.g., minimal, maximal, average duration) to process structure. All
60
Table 1: Constraint Types
this information is necessary at run-time, for a temporal monitor as a basis to verify if a
process behaves correctly, or an RBAC system to select actors in a worklist. The only
difference between these two is, that during change-time timing information may have to
be adapted to account for the duration of changes. E.g. constraint C11 in Fig. 1 can
represented as follows:
LinkageC11
SPC11
ConditionC11
BehaviorC11
: ((Invasive Surgery, ALL), SPC11 , ∅)
: ∃a1 Is(a1 , affirm diagnosis)
:∅
: {ROLE := Doctor}
with
For attribution TPc is always empty, as the matched structural pattern implies, that the
attribution is available all the time.
The third constraint type describes Business Compliance Constraints in general. We decided to include this very generic constraint type (which should not be confused with
several of the other constraint types) to show some specific characteristics we derived by
studying constraint sources. Business compliance constraints, are exclusively compliance
constraints, they never carry a behavioral part, they most of the time verify a certain structure of the process, with a possibility of checking also data, resource and temporal aspects.
Therefore the design-time Structural constraints also in this list are Business Compliance
Constraints, as well as some Temporal Constraints or Data constraints (whenever only
design-time aspects are covered). Separation / Binding of Duty, Resource Attribution,
and Timing Information Attribution are the exception to the rule. They may be very well
applied at design time, just only the impact can only be seen at run-time.
61
Figure 8: Unified Constraint Optimization and Checking
4.3 Application Scenarios for Integrated Process Constraint Representation
Fig. 8 illustrates how presented approach can streamline the maintenance and evolution of
new process constraints within Process-Aware Information Systems (PAIS).
• Process constraints can be identified based on structural patterns and enriched with
process context. At some point it may be necessary to create new processes (structural constraints often trigger the creation or evolution of existing processes).
• Constraints can be expressed and stored in an integrated manner based on the representation presend in this paper.
• Based on the integrated representation, constraint consistency checking and optimization can be supported by an automatic process. Because all constraints are
available in a single consistent representation, (unlike for current approaches) it is
easily possible to check if inconsistencies between constraints exists (particularly if
they stem from different sources), rather than checking consistency between pure
structural process constraints only. In current approaches the constraints are embedded and encapsulated in different components of the PAIS (e.g., process engine and
organizational model) such that integrated checks are not possible.
• Constraint enactment is also different from current approaches. As depicted in Fig. 8
step (1) the Linkage can be matched on behalf of the PAIS, the checking of condition
and the enactment of behavior can be move to (potentially external) components.
5 Related Work
Many approaches for checking BPC either at design or run time exist, e.g., [LSG07,
SGN07]). As argued before, an integrated constraints representation has been outside
62
Representation
DECLARE
[PSvdA07]
control flow pat.
BPMN-Q [AWW11]
Linkage
SeaFlows
[LRD10]
activity set
Condition
Behavior
data
–
–
–
data
–
control flow pat.
global scope
Integrated Representation
control flow pat.
process context
trigger position
data, time, resource
!
Table 2: Comparison of Existing Approaches
the scope of these approaches so far. Validation of the integrated representation compared
to selected existing approaches is presented in Table 2.
The above mentioned approaches try to ensure BPC either a-priori at design time or by
detecting inconsistencies at run-time. In addition, there are also a-posteriori approaches
that offer techniques to analyze process logs (i.e., data of already executed and finished
processes) with respect to certain properties such as adhering to compliance constraints
[vdAdBvD05]. In these approaches, different aspects logged during process execution can
be checked, e.g., separation of duties. Doing so the a-posteriori approach reflects the need
for unified compliance checking, not only a-posteriori, but also a-priori. Other approaches
use constraints to synchronize between business processes of different type (e.g., between
a chemotherapy and a radiation process) [Hei01]. Based on our approach, synchronization
constraints can be unified and managed in combination with the other constraints in PAIS.
6 Summary and Outlook
This paper introduced a framework for the integrated representation of process constraints
in PAIS. First of all, the basic concepts of the framework were introduced, i.e., Linkage
that contains the structural pattern of the process the constraint refers to, the Context, i.e.,
the process types or instances for which the constraint is imposed, and the Trigger Position that equips the constraints with actions that migth be triggered by the constraint, e.g.,
for synchronization. Further, the framework was enriched by coping with data, time, and
resource information provided within the constraints. Finally Behavior was specified that
serves for resource attribution or exception handling, for example. Based on different realworld examples the feasibility of the representation has been shown. Further perspectives
on the usage of the presented framework for consistency checks and compliance verification were discussed. Finally, it was shown how existing formalisms for specifying process
constraints can be expressed by the representation framework at hand.
Currently we are working on an implementation representation framework on top of our
adaptive process engine CPEE. Particular focus will be put on maintenance of the constraint base and an integrated verification component for the engine.
63
References
[AWW11]
A. Awad, M. Weidlich, and M. Weske. Visually specifying compliance rules and
explaining their violations for business processes. Journal of Visual Languages &
Computing, 22(1):30–55, 2011.
[Bas06]
Basler Ausschuss fuer Bankenaufsicht. Internationale Konvergenz der Eigenkapitalmessung und Eigenkapitalanforderungen, 2006.
[Hei01]
C. Heinlein. Workflow and Process Synchronization with Interaction Expressions
and Graphs. In Proc. Int’l Conf. on Data Engineering, page 243–252, 2001.
[HG09]
Barbara Halle and Larry Goldberg. Business Process Models and Business Rules:
How They Should Work Together. In The Decision Model: A Business Logic Framework Linking Business and Technology. Taylor & Francis, LLC, 2009.
[KLR+ 10]
D. Knuplesch, L.T. Ly, S. Rinderle-Ma, H. Pfeifer, and P. Dadam. On Enabling DataAware Compliance Checking of Business Process Models. In Int’l Conf. Conceptual
Modeling, volume 6412, pages 332–346, 2010.
[LRD10]
L.T. Ly, S. Rinderle-Ma, and P. Dadam. Design and Verification of Instantiable Compliance Rule Graphs in Process-Aware Information Systems. In Proc. Int’l Conf. on
Advanced Systems Engineering, pages 9–23, 2010.
[LRGD09]
T. Ly, S. Rinderle-Ma, K. Göser, and P. Dadam. On Enabling Integrated Process
Compliance with Semantic Constraints in Process Management Systems - Requirements, Challenges, Solutions. Information Systems Frontiers, pages 1–25, 2009.
[LSG07]
R. Lu, S. Sadiq, and G. Governatori. Compliance Aware Process Design. In Proc.
Int’l Business Process Management Workshops ’07, 2007.
[MRM11]
Juergen Mangler and Stefanie Rinderle-Ma. IUPC: Identification and Unification of
Process Constraints. CoRR, abs/1104.3609, 2011.
[NS07]
K. Namiri and N. Stojanovic. Pattern-Based Design and Validation of Business
Process Compliance. In Int’ OTM Conferences 2007, Part I, volume 4803 of LNCS,
page 59–76. Springer, 2007.
[PSvdA07]
M. Pesic, H. Schonenberg, and W.M.P. van der Aalst. DECLARE: Full Support for Loosely-Structured Processes. In Int’l Enterprise Computing Conference
(EDOC’07), page 287–300, 2007.
[PT09]
M. Peleg and S. W. Tu. Design patterns for clinical guidelines. Artificial Intelligence
in Medicine, 47(1):1–24, 2009.
[SGN07]
S. Sadiq, G. Governatori, and K. Naimiri. Modeling Control Objectives for Business
Process Compliance. In Proc. BPM ’07, pages 149–164, 2007.
[SM10]
M. Strembeck and J. Mendling. Generic Algorithms for Consistency Checking of
Mutual-Exclusion and Binding Constraints in a Business Process Context. In OTM
Conferences (1), pages 204–221, 2010.
[U.S03]
U.S. Department of Health and Human Services. 7th Report of the Joint National
Committee on Prevention, Detection, Evaluation, and Treatment of High Blood Pressure. Technical Report 03-5233, National Heart Lung and Blood Institute, 2003.
[vdAdBvD05] W. van der Aalst, H. de Beer, and B. van Dongen. Process Mining and Verification of Properties: An Approach based on Temporal Logic. In Int’l OnTheMove
Conferences, volume 3761, page 130–147, 2005.
64
Workflow Nets with Roles
Robin Bergenthum, Jörg Desel, Sebastian Mauser
FernUniversität in Hagen, Germany
firstname.lastname@fernuni-hagen.de
Abstract: We formalize the usual static role concept for workflow nets, introduce
dynamic roles and define soundness as well as a second correctness criterion, called
consistency, for workflow nets with roles. We study the relation between the notions
of consistency and soundness of workflow nets with and without roles. In particular,
we show that a sound workflow net extended by consistent roles is again sound.
1 Introduction
The focus of formal business process modeling and related analysis methods is mainly
on the control flow perspective. However, there is also research [HSV06, JKJ10, BBS07,
RAHE05, PA07, RAH09, Pri08] on the resource perspective of business processes, i.e.
on the people, systems and machines actually doing the work. In this paper we formally
model human resources, i.e. the actors enabling a business process. In this context, a central question is which actors are allowed to work on a given task of a process. Similar to the
concept of role-based access control (RBAC) in the domain of IT-security [SCFY96], business process modeling languages as well as industrial workflow systems usually consider
roles as an intermediary concept between tasks and actors [TCG04, RAHE05, AH02]. The
permission to execute a task is associated with one or more roles, and conversely also actors are assigned to roles, thereby acquiring the permissions of the roles. Roles are mostly
determined by the organizational units (also called groups) and the positions within an
enterprise, but also skills and responsibilities are regarded. Typical modeling constructs
for assigning roles to tasks are swimlanes referring to roles as in BPMN and EPCs or an
annotation of tasks with roles as common for workflow (Petri) nets [AH02, Aal98, Wes07].
In this paper, we provide a formal semantics of the usual role concept for workflow nets
by a translation into Colored Petri nets (CP-nets) [Jen97]. There already exist papers that
show how to model resource allocation and the handling of resources available for process
instances by CP-nets [PA07, RAH09] (there is also some work considering nets in nets
[Pri08] and plain Petri nets [JKJ10, HSV06]). These papers focus on the distribution of
work items, resource allocation patterns, resource management, etc. These aspects are
important for workflow systems, but they go beyond the role concept of workflow nets.
Moreover, we do not restrict our considerations to static roles but also define a formal semantics for workflow nets with dynamic roles which have informally been introduced in
[BDHM11] in the context of learning processes. In learning processes it is common that
the learners dynamically change their roles. Depending on executed learning tasks they
gain additional knowledge or responsibilities and therefore their role in the process can
65
change which then influences the allocation of further tasks. Besides the intended applicability for learning and teaching processes, workflow nets with dynamic roles are useful
for formally modeling and analyzing business processes with complex authorization constraints. In particular, the principles of separation of duty (SoD) and binding of duty (BoD)
[TCG04] can nicely be represented by dynamic roles. For instance, when writing a review
an actor gets a new role which is later on required for presentation of the review (BoD)
but prohibits writing a second review (SoD). In the literature, there are several interesting
approaches, mostly based on RBAC concepts [SCFY96], which introduce specific additional model constructs and notations for explicitly representing authorization constraints
in the context of workflows, e.g. [TCG04].
The definitions of workflow nets with static and dynamic roles in this paper enable formal
methods. In particular, it is possible to define behavioral correctness criteria regarding
these role concepts. The most popular correctness criterion for workflow nets is soundness
[Aal98, Wes07]. We first extend this notion in a natural way to workflow nets with static
and dynamic roles. Then, we introduce a second correctness criterion, called consistency,
which requires that the role perspective does not influence the control flow of a model.
We briefly show that, in the case of static roles, soundness and consistency have a very
simple and intuitive meaning, namely a workflow net with static roles is consistent iff, for
each non-dead task, there is at least one actor having one of the roles assigned to the task,
and it is sound iff it is consistent and the underlying workflow net is sound. Concerning
workflow nets with dynamic roles the notions of soundness and consistency become more
difficult. We show that a consistent model is sound iff the underlying workflow net is
sound. Counterexamples show that further relations between consistency and soundness
do not hold. Lastly, we discuss a characterization of consistency for dynamic roles which
regards the role perspective separately from the underlying workflow net.
The paper is organized as follows. In Section 2 we recapitulate the standard role concept
for workflow nets and define a formal semantics for this concept using CP-nets. In Section
3 we formally introduce and define dynamic roles for workflow nets. In Section 4 we
discuss soundness and consistency for workflow nets with static and with dynamic roles.
2 Standard Role Concept
The role concepts of most business process modeling languages are similar. They basically
allow an assignment of users and tasks to roles. This is also the case for the standard role
concept of workflow nets [AH02, Aal98]. This section presents a formal semantics of this
concept by translating it into a CP-net-notation. We assume that the reader is familiar with
workflow nets and CP-nets. We use the following notations. N denotes the non-negative
the powerset of A and NA the set of multisets over
integers. For a finite set A, 2A denotes
"
A
A. For m ∈ N we write m = a∈A m(a) · a. If m(a) > 0 we write a ∈ m.
Definition 1. A workflow net (WF-net) is a tuple N = (P, T, F, i, f ), where
•
•
•
•
P and T are finite sets of places and transitions fulfilling P ∩ T = ∅,
F ⊆ (P × T ) ∪ (T × P ) is a flow relation,
i, f ∈ P are places satisfying (T × {i}) ∩ F = ∅ and ({f } × T ) ∩ F = ∅,
for any node n ∈ P ∪ T there exists a path from i to n and a path from n to f .
66
Figure 1: A WFR-net.
The behavioral semantics, i.e. the occurrence sequences, of a WF-net is given by considering the corresponding marked Petri net with the initial marking 1 · i.
Definition 2. A workflow net with roles (WFR-net) is a tuple N R = (N, R, A, l, r), where
•
•
•
•
•
N = (P, T, F, i, f ) is a WF-net,
R is a finite set of roles,
A is a finite set of actors,
l : T → 2R is a labeling function assigning a set of roles to each transition,
r : A → R is a function assigning a role to each actor.
This definition represents roles and actors in an intuitive and simple way. It is similar to
most standard role concepts for business processes. Each actor having one of the roles
associated to a transition is permitted to execute a respective task. Unlabeled transitions
(having assigned the empty set) can occur automatically without requiring an actor.
Figure 1 shows an example workflow net with roles. There are four actors, one with role
R1, two with role R2 and one with role R3. Task A has to be done by an actor with role
R1. Then, B and C can be executed concurrently by R2-actors. Afterwards, D requires the
role R3. Finally, E can be accomplished by an actor with role R1 or with role R2.
To provide a formal operational semantics for WFR-nets, we define a translation into CPnets. We extend the underlying WF-net by a place which serves as a resource pool, containing all actors with their associated roles. The role annotations of transitions are considered
by appropriate guards. In this way roles, actors and execution permissions are regarded.
We here just provide the definition of a CP-net [Jen97] (black tokens are given by the color
set U N IT = {()}). For the operational semantics of CP-nets, see [Jen97].
Definition 3. A colored Petri net (CP-net) is a tuple CP N = (C, P, T, F, V, c, v, g, e,
m0 ), where
•
•
•
•
•
•
•
•
•
C is a finite set of non-empty types (each type is a set called color set),
P and T are finite sets of places and transitions fulfilling P ∩ T = ∅,
F ⊆ (P × T ) ∪ (T × P ) is a flow relation defining a set of arcs,
V is a finite set of variables,
c : P → C is a coloring function assigning a type to every place,
v : V → C is a coloring function assigning a type to each variable,
g assigns a boolean expression using variables from V to every transition,
e assigns an!expression of type Nc(p) using appropriate variables from V to every arc,
m0 : P → p∈P Nc(p) assigns an initial marking m0 (p) ∈ Nc(p) to every place p.
67
Figure 2: CP-net corresponding to a WFR-net.
Definition 4. Given a WFR-net N R = (P, T, F, i, f, R, A, l, r), we define the corresponding CP-net CP NN R = (C, P & , T, F & , V, c, v, g, e, m0 ) by
•
•
•
•
•
•
•
•
C = {U N IT, A, R, A × R},
P & = P ∪ {pres },
F & = F ∪ ({t ∈ T | l(t) 3= ∅} × {pres }) ∪ ({pres } × {t ∈ T | l(t) 3= ∅}),
V = {x, y},
c(p) = U N IT for p ∈ P , c(pres ) = A × R,
v(x) = A, v(y) = R,
g(t) = [y = g1 orelse . . . y = gn ] for t ∈ T satisfying l(t) = {g1 , . . . , gn },
e(z) = 1 · () for z ∈ F , e(z) = (x, y) for z ∈ ({t ∈ T | l(t) 3= ∅} × {pres }) ∪
({pres } × {t ∈ T | l(t) 3= ∅}),
"
• m0 (i) = 1 · (), m0 (p) = ∅ for p ∈ P \ {i}, m0 (pres ) = a∈A (a, r(a)).
In order to fire a transition of the introduced CP-net, the variables x and y have to be bound.
In this way an actor x having the role y is allocated to the task. The transition guard ensures
that the allocated actor is permitted to execute the task, i.e. it is checked that his role y is
assigned to the task in the WFR-net. When an actor executes a task, he is removed from
the place pres . As soon as the task is accomplished, the actor is released by giving it back
to the place pres . In this way the place pres guarantees that at any time an actor can only
be allocated to one task. An exception are tasks of the original WFR-net having an empty
set of roles. These automatic tasks require no actor from the resource pool pres . Note that
the concept of releasing an actor allocated to a task as soon as the task is completed is an
important difference to the approaches in [HSV06, JKJ10, BBS07], where also WF-nets
regarding resources are formally discussed. Figure 2 depicts the CP-net corresponding to
our example WFR-net from Figure 1.
We have restricted ourselves to a basic role model in this section. This basic model can be
extended in different directions:
• The basic role model does not regard process instances. However, different process
instances can easily be distinguished by using a copy of the WFR-net for each instance and connecting each copy with the place representing the resource pool. Then,
this place does not only ensure that an actor cannot execute two tasks of one process
68
instance at once but also that an actor cannot execute two tasks of different process
instances at once. One resource place can also be used for different process models.
Then, the same actors are shared among the process instances of several processes.
• In the basic role model an actor can only be assigned to one role, and we do not consider a hierarchy among roles. While these restrictions reduce the modeling comfort,
they do not restrict the modeling capabilities, since these aspects can equivalently be
represented by the concept of alternative role annotations. For explicitly modeling
multiple roles of actors, we can associate sets of roles instead of single roles to each
actor. Then, the transition guards check whether a specific role is contained in the
set of roles of an actor. A hierarchy among roles can be expressed by a consistency
condition on the function assigning sets of roles to actors.
• The basic model does not regard collaborative tasks which require a joint execution by
several actors with certain roles, e.g. two authors that write a paper together. We can
extend the role model by collaborative tasks as follows (see [BDHM11] for details).
Instead of one actor, a collaborative task consumes a certain number of actors from the
resource pool. To ensure that each of the actors has an appropriate role, we can use the
“andalso”-operation in the guards.
3 Dynamic Roles
The idea of dynamic roles is to change role assignments of actors depending on tasks
executed by an actor, i.e. the role of an actor depends on his task history. We will extend
the intuitive modeling language of WFR-nets by this concept. If an actor having a certain
role changes his role when executing a task, we represent this by extending the label of the
transition. Formally, we simply consider pairs of roles consisting of the old and the new
role, i.e. the labeling function now assigns a set of pairs of roles instead of a set of single
roles to the transitions. If a role assignment does not change when executing a task, the
old and the new role coincide.
Definition 5. A workflow net with dynamic roles (WFDR-net) is a tuple NR= (N, R,
A, l, r), where
•
•
•
•
•
N = (P, T, F, i, f ) is a WF-net,
R is a finite set of roles,
A is a finite set of actors,
l : T → 2R×R is a labeling function assigning a set of pairs of roles to each transition,
r : A → R is a function assigning a standard role to each actor.
Figure 3 shows an example workflow net with dynamic roles. There are two actors, both
initially having the role R1. Both are allowed to execute task A. Also the concurrent tasks
B and C require the role R1, i.e. B as well as C can be accomplished by any of the two
actors. However, one and the same actor cannot execute both tasks (separation of duty),
since each of them causes a role change. The actor executing B gets the role R2, the one
executing C gets the role R3. Therefore, the two actors have to share the two tasks among
each other. After B and C, task D can be done by either an R2-actor or an R3-actor, i.e. by
any of the two actors. Finally, E has to be executed by an actor with role R2. Thus, this
task requires the actor that executed B before (binding of duty).
69
Figure 3: A WFDR-net.
Analogously to the last section, we define the semantics of a WFDR-net by a translation
into a CP-net. In addition to Definition 4, the dynamic roles of actors are regarded. When
firing a transition three variables have to be bound. Besides the actor x, the variable y1
represents the role required to execute the task and y2 represents a role change.
Definition 6. Given a WFDR-net NR= (P, T, F, i, f, R, A, l, r), we define the corresponding CP-net CP NN R = (C, P & , T, F & , V, c, v, g, e, m0 ) by
•
•
•
•
•
•
•
C = {U N IT, A, R, A × R},
P & = P ∪ {pres },
F & = F ∪ ({t ∈ T | l(t) 3= ∅} × {pres }) ∪ ({pres } × {t ∈ T | l(t) 3= ∅}),
V = {x, y1 , y2 },
c(p) = U N IT for p ∈ P , c(pres ) = A × R
v(x) = A, v(y1 ) = R, v(y2 ) = R,
g(t) = [(y1 = g1,1 andalso y2 = g1,2 ) orelse . . . (y1 = gn,1 andalso y2 = gn,2 )] for
t ∈ T satisfying l(t) = {(g1,1 , g1,2 ), . . . (gn,1 , gn,2 )},
• e(z) = 1 · () for z ∈ F ,e(z) = (x, y1 ) for z ∈ {pres } × {t ∈ T | l(t) 3= ∅},
e(z) = (x, y2 ) for z ∈ {t ∈ T | l(t) =
3 ∅} × {pres },
"
• m0 (i) = 1 · (), m0 (p) = ∅ for p ∈ P \ {i}, m0 (pres ) = a∈A (a, r(a)).
Figure 4 shows the CP-net corresponding to the WFDR-net from Figure 3.
Figure 4: CP-net corresponding to a WFDR-net.
70
4 Soundness
In this section we discuss soundness of WFR-nets and WFDR-nets. A net may exhibit
errors such as deadlocks, livelocks or garbage being left in the process after termination.
Soundness is a basic behavioral property that each proper procedure should satisfy. For
WF-nets the property states that “for any case, the procedure will terminate eventually, and
at the moment the procedure terminates there is a token in place f and all the other places
are empty” [Aal98]. Moreover, there should be no dead tasks.
Definition 7. A WF-net N = (P, T, F, i, f ) is sound if
(S1) From each marking reachable from 1 · i, the marking 1 · f is reachable.
(S2) There are no dead tasks w.r.t the initial marking 1 · i.
As an example, the WF-net underlying the nets shown in Figure 1 and 3 is sound. Remark that the soundness definition originally included a third property stating that for each
marking m reachable from 1 · i with m(f ) ≥ 1, there holds m = 1 · f . However, it was
later shown that this property follows from (S1).
We now define a notion of soundness of WFR- and WFDR-nets which integrates the resource perspective of business processes. For this purpose we consider the corresponding
CP-nets and formulate two requirements generalizing the properties (S1) and (S2).
Definition 8. Let NR= (N, R, A, l, r) be a WFR- resp. a WFDR-net and CP NN R =
(C, P & , T, F & , V, c, v, g, e, m0 ) the corresponding CP-net. A marking mf of CP NN R is
called final marking if mf (f ) = 1 · () and mf (p) = ∅ for p ∈ P \ {f }
Definition 9. Let NR= (N, R, A, l, r) be a WFR- resp. a WFDR-net and CP NN R =
(C, P & , T, F & , V, c, v, g, e, m0 ) the corresponding CP-net. The net NR is sound if
(S1’) From each reachable marking of CP NN R , a final marking is reachable.
(S2’) CP NN R has no dead tasks.
It can easily be verified that the nets in Figure 2 and 4 fulfill the properties (S1’) and (S2’).
That means, the WFR-net in Figure 1 and the WFDR-net in Figure 3 are sound.
In WFR-nets and WFDR-nets the resource perspective is defined on top of the control flow
perspective. The latter is given by a WF-net and the former by role annotations and actors.
This concept clearly separates the resource view from the control flow view. Therefore,
the actors and roles of a WFR-net resp. a WFDR-net should have no influence on the
control flow of the model. They should only describe the resource allocation allowed
within the business process. Otherwise, the two views are not consistent. Consequently,
besides soundness we in the following introduce a second correctness property which is
concerned with consistency of the resource perspective to the control flow perspective.
By requiring actors with certain roles for firing transitions, the control flow can only
be restricted. That means, if a transition can occur in a certain marking of a WFR-net
resp. a WFDR-net this transition is also enabled in the corresponding marking of the underlying WF-net where the corresponding marking is given by just neglecting the place
pres . Thereby, a transition occurrence of a CP-net, called binding element, is given by
71
a pair consisting of the fired transition and the firing mode (i.e. the binding of the variables) of the transition [Jen97]. For instance, in the net of Figure 4 the binding element
(A, < x = andy, y1 = R1, y2 = R2 >) is enabled.
Definition 10. Let NR= (N, R, A, l, r) be a WFR-net resp. a WFDR-net, CP NN R =
(C, P & , T, F & , V, c, v, g, e, m0 ) the corresponding CP-net and m a marking of CP NN R .
Then the marking mu of N given by mu (p) = |m(p)| for p ∈ P is called corresponding
marking of m.
Lemma 1. Let m be a marking of CP NN R . If the binding element (t, b) is enabled in m
leading to the follower marking m& , then t is enabled in the marking mu of N leading to
u
the follower marking m& .
Proof. By construction, compared to N , the CP-net CP NN R just has the additional place
pres (together with arcs connecting the place with transitions). A well-known property of
Petri nets is that adding a place to a Petri net can only restrict the enabledness of transitions
and does not influence the markings w.r.t. other places.
Figure 5: WFDR-net which is sound but not
consistent.
It is possible that an enabled transition of N is
prohibited in NR, since appropriate actors are
missing, i.e. the place pres prohibits the occurrence. If behavior specified by the control
flow model cannot occur due to missing actors,
this indicates inconsistency between the control
flow and the resource perspective, i.e. the resources are not appropriate for the given control flow. Therefore, we formulate the following
correctness criterion which ensures the reverse
implication to Lemma 1.
Definition 11. Let NR= (N, R, A, l, r) be a WFR- resp. a WFDR-net and CP NN R =
(C, P & , T, F & , V, c, v, g, e, m0 ) be the corresponding CP-net. The net NR is called consistent if for each reachable marking m of CP NN R and each transition t which is enabled
in the marking mu of N , there is a binding b such that (t, b) is enabled in m.
Using Lemma 1 we can summarize the following relationship between the behavior of a
WFR-net resp. a WFDR-net and the underlying WF-net in terms of occurrence sequences.
Lemma 2. If an occurrence sequence (t1 , b1 ) . . . (tn , bn ) is enabled in the initial marking
m0 of CP NN R leading to the follower marking m, then t1 . . . tn is enabled in the initial
marking 1 · i of N leading to the follower marking mu . In the case NR is consistent we
also have: If t1 . . . tn is enabled in the initial marking 1 · i of N , then there are bindings
b1 , . . . , bn such that (t1 , b1 ) . . . (tn , bn ) is enabled in the initial marking m0 of CP NN R .
Proof. The statements follow from mu0 = 1 · i by inductively applying Lemma 1 and
Definition 11.
72
The WFR-net of Figure 1 and the WFDR-net of Figure 3 are both consistent. Figure 5
shows a WFDR-net which is not consistent. However, this net and also its underlying WFnet are sound. The net contains two subsequent alternatives, first between task A and B,
then between task C and D. The role annotations ensure that whenever the actor initially
executes A, then he next has to execute C, although D is also enabled in the underlying
WF-net. Similarly, when starting with B, the actor then has to execute D. Therefore, the
role perspective of this net forbids behavior which is allowed by the control flow model,
namely the occurrence sequences AD and BC.
4.1
WFR-nets
An important observation is that for WFR-nets the correctness property of consistency is
already included in the soundness property.
Lemma 3. A sound WFR-net NR= (N, R, A, l, r) is consistent.
Proof. If NR is not consistent, there is a reachable marking m of CP NN R and a transition
t such that t is enabled in the marking mu of N , but for each b the binding element (t, b) is
not enabled in m. It follows that the place pres prohibits the firing of t. Since the marking
of pres never changes, the transition t is dead in CP NN R , i.e. NR does not fulfill (S2).
Consequently, NR is not sound.
Since soundness of WF-nets is well investigated, we discuss the relation between soundness of a WFR-net and soundness of the underlying WF-net. It can first be shown that
soundness of the underlying WF-net is a necessary condition for soundness of a WFR-net.
Lemma 4. If NR= (N, R, A, l, r) is a sound WFR-net, then also N is sound.
Proof. If N is not sound, one of the conditions (S1) or (S2) is not satisfied. We show for
each case that NR is not sound.
If N does not fulfill (S1), then there is a reachable marking m from which 1 · f is not
u
reachable. Either a marking m& with m& = m is reachable in CP NN R or this is not the
case. In the first situation, by Lemma 1, a final marking is not reachable from m& , since
a final marking mf fulfills muf = 1 · f . That means NR does not fulfill (S1’). In the
second case, by Lemma 2, NR is not consistent and therefore Lemma 3 shows that NR is
not sound.
If N does not fulfill (S2), then there is a dead task. By Lemma 2 this task is also dead in
NR, i.e. NR does not fulfill (S2’).
In general, consistency does not imply soundness, since it formulates no requirements on
the control flow of the underlying WF-net. Soundness of the underlying net also does not
imply soundness of a WFR-net because the WFR-net can contain dead tasks due to missing
actors. However, we now show that both properties together, consistency and soundness
of the underlying WF-net, ensure soundness of a WFR-net.
Lemma 5. Let NR= (N, R, A, l, r) be a WFR-net. If NR is consistent and N is sound,
then also NR is sound.
73
Proof. If NR is not sound, one of the conditions (S1’) or (S2’) is not satisfied. We show
for each case that either NR is not consistent or N is not sound.
If NR does not fulfill (S1’), then there is a reachable marking m of CP NN R from which
no final marking is reachable. By Lemma 2, the marking mu is reachable in N . In the
case NR is consistent, with Lemma 1 it follows that the marking 1 · f is not reachable from
mu , since a marking mf of CP NN R fulfilling muf = 1 · f is a final marking. That means,
N does not fulfill (S1).
If NR does not fulfill (S2’), then there is a dead task. In the case NR is consistent, by
Lemma 2, this task is also dead in N , i.e. N does not fulfill (S2).
The previous lemmas imply the following characterization of soundness for WFR-nets.
Theorem 1. A WFR-net NR= (N, R, A, l, r) is sound iff NR is consistent and N is sound.
This characterization shows how to design sound WFR-nets. First, a sound WF-net is
constructed. Then, roles and actors are added in a way which does not influence the
control flow given by the WF-net.
So far the structural interpretation of the behavioral property consistency is not clear, and
thus we do not know how to ensure the property when designing the role perspective of a
WFR-net. Therefore, we provide a simple characterization of consistency:
Lemma 6. A WFR-net NR= (N, R, A, l, r) is consistent iff, for each transition t with
l(t) 3= ∅ which is not dead w.r.t. N , there is an actor a ∈ A such that r(a) ∈ l(t).
Proof. If NR is not consistent, the proof of Lemma 3 shows that there is a transition t
which is not dead w.r.t. N but cannot fire w.r.t. CP NN R due to the place pres (which has
a constant marking). It follows: l(t) 3= ∅ and there is no a ∈ A such that r(a) ∈ l(t).
If there exists a transition t with l(t) 3= ∅ which is not dead w.r.t. N such that there is no
a ∈ A with r(a) ∈ l(t), then there is a reachable marking m of N which enables t but for
any reachable marking m& of CP NN R and any b the binding element (t, b) is not enabled
in m& because the constant marking of pres prohibits t. Therefore, if a marking m& with
u
m& = m is reachable in CP NN R , then NR is not consistent. Otherwise, by Lemma 2,
NR is not consistent.
From this characterization we can immediately deduce two very simple sufficient conditions for consistency of a WFR-net NR= (N, R, A, l, r) which are purely structural. In
particular, they are completely independent from the underlying WF-net N . In the first
condition we only remove the restriction to dead tasks from the previous characterization.
The second condition is a further simplification abstracting from tasks. It just requires that
for each role there is at least one actor having the role.
(C1) If, for each transition t with l(t) 3= ∅, there is an actor a ∈ A such that r(a) ∈ l(t),
then NR is consistent.
(C2) If, for each role x ∈ R, there is an actor a ∈ A with r(a) = x, then NR is consistent.
The WFR-net from Figure 1 fulfills (C2), since there is an actor for all three roles.
74
Figure 6: Sound WFDR-net.
4.2
WFDR-nets
For WFDR-nets, soundness and consistency become more difficult notions. In contrast to
Lemmas 3 and 4 for WFR-nets, soundness of a WFDR-net does neither imply consistency
of the WFDR-net nor soundness of the underlying WF-net. The WFDR-net of Figure 6 is
sound, although it is not consistent and the underlying WF-net is not sound. The WF-net
can run into a deadlock when firing tasks C and F or D and E. However, when firing C and
E or D and F the net completes properly. Inconsistent role annotations of the WFDR-net
prohibit the deadlocks of the underlying WF-net.
We have already discussed that soundness alone is not enough for a WFDR-net. For correctness, we are interested in sound and consistent WFDR-nets. For such nets it is possible
to show that the underlying WF-net is also sound.
Lemma 7. If NR= (N, R, A, l, r) is a sound and consistent WFDR-net, then N is sound.
Proof. If N is not sound, one of the conditions (S1) or (S2) is not satisfied. We show for
each case that NR is not sound or not consistent.
If N does not fulfill (S1), then there is a reachable marking m from which 1 · f is not
u
reachable. In the case NR is consistent, by Lemma 2, a marking m& with m& = m is
reachable in CP NN R . By using Lemma 1, a final marking is not reachable from m& ,
since a final marking mf fulfills muf = 1 · f . That means NR does not fulfill (S1’).
If N does not fulfill (S2), then there is a dead task. By Lemma 2 this task is also dead in
CP NN R , i.e. NR does not fulfill (S2’).
Moreover, analogously to Lemma 5 for WFR-nets, consistency of a WFDR-net together
with soundness of the underlying WF-net implies soundness of the WFDR-net.
Lemma 8. Let NR= (N, R, A, l, r) be a WFDR-net. If NR is consistent and N is sound,
then also NR is sound.
Proof. The proof is analogous to Lemma 5.
The two previous lemmas together yield the following equivalence.
Theorem 2. A consistent WFDR-net NR= (N, R, A, l, r) is sound iff N is sound.
75
Altogether, we have shown how to design correct WFDR-nets. First, a sound WF-net is
constructed which is then consistently extended by roles and actors. With this approach
soundness of the resulting WFDR-net is guaranteed. Still, there are also sound WFDRnets which are not consistent. For such a net it is possible that the underlying WF-net is
not sound (Figure 6), but it is also possible that the WF-net is sound (Figure 5). For the
sake of completeness note that in the case of a non-sound and non-consistent WFDR-net,
the underlying WF-net can be both sound (Figure 5 with empty set of actors) or not sound
(Figure 6 with empty set of actors).
We now investigate the property of consistency for WFDRnets in detail. The aim is to find a characterization of consistency which explicitly regards the role perspective separately from the underlying WF-net. Then, as it is natural,
consistency for the role perspective can be checked on top of
the control flow perspective. For this purpose, we introduce
the notion of role diagram which describes the dynamic role
behavior of all the actors of a WFDR-net as given by the role
annotations (neglecting the underlying WF-net). Figure 7 ilFigure 7: A role diagram.
lustrates the role diagram of the WFDR-net from Figure 3.
The role diagram of a WFDR-net is a non-deterministic finite automaton which models
the overall dynamic resource perspective of a WFDR-net. States represent different role
combinations of actors, and each transition represents a task that can be executed by a
certain role combination as well as the role change triggered by this task execution.
Definition 12. A non-deterministic finite automaton is a tuple M = (Q, T, δ, q0 ), where
•
•
•
•
Q is a finite set of states,
T is a finite set of input symbols,
δ ⊆ Q × T × Q is a transition relation and
q0 ∈ Q is an initial state.
Definition 13. Let NR= (P, T, F, i, f, R, A, l, r) be a WFDR-net. The role diagram
RN R = (Q, T, δ, q0 ) of NR is defined inductively (Q ⊆ NR ):
"
• q0 = a∈A r(a)
"
• If q ∈ Q, q(x1 ) > 0 and (x1 , x2 ) ∈ l(t) for t ∈ T then q & = x∈R\{x1 ,x2 } q(x) · x +
(q(x1 ) − 1) · x1 + (q(x2 ) + 1) · x2 ∈ Q and (q, t, q & ) ∈ δ.
A WFDR-net is consistent if the resource perspective, i.e. the place pres does not prohibit
any behavior of the underlying WF-net. That means, for each reachable state, if the WFnet allows the occurrence of a task, then there are actors capable of executing the task.
In particular, the enabledness of a task has to be independent from the assignments of
actors to previous tasks. Based on the concept of role diagram we formulate the following
characterization for consistency of WFDR-nets.
Lemma 9. Let NR= (N, R, A, l, r) be a WFDR-net and RN R = (Q, T, δ, q0 ) the role
diagram of NR. Then, NR is consistent iff it fulfills the following property: For each occurrence sequence t1 . . . tn , n ≥ 1, of N and each (q0 , t1 , q1 ) . . . (qn−2 , tn−1 , qn−1 ) ∈ δ
there exists qn ∈ Q such that (qn−1 , tn , qn ) ∈ δ.
76
Proof. If NR is not consistent, there is a marking m of CP NN R reachable by an occurrence sequence (t1 , b1 ) . . . (tn−1 , bn−1 ) and a transition tn which is enabled in the marking mu of N , such that (tn , bn ) is not enabled in m for each binding bn . By Lemma
2, t1 . . . tn is an occurrence sequence of N . Moreover, by
"construction of RN R it holds
(q0 , t1 , q1 ) . . . (qn−2 , tn−1 , qn−1 ) ∈ δ such that qn−1 = (a,x)∈A×R m(pres )(a, x) · x.
Since tn is not enabled in m w.r.t. the place pres , it follows that there is no (a, x1 ) ∈
m(pres ) such that (x1 , x2 ) ∈ l(tn ). Thus, there is no x1 ∈ qn−1 such that (x1 , x2 ) ∈
l(tn ). It follows that there does not exist a state qn ∈ Q with (qn−1 , tn , qn ) ∈ δ.
If there is an occurrence sequence t1 . . . tn of N and (q0 , t1 , q1 ) . . . (qn−2 , tn−1 , qn−1 ) ∈ δ
such that there is no qn ∈ Q with (qn−1 , tn , qn ) ∈ δ, then by construction of RN R and
Lemma 2 there are bindings b1 . . . bn−1 such that (t1 , b1 ) . . . (tn−1 , bn−1 ) is an occurrence
sequence
of CP NN R which leads to a marking m with the following properties: qn−1 =
"
u
(a,x)∈A×R m(pres )(a, x) · x and m is the follower marking of the occurrence sequence
t1 . . . tn−1 of N . By assumption there is no x1 ∈ qn−1 such that (x1 , x2 ) ∈ l(tn ) and thus
there is no (a, x1 ) ∈ m(pres ) such that (x1 , x2 ) ∈ l(tn ). Consequently, there is no bn such
that (tn , bn ) is enabled in m, although tn is enabled in mu , i.e. NR is not consistent.
With Lemma 9, consistency of the role perspective can nicely be checked on top of a given
WF-net by comparing the marking graph of the WF-net and the role diagram.
The maximal occurrence sequences of the WF-net underlying the WFDR-net from Figure
3 are ABCDE and ACBDE. Thus, to verify consistency of the WFDR-net, for each prefix
of these sequences we have to check that the property formulated in the previous lemma
is satisfied by the role diagram of Figure 7. For instance, given the sequence ABCDE,
starting in the initial state of the role diagram there is only one sequence of state transitions
corresponding to ABCD. For the follower state R2+R3, it has to be checked that there is a
state transition given by the task E.
From Lemma 9 we can also deduce reasonable sufficient conditions for consistency of a
WFDR-net NR= (N, R, A, l, r) with role diagram RN R = (Q, T, δ, q0 ) which are more
simple to check. First, a simplification can be achieved in the case of deterministic role
annotations: If the role diagram is deterministic, then it is enough to check whether each
occurrence sequence of N is included in the role diagram. Second, we can consider the
situation that there are always enough actors to perform each task of the net. For this purpose, we regard the set of all states Q& ⊆ Q of the role diagram reachable by an occurrence
sequence of N . That means, Q& represents all reachable role combinations in the resource
place pres . We formulate the following condition which is analogous to (C1) for WFRnets: If for each task, each reachable role combination contains a role which is allowed to
execute the task, then NR is consistent. Moreover, we can simplify this condition analogously as in the case of (C2) for WFR-nets: If each reachable role combination contains
all roles of NR, then NR is consistent. Thereby, the set of all reachable role combinations
Q& can be computed by projecting the states of the product automaton of RN R and the
marking graph of N onto the RN R -component.
(D) If RN R fulfills the property (q, t, q & ), (q, t, q && ) ∈ δ =⇒ q & = q && , then NR is
consistent iff RN R accepts each occurrence sequence of N (when considering all
states Q as final states).
77
(C1’) If for each t ∈ T with l(t) 3= ∅ and each q ∈ Q& , there exists (x1 , x2 ) ∈ l(t) such
that q(x1 ) > 0, then NR is consistent.
(C2’) If for each q ∈ Q& and each x ∈ R there holds q(x) > 0, then NR is consistent.
Since the role diagram of Figure 7 is deterministic, we can verify consistency of the
WFDR-net in Figure 3 by checking condition (D).
5 Conclusion
We have shown how to formally extend Petri net process models by static and dynamic role
concepts. Then, we have discussed correctness of respective models. An important topic
for future research is a detailed discussion of the extensions of the modeling languages
mentioned in Section 2.
References
[Aal98]
W. van der Aalst. The Application of Petri Nets to Workflow Management. The Journal
of Circuits, Systems and Computers, 8(1):21–66, 1998.
[AH02]
W. van der Aalst and K. van Hee. Workflow Management: Models, Methods, and
Systems. MIT Press, 2002.
[BBS07]
K. Barkaoui, R. Benayed, and Z. Sba. Workflow Soundness Verification Based on
Structure Theory of Petri Nets. IJCIS Journal, 5:51–62, 2007.
[BDHM11] R. Bergenthum, J. Desel, A. Harrer, and S. Mauser. Modeling and Mining of Learnflows. In to appear in ToPNoC. Springer, 2011.
[HSV06]
K. van Hee, N. Sidorova, and M. Voorhoeve. Resource-Constrained Workflow Nets.
Fundam. Inf., 71:243–257, 2006.
[Jen97]
K. Jensen. Coloured Petri Nets. Basic Concepts, Analysis Methods and Practical Use.,
volume 1-3 of Monographs in Theoretical Computer Science. Springer, 1992, 1994,
1997.
[JKJ10]
G. Juhás, I. Kazlov, and A. Juhásová. Instance Deadlock: A Mystery behind Frozen
Programs. In Petri Nets 2010, LNCS 6128, pages 1–17. Springer, 2010.
[PA07]
M. Pesic and W. van der Aalst. Modeling Work Distribution Mechanisms using Colored Petri Nets. International Journal on Software Tools for Technology Transfer, 9(34):327–352, 2007.
[Pri08]
O. Prisecaru. Resource workflow nets: an approach to workflow modelling and analysis. Enterp. Inf. Syst., 2(2):101–120, 2008.
[RAH09] N. Russell, W. van der Aalst, and A. ter Hofstede. Designing a Workflow System Using
Coloured Petri Nets. pages 1–24. Springer, 2009.
[RAHE05] N. Russell, W. van der Aalst, A. ter Hofstede, and D. Edmond. Workflow Resource
Patterns: Identification, Representation and Tool Support. In CAiSE 2005, LNCS 3520,
pages 216–232. Springer, 2005.
[SCFY96] R. Sandhu, E. Coyne, H. Feinstein, and C. Youman. Role-Based Access Control Models. IEEE Computer, 29(2):38–47, 1996.
[TCG04]
K. Tan, J. Crampton, and C. Gunter. The Consistency of Task-Based Authorization
Constraints in Workflow Systems. In CSFW-17, pages 155–169. IEEE, 2004.
[Wes07]
M. Weske. Business Process Management – Concepts, Languages and Architectures.
Springer, 2007.
78
Fast Pattern Matching in Conceptual Models
Evaluating and Extending a Generic Approach
Hanns-Alexander Dietrich, Matthias Steinhorst, Jörg Becker, Patrick Delfmann
European Research Center for Information Systems
University of Muenster
Leonardo-Campus 3
48149 Münster
{dietrich | steinhorst | becker | delfmann}@ercis.uni-muenster.de
Abstract: Identifying structural patterns in conceptual models serves a variety of
purposes ranging from model comparison to model integration and exploration.
Although there are a multitude of different approaches for particular modelling
languages and application scenarios, the modelling community lacks an integrated
approach suitable for conceptual models of arbitrary languages and domains.
Therefore, a generic set-theory based pattern matching approach has recently been
developed. To prove that this approach is beneficial in terms of performance, we
conduct a statistically rigorous analysis of its runtime behaviour. We augment the
original approach to include a caching mechanism that further increases performance. We are able to show that the original algorithm is able to identify arbitrary
patterns within milliseconds. The caching extension further increases performance
by up to fifty per cent given the model base and patterns we used.
1 Introduction
Structurally analysing conceptual models serves a wide variety of purposes. In the domain of Business Process Management (BPM) it helps detect process improvement potential [VTM08]. In case of mergers and acquisitions, identifying structural patterns in
process models allows for comparing heterogeneous process landscapes to one another.
Such a comparison can then be used to integrate various process models containing similar or equivalent structures [Di09]. Identifying patterns in conceptual models is also
useful for business process compliance checking [Kn10] as well as determining the syntactical correctness of a particular model [Me07]. In the domain of database engineering,
detecting structures in conceptual models addresses the problem of schema matching and
integration [PS11; RS10]. In all of these application scenarios a manual analysis is extremely costly, since the number of models to be analysed may range in the thousands
[YDG10]. Furthermore, each model may contain hundreds of elements. Structural pattern matching can therefore only be beneficial if it is conducted in an automated or at
least semi-automated way. To that end, we have developed a generic set-theory based
pattern matching approach that is able to find patterns in conceptual models of arbitrary
modelling languages [De10].
79
The approach is based on the idea that any graph-based conceptual model can be regarded as the set of its objects and relationships. The approach was prototypically implemented as a plug-in for a meta-modelling tool that was available from a previous
research project [DK07]. To prove that our approach is beneficial in terms of performance, we conduct a statistically rigorous analysis of its runtime behaviour. We evaluate
the approach on two sets of Event-driven Process Chains (EPC) [Sc00] that were available from previous research projects as well. In total, we analyse process models having
between 20 and 343 elements. The patterns we search for are based on the works of
[Me07] who defines syntactical errors in EPCs and [Va03] who identify workflow patterns. We are able to demonstrate that the generic set-theory based approach allows for
finding structural pattern matches within milliseconds. We augment the approach to
include a caching algorithm that further improves matching performance by up to fifty
per cent given the model base and patterns we used. The algorithm is based on the idea
of storing frequently used sub-patterns in a data structure that allows access in constant
time.
The remainder of the paper is structured as follows. In Section 2, we discuss related
work. Our analysis indicates that there is no generic pattern matching algorithm that
finds matches within milliseconds. In Section 3, we briefly elaborate on the approach to
be evaluated. We continue by explaining the caching mechanism augmenting the original algorithm. In Section 4, we provide a detailed performance analysis of the approach
both with and without the caching algorithm. We conclude by summarizing our main
findings and providing an outlook to future research.
2 Related Work
Identifying structural patterns in conceptual models serves multiple purposes in the
fields of database engineering and business process management. To identify relevant
literature, the keywords schema matching, business process, similarity, merge,
check , comparison , compliance, and exploration were used in combination with
Google Scholar. Primarily, recent literature from the last five years was analysed. We
furthermore included some sources from unstructured search.
In the field of database engineering, various approaches have been proposed addressing
the problem of schema matching and integration. Two or more schemas are taken as
input and semantically as well as structurally equivalent elements are identified to create
an integrated schema as output [RB01]. Some of these approaches assume the underlying schemas to exhibit a tree-like structure [RS10; TC06; Au05] others identify similar
model parts [PS11; SDH08; DHY08].
In the field of business process management pattern matching is closely related to the
topic of model analysis, which serves multiple purposes ranging from model comparison
[Di10; YDG10; DDG09; AVW08; Kü08; VDM08; EKO07; SM07a; VAW06;] and
model merge [La10; WDM10; Di09; SM07b; Gr05; UC04] to model exploration [We11;
Kn10; Sm10; We10; Fa09; Sm09; WHM08; Aw07].
80
All of these approaches address very particular tasks related to the identification of patterns in conceptual models. Furthermore, some of them restrict themselves to specific
modelling languages or do not provide exact matching algorithms but approximationbased heuristics. We therefore argue that the modelling community would benefit from a
more generic pattern matching approach that is not limited to a particular modelling
language or application scenario and that at the same time is able to identify exact pattern matches. Moreover, to be beneficial such an approach has to return results with
good performance. By providing a statistically rigorous performance evaluation, this
paper can therefore be seen as a continuation of the work of [De10].
3 Set-Theory Based Pattern Matching
3.1 Research Design
Our research follows the design science approach [He04] and more specifically the design science research methodology put forth by [Pe07]. Design science is concerned with
the creation of innovative artefacts solving previously unsolved problems. In our case
the artefact constitutes a generic pattern matching algorithm that is able to find structures
in conceptual models created in arbitrary modelling languages. The methodology of
[Pe07] stipulates six steps for the process of conducting design science, namely problem
identification and motivation, definition of the objectives for a solution, design and development, demonstration, evaluation, and communication. In this paper, we conduct a
statistically rigorous performance evaluation of the artefact presented by [De10]. We
assess its capability of efficiently contributing to the analysis of conceptual models. In an
additional development iteration we augment the artefact to include a caching algorithm
that further improves performance.
3.2 Pattern Matching Approach
The idea of this approach is to apply set operations to a set of model elements, representing the model to be analysed. Based on graph theory, the approach recognizes any
conceptual model as a graph G, consisting of vertices V and edges E, where G=(V,E)
with E$V×V. Therefore, the approach distinguishes model objects, representing nodes,
and model relationships, representing edges, interrelating model objects. Starting from a
particular basic set, the approach searches for pattern matches by performing set operations on this basic set. We define the set O of all objects, the set R of all relationships,
and the set E = O ! R to be the basic sets required for this approach. By combining
different set operations, patterns are built up successively. Given a pattern definition, the
matching process returns a set of model subsets representing the pattern matches found.
Every match found is put into a separate subset. In the following, we introduce the available operations of the approach briefly. A detailed formal specification can be found in
[De10].
81
Each operation has a defined number of input sets and returns a resulting set. In the explanation of the operations, we use additional sets (X: arbitrary set of elements; Y: arbitrary set of objects; Z: arbitrary set of relationships) specifying which kinds of inputs an
operation expects.
" ElementsOfType(X,a) returns a set of all elements of X, belonging to the given element type a.
" ElementsWithTypeAttributeOfValue(X,a,b) returns a set of all elements of X whose
type is assigned an attribute a. In addition, the instance of the type attribute has to be
of value b. A type attribute might be the label of an element. The function then returns all elements having a particular label b.
In order to assemble complex pattern structures successively, the following operations
combine elements and their relationships and elements being related, respectively:
" ElementsWith{In|Out}Relations(X,Z) returns a set of sets containing all elements of X
and their {ingoing | outgoing} relationships of Z.
" ElementsDirectlyRelated(X1,X2) returns a set of sets containing all elements of X1 and
X2 that are connected directly via undirected relationships of R, including these relationships. Each inner set contains one occurrence.
" DirectSuccessors(X1,X2) is the directed analogue to ElementsDirectlyRelated.
A further category of operation is needed to build patterns representing recursive structures (e.g. a path of arbitrary length):
" {Directed}Paths(X1,Xn) returns a set of sets containing all sequences with undirected
{directed} relationships, leading from any element of X1 to any element of Xn. The
elements that are part of the paths do not necessarily have to be elements of X1 or Xn,
but can also be of E\X1\Xn. Each path found is represented by an inner set.
" {Directed}Loops(X) is defined analogously to {Directed}Paths. It returns a set of sets
containing all undirected {directed} sequences, which lead from any element of X to
itself.
To avoid infinite sets, only finite paths and loops are returned. To provide a convenient
specification environment for structural model patterns, we define some additional functions that are derived from those already introduced:
" ElementsWith{In|Out}RelationsOfType(X,Z,c) returns a set of sets containing all
elements of X and their {undirected} directed, {incoming | outgoing} relationships of
Z of the type c. Each occurrence is represented by an inner set.
" ElementsWithNumberOf{In|Out}Relations(X,n) returns a set of sets containing all
elements of X, which are connected to the given number n of {undirected} directed
{incoming | outgoing} relationships of R, including these relationships. Each occurrence is represented by an inner set.
" ElementsWithNumberOf{In|Out}RelationsOfType(X,c,n) returns a set of sets containing all elements of X, which are connected to the given number n of {undirected} directed {incoming | outgoing} relationships of R of the type c, including these relationships. Each occurrence is represented by an inner set.
82
" {Directed}PathsContainingElements(X1,Xn,Xc) returns a set of sets containing elements that represent all undirected {directed} paths from elements of X1 to elements
of Xn, which each contain at least one element of Xc. The elements that are part of the
paths do not necessarily have to be elements of X 1 or Xn, but can also be of E\X1\Xn.
Each such path found is represented by an inner set.
" {Directed}PathsNotContainingElements(X1,Xn,Xc) is defined analogously to {Directed}PathsContainingElements. However, it returns only paths that do not contain
any element of Xc.
" {Directed}Loops{Not}ContainingElements(X,Xc) is defined analogously to {Directed}Paths{Not}ContainingElements.
" {Longest|Shortest}{Directed}Path{Not}ContainingElements(X 1,Xn,Xc) is defined
analogously to {Directed}Paths{Not}ContainingElements. However, the function
only returns the shortest or longest paths from elements of X 1 to elements of Xn containing or not containing at least one element of Xc.
By nesting the functions introduced above, it is possible to build structural model patterns successively. The results of each function can be reused adopting them as an input
for other functions. In order to combine different results, the basic set operators union
(!), intersection (#), and complement (\) can generally be used. Since it should be possible to not only combine sets of pattern matches (i.e., sets of sets), but also the pattern
matches themselves (this refers to the inner sets), the approach incorporates additional
set operators. These operate on the inner sets of two sets of sets respectively. The Join
operator performs a union operation on each inner set of the first set with each inner set
of the second set having at least one element in common. The InnerIntersection operator intersects each inner set of the first set with each inner set of the second set. The
InnerComplement operator applies a complement operation to each inner set of the first
outer set combined with each inner set of the second outer set. Only inner sets that have
at least one element in common are considered. As most of the set operations introduced
expect simple sets of elements as inputs, further operators are introduced that turn sets of
sets into simple sets. The SelfUnion operator merges all inner sets of one set of sets into
a single set performing a union operation on all inner sets. The SelfIntersection operator
performs an intersection operation on all inner sets of a set of sets successively. The
result is a set containing elements that each occur in all inner sets of the original outer
set.
DirectSuccessors(ElementsOfType(O, Function),ElementsOfType(O, Event))
The example of an EPC pattern given above illustrates the application of the approach.
This pattern represents all functions that are directly succeeded by an event. The ElementsOfType calls return the set of all functions and events respectively. They take the
basic set O as input which represents the set of all objects contained in the model to be
searched. The second parameter specifies the type of the objects contained in the intermediate result sets. These are then passed on to the DirectSuccessors call returning a set
of sets. Each inner set contains a function, the succeeding event and the edge between
these two objects.
83
Figure 1 illustrates the prototypical implementation of the approach as a plugin for a
meta-modelling tool. The patterns are defined in a specification environment that is integrated into the meta-modelling environment of the tool. The modelling environment
allows for executing the pattern matching algorithm. All elements that belong to a particular pattern match are highlighted. In the example presented in Figure 1 two pattern
matches were found. Each match consists of a function, its succeeding event and the
relationship between these two objects.
Pattern specification environment
First match
Second
match
Figure 1: Application of the pattern matching approach
3.3 Extensions of the approach
Our approach is implemented using the visitor design pattern known from software engineering [Ga95]. A nested pattern definition is represented as a search tree. The algorithm performs a depth-first search on that tree calculating the result of the leaf nodes
first and then returning this intermediate result to the next higher level of the tree. This
structure allows for efficiently caching intermediate results, since a given sub-tree may
appear multiple times within one pattern definition. Before calculating the result on any
given level in the search tree, the algorithm checks a hash table to see if the given subtree has already been calculated. This hash table can be read in constant time, thus allowing for extremely fast access.
84
4 Performance Evaluation
4.1 Model Base
As model base, we chose 19 different EPC models that were available from two previous
research projects in the fields of public administration (nine models) [Be05] and the
retail industry (ten models) [BS04]. In terms of the former, the largest process model
contains 343 elements and describes the process of refunding student travel expenses.
The smallest model consists of 26 elements and depicts the application for a dog licence
fee. As far as the process models of the retail industry are concerned, the largest model
contains 150 elements and depicts the process of goods receipt. The smallest model
consists of 20 elements representing the return of empties. In terms of the structure of the
models, in both model pools the large models contain a disproportionately high number
of split-connectors.
We chose these model pools for two reasons. First, these processes were surveyed in real
life scenarios. Consequently, they allow us to test the pattern matching approach in a
realistic context. Second, these pools explicitly include very small process models having a minimum of 20 elements as well as extremely large models of up to 343 elements.
This range of different model sizes allows us to evaluate the scaling characteristics of the
pattern matching approach. We argue that, other than the complexity of the pattern and
the abovementioned structure of the model, the size of the model is the factor that most
influences runtime performance. Additionally, we argue that most conceptual models
exhibit model sizes that fall into the range we cover here. We therefore conclude that the
results we obtained for EPC models are also representative of other process modelling
languages as well as application domains.
4.2 Patterns
In total we searched for seven patterns in each of the 19 process models. All but one
pattern check the syntactical correctness of EPCs. Although the matching approach can
be used in a variety of scenarios, we restrict ourselves to syntactical soundness checking
as this task requires the most complex patterns. Furthermore, the patterns we searched
for are based on the workflow patterns presented by [Va03] and common syntactical
errors in EPCs described by [Me07].
The first pattern we searched for is the AND might not get control from XOR/OR
(AND)-pattern reported in [Me07]. It returns paths of arbitrary length that begin in an
XOR/OR-split and end in an AND-join which is the successor of an event having no
incoming edges. The exact definition of that pattern is given below. Another very common syntactical error in EPCs constitutes a decision split after an event (DSAE) [Me07].
Other than that, we also searched for the syntactically correct version of that pattern, the
decision split after a function (DSAF). The exact pattern definitions can be found in
[De10]. Furthermore, we searched for connector loops (CL) which are paths of arbitrary
length that start and end in the same element and contain only connectors.
85
DirectedPaths(
COMPLEMENT(
UNION(ElementsOfType(O,'OR'),ElementsOfType(O,'XOR')),
UNION(
SELFUNION(INNERINTERSECT(UNION(ElementsOfType(O,'OR'),
ElementsOfType(O,'XOR')),ElementsWithNumberOfOutRelations(
UNION(ElementsOfType(O,'OR')ElementsOfType(O,'XOR')),1)))
SELFUNION(INNERINTERSECT(UNION(ElementsOfType(O,'OR'),
ElementsOfType(O,'XOR')),ElementsWithNumberOfOutRelations(
UNION(ElementsOfType(O,'OR'),ElementsOfType(O,'XOR')),0)))
)),
INNERINTERSECT(ElementsOfType(O,AND),DirectSuccessors(
ElementsWithNumberOfInRelations(ElementsOfType(O,'Event'),0),
ElementsOfType(O,'AND'))))
We determined the longest sequence of activities (LSA) in each model which we define
as the longest paths not containing connectors. Additionally, we checked if every XORsplit is joined by the appropriate counterpart XOR (XOR) [Me07]. This pattern returns
all paths from XOR-split connectors to XOR-join connectors not containing AND- or
OR-join connectors. We also searched for all paths not containing elements that have a
particular label (PRC). We define this pattern to calculate all paths from the start to the
end elements of the model. From this set of paths we subtract all paths that do contain
elements having that particular label. This second set of paths is calculated by joining the
set of paths from all start to all end elements with the set of elements exhibiting the label.
By defining the pattern in this manner, we calculate the paths from all start to all end
elements twice. In doing so, we are able to demonstrate the effect of the caching mechanism on a computationally expensive operation like a path search. For reasons of brevity
we omit the exact definitions of the CL-, LSA-, XOR-, and PRC-patterns here.
4.3 Evaluation Process
To obtain the performance data in a statistically rigorous manner, we applied the steadystate performance methodology presented by [GBE07]. The methodology was developed
to control measurement errors in Java programs caused by the non-deterministic execution in the virtual machine due to garbage collection, just-in-time (JIT) compilation and
adaptive optimisation. As described above, the pattern matching approach was implemented as a plugin for a meta-modelling tool written in C#. This programming language
suffers from the same measurements errors. Therefore we adapted the approach of
[GBE07] to C#. To compensate for the non-deterministic behaviour of the virtual machine, we chose the so-called steady-state performance measurement determining program performance after JIT compilation and optimisation.
The performance evaluation was conducted on an Intel® Core 2 Duo CPU E8400 3.0
GHZ with 3.25 GB RAM and Windows 7 (32-Bit edition). We disabled the energy saving settings in Windows and executed the process as a real-time process to avoid any
unnecessary hardware slow down or process switching. Prior to each search run a complete garbage collection was forced. In doing so, we further eliminate systematic errors.
For the time measurement we used the high resolution QueryPerformanceCounter-API
in Windows in concert with the adapted methodology.
86
During the initial warm-up phase, the JIT compiler optimized the pattern search algorithm. This implies running the algorithm until the variation coefficient drops under two
per cent [GBE07]. This coefficient is calculated by dividing the standard deviation of all
runs by its mean execution time. This led to searching each pattern in each model at least
five times. Once this warm-up phase was completed and a steady state was reached, the
measurement phase started. During this phase, each pattern was searched in each model
ten times. After ten search runs, the meta-modelling tool was restarted triggering another
ten search runs. This process was repeated 35 times to compensate for the non-deterministic behaviour of the virtual machine [GBE07]. For each process, the mean execution times were determined for each combination of pattern and model with caching
enabled and disabled respectively. This leads to a total of 70 measurements per combination of pattern and model, 35 for caching enabled and 35 for caching disabled.
4.4 Performance Data
We first consider the original algorithm without the caching extension. Runtime measurements are depicted in Figure 2. Each model is represented on the abscissa by its number of elements. In each model all seven patterns were searched for resulting in one bar
for each pattern. The ordinate represents the time unit in milliseconds on a logarithmic
scale. Each measurement depicts the mean execution time of all 35 search processes. In
addition, each measurement exhibits a confidence interval smaller than one per cent
given a confidence level of 99 per cent.
1000000
100000
10000
Milliseconds
1000
100
10
1
0,1
0,01
343
150
142
121
104
89
79
67
67
61
57
50
50
46
42
35
26
24
20
Number of elements
LSA
CL
XOR
PRC
AND
DSAE
DSAF
Figure 2: Performance evaluation of the original algorithm
Figure 2 demonstrates that the majority of patterns are found within less than one
millisecond. In all but six cases results were obtained in considerably less than one second. The data suggest that our pattern matching algorithm returns results with an overall acceptable performance. Furthermore, the figure suggests that the performance of the
matching algorithm depends on three factors, namely the size and the structure of the
model (cf. Section 4.1) as well as the complexity of the pattern. The latter refers to what
and how many functions and operators are used in the pattern definition and to what
extent they are nested. Since the measurements are right-skewed, Figure 2 shows that the
larger the model the longer takes the matching process.
87
The structure of the model refers to its number of split-connectors. A linear sequence of
elements can be more efficiently searched than a highly branched model. This explains
the execution times observed in the third largest model representing a procurement process in the retail industry. This process model contains a large number of splitting XOR
connectors, thus prolonging the matching process. In terms of the pattern complexity, in
all but one model the LSA pattern took the longest to compute. Furthermore, identifying
events or functions that are followed by an XOR-split connector can be computed in less
than one millisecond in all models except the largest one. The interaction of these three
factors influences runtime performance. Although overall performance is acceptable,
Figure 2 indicates that in case of large, highly branched models in combination with
complex patterns the execution time increases exponentially. This is especially the case
if path functions are used in the pattern definition. As explained in Section 3.2 these
functions determine all paths from all start to all target elements. Depending on the
number of objects in a given model in combination with the number of their relationships, an exponential number of paths have to be calculated. This theoretical complexity
explains the execution times of the LSA and PRC patterns in the large, highly branched
models.
The exponential increase in runtime performance motivates augmenting the original
matching approach to include a caching mechanism that stores previously calculated sub
patterns. The execution times of the augmented algorithm are depicted in Figure 3 and
Figure 4. Each figure depicts runtime performance of only one pattern, namely the PRC
pattern in Figure 3 and the XOR pattern in Figure 4. The models are again represented
on the abscissa by their number of elements. The figures depict two measurements for
each model, one with caching enabled and one with caching disabled. The execution
times of the original algorithm without the caching mechanism are taken as a reference
value for the measurements of the extended version. Again, the measurements represent
the mean value of all 35 search processes.
Figure 3 suggests that the caching mechanism improves the overall search performance
in every model. It represents measurements for the PRC pattern that allows for caching a
computationally expensive path search (cf. Section 4.2). The data suggest that the caching mechanism works particularly well on large and highly branched models in combination with complex patterns allowing to cache expensive operations. In the five largest
models the execution time decreased by approximately 50 per cent. Given this pattern,
the performance gain tends to decrease with the size of the model. For the smallest
model only a 20 per cent gain was measured.
The fact that the caching mechanism improves performance particularly well for computationally expensive operations is also supported by Figure 4. Here the execution
times of the XOR pattern are compared. This pattern only allows for caching relatively
cheap calls to the ElementsOfType function. The expensive path search is performed
only once. In this case, the caching mechanism yields hardly any performance increase
in large models. For smaller models, however, execution times can be decreased by up to
25 per cent. Since in models of up to 100 elements this pattern can be searched for in
less than one millisecond anyway, the actual performance gain is negligible. Still, the
data presented in Figure 4 indicates that execution times decrease in all but two models.
88
100%
90%
80%
70%
60%
50%
40%
30%
20%
10%
0%
343 150 142 121 104 89
79
67
67
61
Without caching
57
50
50
46
42
35
26
24
20
With caching
Figure 3: Comparison of the caching algorithm to the original algorithm executing the PRC pattern
These performance statistics prove that although our pattern matching algorithm is generic and not optimized to suit a particular application scenario it returns results with
acceptable performance. The caching extension further decreases execution time. Its
performance gain depends on the size and structure of the models to be searched as well
as the complexity of the patterns. It is particularly beneficial on large, highly branched
models in combination with patterns that contain computationally expensive operations
like the path search. In those cases, the caching mechanism increases runtime performance by up to fifty per cent. In the worst case, caching previously calculated sub-patterns does not yield any performance gain. As expected, however, it does not decrease
execution time either.
100%
80%
60%
40%
20%
0%
343 150 142 121 104 89
79
67
67
Without caching
61
57
50
50
46
42
35
26
24
20
With caching
Figure 4: Comparison of the caching algorithm to the original algorithm executing the XOR pattern
89
5 Summary and Outlook
In this paper, we conduct a statistically rigorous performance evaluation of a generic settheory based pattern matching approach. On the basis of a literature review we conclude
that there is no generic pattern matching approach suited for any application scenario
that at the same time is able to identify structures in conceptual models of arbitrary modelling languages. To prove that such a generic approach is beneficial in terms of performance, a detailed analysis of its runtime behaviour is required. After briefly introducing
the matching approach, we augmented the algorithm by a caching mechanism that stores
previously calculated sub-patterns. We introduced the model base as well as the patterns
we used to conduct the evaluation. Since the size of the models we analysed ranges from
20 to 343 elements, we argue that the results we obtained are also representative of other
process modelling languages as well as application scenarios. The evaluation process
was conducted according the methodology presented by [GBE07]. Our results were
therefore obtained in a statistically rigorous manner. The performance data suggest that
despite its generic nature our pattern matching approach returns results with acceptable
performance. Most of the patterns could be identified within a few milliseconds. The
caching mechanism we introduce continues to decrease execution time by up to fifty per
cent.
Having conducted a statistical analysis of its runtime behaviour, future research activities
will focus on a theoretical performance evaluation of the matching approach to determine the complexity class of the algorithm. To confront the exponential increase in execution time given large, highly branched models in combination with complex patterns,
we furthermore intent to introduce an additional function called ElementsOfTypeFast.
This function will work on a similar basis than the caching mechanism presented in this
paper. It will access a data structure that holds all elements of the respective element
types of a given modelling language. This data structure will furthermore allow for access in constant time. The idea is to fill this data structure only once before the actual
calculation of the pattern matches. ElementsOfTypeFast will then call this structure and
thus return all elements of a particular type in constant time which will lead to a further
increase in execution time.
References
[Au05]
Aumueller, D. et. al.: Schema and Ontology Matching with COMA++. In: Proc. of the
2005 ACM SIGMOD Int. conf. on Management of data, Baltimore, 2005; pp. 906-908.
[AVW08] Alves de Medeiros, A.K.; van der Aalst, W.M.P.; Weijters, A.J.M.M.: Quantifying
Process Equivalence Based on Observed Behavior. Data & Knowledge Engineering
64, 2008; pp. 55-74.
[Aw07]
Awad, A.: BPMN-Q: A Language to Query Business Processes. In (Reichert, M.,
Strecker, S., Turowski, K. Eds.): Proc. of the 2nd Int. Workshop on Enterprise Modelling and Information Systems Architectures, St. Goar, 2007; pp. 115-128.
[Be05]
Becker, J. et. al.: Referenzmodellierung in öffentlichen Verwaltungen am Beispiel des
prozessorientierten Reorganisationsprojekts Regio@KomM. In (Ferstl, O. K.; Sinz, E.
J.; Eckert, S.; Isselhorst, T. Eds.): Wirtschaftsinformatik 2005, Physica-Verlag HD,
Heidelberg, 2005; pp. 729-745.
90
Becker, J.; Schütte, R.: Handelsinformationssysteme. 2nd edition, Redline, Frankfurt,
2004.
[DDG09] Dijkman, R. M., Dumas, M., García-Bañuelos, L.: Process Model Similarity Search. In
(Dayal, U., Eder, J., Koehler, J., Reijers, H. A. Eds.): Proc. of the 7th Int. Conf. on
Business Process Management, Ulm, 2009; pp. 48-63.
[De10]
Delfmann, P. et. al.: Pattern Specification and Matching in Conceptual Models. A
Generic Approach Based on Set Operations. In: Enterprise Modelling and Information
Systems Architectures 5, 2010; pp. 24-43.
[DHY08] Dong, X. L.; Halevy, A.; Yu, C.: Data Integration with Uncertainty. In: Proc. of the
33rd Int. Conf. on Very large data bases, Vienna, 2008; pp. 687-698.
[Di09]
Dijkman, R. et. al.: Aligning Business Process Models. In: Proc. of the 13th IEEE Int.
Conf. on Enterprise Distributed Object Computing, Auckland, 2009; pp. 40-48.
[Di10]
Dijkman, R. et. al.: Similarity of Business Process Models. Metrics and Evaluation. In:
Information Systems 36, 2010; pp. 498516.
[DK07]
Delfmann, P.; Knackstedt, R.: Towards Tool Support for Information Model Variant
Management A Design Science Approach. In: Proc. of the 15th European Conf. on
Information Systems, St. Gallen, 2007; pp. 2098-2109.
[EKO07] Ehrig, M., Koschmider, A. Oberweis, A.: Measuring similarity between semantic
business process models. In (Roddick, J.F., Hinze, A. Eds.): Proc. of the 4th Asia-Pacific Conf. on Conceptual Modelling, Ballarat, 2007; pp. 7180.
[Fa09]
Fahland, D. et. al.: Instantaneous Soundness Checking of Industrial Business Process
Models. In (Dayal, U., Eder, J., Koehler, J., Reijers, H. A. Eds.): Proc. of the 7th Int.
Conf. of Business Process Management, Ulm, 2009; pp. 278-293.
[Ga95]
Gamma, E. et. al.: Design Patterns. Elements of Reusable Object-Oriented Software.
Addison-Wesley Longman, Amsterdam, 1995.
[GBE07] Georges, A., Buytaert, D., Eeckhout, L.: Statistically rigorous java performance
evaluation. ACM SIGPLAN Notices, vol. 42, 2007; pp. 57-76.
[Gr05]
Grossmann, G. et. al.: Behavior based integration of composite business processes. In
(van der Aalst, W.M.P., Benatallah, B., Casati, F. Curbera, F. Eds.): Proc. of the 3rd
Int. Conf. on Business Process Management, Nancy, 2005; pp. 186-204.
[He04]
Hevner, A. R. et. al.: Design Science in Information Systems Research. MIS Quarterly
28, 2004; pp. 75-105.
[Kü08]
Küster, J. et. al.: Detecting and Resolving Process Model Differences in the Absence
of a Change Log. In (Dumas, M., Reichert, M., Ming-Chien, S. Eds.): Proc. of the 6th
Int. Conf. on Business Process Management, Milan, 2008; pp. 244-260.
[Kn10]
Knuplesch, D. et. al.: On Enabling Data-Aware Compliance Checking of Business
Process Models. In (Parsons, J., Saeki, M., Shoval, P., Woo, C. C., Wand, Y. Eds.):
Proc. of the 29th Conf. on Conceptual Modeling, Vancouver, 2010; pp. 332-346.
[La10]
La Rosa, M.et. al.: Merging Business Process Models. In (Meersmann, R., Dillon, T.,
Herrero, H. Eds.): Proc. of the 2010 Int. Conf. on the move to meaningful internet
systems, Hersonissos, 2010; pp. 96-113.
[Me07]
Mendling, J.: Detection and Prediction of Errors in EPC Business Process Models.
Doctoral Thesis, Vienna University of Economics and Business Administration. Vienna, 2007.
[Pe07]
Peffers, K. et. al.: A Design Science Research Methodology for Information Systems
Research. Journal of Management Information Systems 24, 2007; pp. 45-77.
[PS11]
Po, L.; Sorrentino, S.: Automatic generation of probabilistic relationships for improving schema matching. Information Systems 36, 2011; pp. 192-208.
[RB01]
Rahm, E.; Bernstein, P. A.: A survey of approaches to automatic schema matching.
The Very Large Data Base Journal 10, 2001; pp. 334-350.
[RS10]
Rajesh, A.; Srivatsa, S. K.: XML Schema Matching Using Structural Information.
International Journal of Computer Applications 8, 2010; pp. 34-41.
[BS04]
91
[Sc00]
[SDH08]
[SM07a]
[SM07b]
[Sm09]
[Sm10]
[TC06]
[UC04]
[Va03]
[VAW06]
[VDM08]
[VTM08]
[WDM10]
[We10]
[We11]
[WHM08]
[YDG10]
Scheer, A.-W.: ARIS Business Process Modelling. 3rd edition, Springer, Berlin,
2000.
Sarma, A.; Dong, X.; Halevy, A.: Bootstrapping Pay-As-You-Go Data Integration
Systems. In: Proc. of the 2008 ACM SIGMOD Int. Conf. on Management of data,
Vancouver, 2008; pp. 861-874.
Simon, C., Mendling, J.: Integration of Conceptual Process Models by the Example of
Event-driven Process Chains. In Proc. of the 8th Int. Tagung Wirtschaftsinformatik,
Karlsruhe, 2007; pp. 677-694.
Simon, C., Mendling, J.: Integration of Conceptual Process Models by the Example of
Event-driven Process Chains. In (Oberweis, A., Weinhardt, C., Gimpel, H., Koschmider, A., Pankratius, V., Schnizler, B. Eds.): Proc. of the 8th Int. Tagung Wirtschaftsinformatik, Karlsruhe, 2007; pp. 677-694.
Smirnow, S. et. al.: Action Patterns in Business Process Models. In (Baresi, L., ChiHung, C., Suzuki, J. Eds.): Proc. of the 7th Int. Conf. on Service-Oriented Computing,
Stockholm, 2009; pp. 115-129.
Smirnov, S. et. al.: Meronymy-Based Aggregation of Activities in Business Process
Models. In (Parsons, J., Saeki, M., Shoval, P., Woo, C. C., Wand, Y. Eds.): Proc. of
the 29th Int. Conf. on Conceptual modelling, Vancouver, 2010; pp. 1-14.
Tansalarak, N.; Claypool, K.T.: QMatch Using paths to match XML schemas. Data
& Knowledge Engineering 60, 2006; pp. 260-282.
Uchitel, S., Chechik, M.: Merging Partial Behavioural Models. SIGSOFT Software
Engineering Notes 29, 2004; pp. 43-52.
Van der Aalst, W. et. al.: Workflow Patterns. Distributed and parallel databases 14,
2003; pp. 5-51.
van der Aalst, W. M. P., Alves de Medeiros, A. K., Weijters, A. J. M. M.: Process
Equivalence: Comparing Two Process Models Based on Observed Behavior. In (Dustdar, S., Fiadeiro, J.L., Sheth, A. Eds.): Proc. of the 4th Int. Conf. on Business Process
Management, Vienna, 2006; pp. 129-144.
van Dongen, B., Dijkman, R., Mendling, J.: Measuring Similarity between Business
Process Models. In (Bellahsène, Z., Léonard, M. Eds.): Advanced Information Systems
Engineering, LNCS 5074, Heidelberg, 2008; pp. 450-464.
Vergidis, K., Tiwari, A., Majeed, B.: Business process analysis and optimization:
beyond reengineering. IEEE Transactions on Systems, Man, and Cybernetics 3, 2008;
pp. 6982.
Weidlich, M., Dijkman, R., Mendling, J.: The ICoP Framework. Identification of
Correspondences between Process Models. In: (Pernici, B. Eds.): Proc. of the 22nd
Conf. on Advanced Information Systems Engineering, Hammamet, 2010; pp. 483-498.
Weidlich, M. et. al.: Process Compliance Measurement based on Behavioral Profiles.
In (Pernici, B. Ed.): Proc. of the 22nd Conference on Advanced Information Systems
Engineering, Hammamet, 2010; pp. 499-514.
Weber, B. et. al.: Refactoring large process model repositories. Computers in Industry
62, 2011; pp. 467-486.
Weber, I., Hoffmann, J., Mendling, J.: Semantic Business Process Validation. In
(Hepp, M., Stojanovic, N., Hinkelmann, K., Karagiannis, D., Klein, R. Eds.): Proc. of
the 3rd Int. Workshop on Semantic Business Process Management, Tenerife, 2008; pp.
22-36.
Yan, Z.; Dijkman, R.; Grefen, P.: Fast business process similarity search with featurebased similarity estimation. In (Meersmann, R., Dillon, T., Herrero, H. Eds.): Proc. of
the 2010 Int. Conf. on the move to meaningful internet systems, Hersonissos, 2010; pp.
60-77.
92
Some Guidelines for the Conception of Domain-Specific
Modelling Languages
Ulrich Frank
Information Systems and Enterprise Modelling
University of Duisburg-Essen
ulrich.frank@uni-due.de
Abstract: While the potential prospects of domain-specific modelling languages
(DSML) are undisputed, the design of a DSML faces specific challenges that have
raised only little attention so far. They relate to the boundaries between a DSML
and corresponding models as well as to the question how specific a DSML should
be. Addressing these challenges does not only contribute to the development of
meta modelling methods, it also relates to judging the economics of a DSML. This
paper suggests guidelines to support typical decisions that occur with the design of
a DSML. They mainly concern the level of abstraction a potential language
concept should be defined on. The guidelines are not intended to serve as recipes
for design, but rather to improve the transparency of design decisions.
1. Motivation
In recent years, the conception of domain-specific modelling languages (DSML) has
gained increasing attention. This is for comprehensible reasons: A general purpose
modelling language (GPML) provides modellers with rudimentary concepts such as
Class, Attribute etc. only. Different from that, a DSML includes higher level concepts
that are reconstructions of technical terms in the targeted domain, e.g. Organisational
Unit, Business Process, Strategic Resource etc. Therefore, the modeller is not forced
to specify these high level terms on his own. Instead he may reuse the concepts of the
DSML. This will not only promote modelling productivity. It also contributes to model
quality: Different from a GPML, the concepts of a DSML include implicit integrity
constraints that prevent the construction of nonsensical models. Furthermore, it should
be expected that the concepts of a DSML are developed with specific expertise and care.
Also, a DSML will typically feature a specific graphical notation that should foster
comprehensibility of respective models. The small, simplified example in Figure 1
illustrates the benefits of DSML compared to GPML. The prospects of deploying
DSMLs are contrasted with remarkable challenges. They concern both, the economics of
DSMLs and their development as well as principal considerations of language design.
Against this background, it is remarkable that there is hardly any method for guiding the
development of modelling languages in general, the design of DSML in particular.
Wands and Webers approach to evaluate the quality of modelling languages by
referring to ontological categories [WaWe93] does not help much with the specification
of a DSML in general, nor with differentiating levels of abstraction in particular. Take,
93
for instance, the construct ontological completeness: Depending on the purpose of a
DSML, completeness in this sense may be not relevant.
Work on modelling language design is usually focusing on formal features of meta
modelling languages or on meta modelling tools (e.g. [Jeus07], [JaSz06]). Since a
modelling method consists of at least one modelling language and a corresponding
process model, one would assume that work on method engineering has produced
support for the development of modelling languages.
Figure 1: Illustration of DSML in comparison to GPML
However, this is not the case. While there is a plethora of approaches on method
engineering (e.g. [Brin96], [Roll09], for an overview see [HeRa10]), they mainly focus
on the construction and configuration of process models and take the modelling language
as given. Recent publications on DSML such as [Klep08] and [KeTo08] describe the
construction and use of DSMLs, but do not go into the details of meta model design.
The current lack of support may be attributed to the fact that those who develop
modelling languages are usually highly specialized experts. However, that does not
mean that they are not in need for support, since the design of modelling languages can
be an extremely demanding task. For several years, an essential part of our research has
been directed towards the development of domain-specific modelling languages and
corresponding modelling methods. During this time we have faced numerous challenges
that are related to certain recurring design decisions and to peculiarities of the
development process. Many discussions with various designers of modelling languages
have shown that we share this experience with others. This may be regarded as a deluxe
problem that is relevant for a few only. However, the lack of substantiated guidance
increases the risk of poorly designed modelling languages, which will eventually affect
all prospective language users. With the increasing popularity of DSML and
corresponding model-based approaches to software development, this problem becomes
even more crucial. While a number of tools support the specification of DSML and the
94
realisation of corresponding model editors, prospective users can expect only little
guidance with designing a language that fits its purpose.
This paper is intended to suggest guidelines for supporting design decisions that are
specific to the development of DSML. For this purpose, it addresses mainly three
aspects: scope and economics, differentiating between language and language
application and dealing with abstraction conflicts. The background of this research is
work on languages for enterprise modelling. Nevertheless, the results should be
applicable to DSML in other areas as well.
2. Scope and Economics of DSML
Usually, the decision for developing a DSML will require some sort of economic
justification. The costs for developing and maybe distributing a DSML should not
exceed the additional benefit to be expected from its deployment. Unfortunately,
development costs and benefits are hard to judge in advance. If a DSML is developed for
one particular project only (as recommended e.g. by [KeTo08]), it may be possible to
estimate the expected cost/benefit ratio in a seemingly satisfactory way. However, as
soon as the development costs require a higher scale of reuse, the decision whether a
DSML will pay off or not will be of remarkable complexity and contingency. It seems
reasonable to assume that ceteris paribus the development costs will be the higher,
the more elaborate and specific a language is, i.e. the more semantics (in the sense of
information content) its concepts incorporate. However, the effect of semantics on a
DSMLs benefit is ambivalent. On the one hand, semantics should promote modelling
productivity and model quality, hence the benefit of reuse in a particular case. On the
other hand, semantics will compromise the scale of reuse: The more specific a DSML is,
the fewer the number of cases it can be applied to. Note, however, that trying to
determine an optimal level of semantics would be a hopeless undertaking: Not only that
measuring the level of semantics provided by a DSML on a metric scale is hardly
possible, the benefits of additional semantics will be hard to judge in a single case not
to speak of numerous cases with considerable variance. Despite these obstacles, it is
nevertheless required to make corresponding decisions, namely to determine whether or
not to develop a DSML for a certain domain and how specific it should be. At the
same time, there is need for reducing the costs and risks related to the development of a
DSML. Due to the contingency of the subject, deciding on the scope and economics of a
DSML in advance needs to account for specific aspects of a particular case. Hence, the
following criteria serve as general guidelines only that need to be interpreted against the
background of a particular project.
To outline the scope of the targeted DSML, modelling projects can be categorized with
respect to modelling subject and purpose, complexity and frequency. The higher the
complexity and frequency of modelling a certain subject for certain purposes, the higher
the benefit to be expected from a respective DSML. Also, the prospective users attitude
and skills need to be accounted for. If they are reluctant or even reject the idea of a
DSML, a negative impact on economics can be expected. To develop an elaborate idea
of the benefits to be expected from a DSML, a thorough requirements analysis is
95
mandatory. Unfortunately, analyzing the requirements a DSML should satisfy is a
particular challenge. Often, prospective users will not be familiar with the idea of a
DSML and, hence, will not know what they can expect from such an artefact. Our
experience in developing DSML suggests that it is a promising approach to develop
prototypical diagrams for certain analysis scenarios, e.g. the economic analysis of an IT
infrastructure (see [Fran10]).
Against this background, one can conduct a high-level assessment of benefits with
respect to certain classes of modelling tasks and corresponding users. To evaluate
prospective economics, costs and risks of introducing, using and maintaining the DSML
need to be addressed. This requires accounting for the availability of development tools,
the qualification of developers and users as well as the volatility of requirements to be
expected in future times. Sometimes, the availability of a DSML will enable additional
options such as the integration with other DSML or the use of models at run time. With
respect to evaluating the investment into the development of a DSML, these options
should be taken into account. Even if the economic perspective of a DSML seems to be
promising, its feasibility may still be a challenge especially if no previous experiences
with similar projects exist. In these cases, it may be a good idea to start with a smaller
scale project.
3. Differentiating between Language and Language Application
In general, a domain-specific modelling language is aimed at providing its prospective
users with technical terms suited to describe the targeted domain. However, sometimes it
is not evident, whether a concept should be part of the language specification or instead
be defined through the language. To illustrate this problem, it is useful to clarify the
targeted level of abstraction. The quest for abstraction suggests to not represent instances
for a good reason: Instances may change over time, which would compromise a
models validity. On the other hand, models that represent the M2 or even higher levels
of abstraction are mainly restricted to specialized applications such as the specification
of modelling languages. Therefore, we focus on the M1 level, since we assume that the
vast majority of conceptual models are assigned to it. As a consequence, designing a
DSML requires analyzing whether a domain-level term is suited to be represented as a
meta type and therefore can be incorporated into the language or not, which would
suggest representing him in a model specified with the prospective DSML. In natural
language, there is no explicit distinction between the type and the meta type level.
Sometimes a term can be used on both levels and often it is used for representing
instances, too. For instance: Business Process can be instantiated into various business
process types, which would qualify it as a meta type. One could argue that it could also
represent a type that then could be specialized into further business process types.
However, this is no option, since there is no suitable specialization concept for process
types.1 Organisational Unit can be seen as the meta type of Department, which in turn
could have instances like Sales Department, IT Service Group etc. But does Sales
1
While there are a few approaches to introduce relaxed concepts of process specialization, a specialization
concept that accounts for substitutability is not feasible, because it would require monotonic extensions.
96
Department need to represent an instance? Take for example a multi-national
corporation that defines a certain organisational structure for all its national subsidiaries.
In this case, Sales Department may serve as a type. While it seems undisputed that it
makes sense to model goals, it is not clear whether a goal like minimize throughput
time should be regarded as a meta type, a type or an instance. As a consequence, the
respective design decisions are even more demanding: In some cases, the foundational
presupposition that models are represented on the type level (M1) has to be dismissed.
For example: A DSML for modelling logistic networks should include concepts to
represent cities. If City was used as a type within a particular model, the level of
abstraction would usually be too high. Instead, it will often be the case that one wants to
represent concrete cities, hence, particular instances in a logistic network.
Table 1 illustrates the decision to be made when analyzing a term with respect to its
suitability for being incorporated into a DSML. In the regular case, the decision will
focus on the two alternatives in the upper row: whether a term should be represented as a
meta type to become part of the language or whether it is supposed to be specified by the
language. In some cases, it will be required to decide whether a term should rather be
represented as a meta type or a type, where the latter implies that it is appropriate to
represent instances in respective models.
DSML
model
Meta type
Business Process
Type
Order Management
Type
Location
Instance
Berlin
Table 1: Language vs. language application on different levels of abstraction
3.1 General Criteria
To develop guidelines for the above decision scenario, we will first look at criteria that
need to be accounted for with every prospective language concept. Hence, we focus on
the principle decision whether a term is suited to be incorporated in a language. To
illustrate the criteria, we refer to examples form the domain of organisational analysis
and design. Note that the following criteria do not account for the question whether a
concept should be reconstructed as a type or a meta type, but only whether or not it is
suited for becoming part of a DSML:
a)
Noteworthy level of invariant semantics: If a term can be expected to have the
same meaning across all application areas a DSML is supposed to cover, it can
be regarded as a candidate for being incorporated in the language. In other
words: It should be possible to define the essential meaning of the concept.
Essential means that the features used to specify the concept are sufficient to
capture at least a minimal set of characterizing features, i.e. features that would
allow for distinguishing it from other concepts. Furthermore, one should know
how to describe the range of its more concrete specifications. In other words:
There should be substantial knowledge about the conceptual variance the
97
abstraction covers. These aspects are related to the level of semantics, a
potential language concept includes. The level of semantics corresponds to its
information content: The more interpretations are excluded, the higher the level
of semantics. Hence, the level of semantics depends on the semantic
determination of its features, i.e. its attributes and the associated (meta) types.
The level of semantics embedded in a language concept is a core indicator of its
value as an abstraction: The more we know about possible instances (types or
instances), the more can we express about them on an abstract level without
having to care for the peculiarities of particular instances. If the meaning of a
concept changes significantly with the application context, this could be
handled only with a specification on a lower level of semantics. Alternatively, it
may be a reasonable decision to reduce the scope of a language, i.e. to make it
more special. Hence, applying this criterion is interrelated with determining a
DSMLs intended scope (see 2).
Test: At first sight, it may appear that the most satisfactory approach to test an
assumption about a concepts meaning across the set of intended applications
would be to conduct an elaborate and representative empirical investigation.
However, such an approach faces two major drawbacks. First, it will often be
not appropriate to focus on the actual meaning of a concept, but rather to pursue
a semantic reconstruction that fits the purpose of the DSML. Second, for
economic reasons, a comprehensive empirical investigation will usually be no
option. Therefore, introspection (reflecting upon known uses of a term which
makes it empirical, too) and discursive evaluation will usually be the option of
choice. In any case, the results of such a test should be regarded as a hypothesis
that may still be refuted.
b) Relevance: If a concept is expected to be used on a regular base, making it part
of the language would increase the languages value for most users. If it is
required in rare cases only, it would increase the size of the language, hence
make it more difficult to learn, while at the same time most users would not
benefit from it. While this criterion is related to all other criteria, it is different
in the sense that it puts explicit emphasis on the purpose of the language and the
actual range of applications.
Test: Creating exemplary models within the scope of intended applications
helps to find out how relevant the use of the concept would be with respect to
intended purposes of corresponding models. Analyzing intended applications
and related purposes may suggest empirical studies. However, such an approach
has its limitations: It does not seem plausible to expect a comprehensive
assembly of relevant purposes from prospective users if they are not familiar
with the considered kind of modelling (which will be especially the case with
DSMLs that target new domains).
98
3.2 Language Concept as Meta Type or Type?
In addition to determining whether a term is in principle suited to become part of a
DSML, it needs to be decided whether it should be reconstructed as a meta type which
can be regarded as the regular case or rather as a type. The first two criteria focus on
checking the suitability as a meta type, while the third and fourth address the suitability
as a type.
c)
Noteworthy semantic differences between types: From a formal point of view, a
meta type should allow for a range of types with clear semantic differences. If
the types that can be instantiated from the potential meta type are all too
similar, the effort it takes to define the types is questionable. This criterion, too,
does not only apply to the entire range of intended applications, but also to the
semantic variance of types within particular models: It does not contribute to
the benefit of the abstraction in a particular case, if there are no noteworthy
semantic differences between potential types. Again, if the variance is
noteworthy only in some cases, while in others it is not, it would indicate the
contingency of a subject.
Test: From a formal point of view, the meta type should allow for a range of
significantly different types. That implies that the types can be defined on a
high level of semantics otherwise there would not be much chance to express
clear differences between types. If this is the case, one would have to check
whether there is a noteworthy number of clearly different types within the range
of possible types.
d) Instance as type intuitive: This criterion emphasizes language ergonomics.
Would it correspond to the common use of the term to regard its instances as
types? Hence, would instances be regarded as types intuitively by prospective
language users or would they rather be interpreted as representing instances of
the respective domain?
Test: Since many people do not consciously differentiate between type and
instance level in many cases, an empirical test would have to face a remarkable
challenge. Therefore, referring to our Sprachgefühl (sense of language) and to
its discursive evaluation will usually be the only feasible option.
The following criteria serve to check whether a term that does not qualify as a meta type
could instead be represented as a type offered by a DSML.
e)
“Instance” as abstraction: There are terms that emphasize abstractions by their
very nature. Therefore, they resist against a clear distinction between the type
and the instance level. For example: Is a goal such as minimize throughput
time of a set of business processes a type or rather an instance? It may be
regarded as a type, since it applies to a set of processes. At the same time, it
hardly allows for further instantiation. Nevertheless, it may be important to
represent a corresponding goal in a model. In these cases, it is suitable to make
a respective type, e.g. Goal, part of the language.
99
Test: Since there is hardly a formal criterion to guide this test, it depends chiefly
on reflection and introspection to find out whether a term that does not allow
for further instantiation can still be regarded as an abstraction.
f)
Invariant and unique instance identity: Sometimes, it makes sense to represent
objects in a model that clearly qualify as instances. Examples include cities,
countries or particular organisations. Apparently, these examples have in
common that the respective instances have a unique identity that is relevant, i.e.
it makes a difference, for the modelling purpose. Also, the features that are
relevant for the modelling purpose are (widely) invariant within the intended
lifetime of a model (e.g. the geographical location of a city). Therefore,
including them in a model does not mean to compromise the quest for
abstraction.
Test: To test a term against this criterion, a number of models where it is
represented as an instance need to be analyzed. The focus of the test is on the
following questions: Are there useful analysis scenarios related to the intended
type of models that require accounting for instances? Are these instances
(widely) invariant with respect to the modelling purpose?
Against this background, we suggest the following guidelines.
Modelling Guideline R1: If a concept fulfils criteria a) to b), it should be
considered for becoming part of the language.
Modelling Guideline R2: If a concept fulfils criteria c) and d), it should be
considered for being represented as meta type.
Modelling Guideline R3: If a concept does not satisfy guideline R2 and fulfils
either criterion e) or f), it should be considered for being represented as type.
To illustrate the use of the suggested criteria, we consider three terms selected from the
examples discussed above. The following tables show the result of evaluating them
against the criteria of the modelling guidelines. + indicates that the term clearly fulfils
the corresponding criterion.
Organisational Unit
An organisational unit is a part of an organisation that reflects a permanent principle of the division of
labour. An organisational unit may contain other organisational units.
a) invariant semantics
+
The term is used on a high level of abstraction. The
essence of its semantics should not vary substantially.
b) relevance
This is a key term for describing organisation structures.
+
c) variance of type semantics
+
The semantics of types of organisational units can vary
significantly.
d) instance as type intuitive
+
At least some of the potential instances will be regarded
c
as types almost intuitively, e.g. Department, Division.
Other potential instances, such as Marketing
Department, Consumer Electronics Division, will
probably not be regarded as types by many. Hence, the
final assessment of this criterion depends on the
recommended instantiation.
100
o expresses that it fulfils the criterion to a satisfactory degree, while - indicates that
it fails to satisfy the corresponding criterion. c means that with respect to the
considered criterion the use of the term is contingent, i.e. in some cases it fulfils the
criterion, in others it does not. Note that guidelines R2 and R3 are considered only, if
guideline R1 applies. R3 is relevant only, if R2 does not apply. Note that there are no
precise guidelines based on hard facts for evaluating domain terms against the
suggested criteria. Instead, it is recommended to make use of discursive evaluation that
is aimed at developing reasons that are regarded as convincing by the participants.
Apparently, the term Organisational Unit qualifies as language concept to be
represented as a meta type. Nevertheless, it may be an option to include a more specific
term in a DSML, e.g. Department: It may be regarded as a relevant term of the
corresponding technical language and since it is more specific than Organisational Unit
it could increase a DSMLs contribution to productivity and integrity.
Department
A department is an organisational unit, which will usually include further organisational units. It may
be part of a superordinate organisational unit.
a) invariant
semantics
b) relevance
As the description of the term indicates, there is not much that is
specific to the term department, which would clearly distinguish it
from the more general term organisational unit. Even on a high level
of abstraction, the meaning of the term varies. In some cases, a
department will be part of a superordinate head department or a
division, in other cases not.
This term is frequently used for describing organisation structures.
c
+
Since criterion a) is not satisfied, guideline R1 does not apply to Department.
Therefore, testing it against guidelines R2 and R3 is obsolete. Note, however, that
criterion a) is contingent, which means that for some domains the terms semantics may
be invariant (see 4.1).
Location
A location represents a geographical area and a corresponding settlement such as a village or a city.
a) invariant
semantics
The essence of the term as it is outlined in the above description
should not vary substantially.
+
b) relevance
In various domains, e.g. in cross-organisational logistics, this is a key
term.
+
c) variance of
instances
The semantics of instances may vary to some degree.
d) instance as type
intuitive
It is conceivable to regard an instance as type, e.g. the type City as
opposed to the type Village. Often, however, it will be more
appropriate to instantiate directly into a particular location such as
Berlin or Hamburg.
101
o
o
c
Since guideline R2 does not clearly apply, R3 can be tested additionally. Note that in this
case, it is assumed that the concepts instances represent particular instances such as
cities.
e) Instance as
abstraction
If an instance is a particular location, then this criterion is not
satisfied.
-
f)
Cities, urban districts and villages satisfy this criterion without a
doubt. Buildings will be regarded as invariant and unique in most
applications, too.
+
invariant and
unique instance
identity
The result of the last evaluation shows that Location is suited as a language concept.
The decision on the level of abstraction depends on the intended purpose. If models are
supposed to represent particular locations, then the term is suited for being represented in
a DSML as a type. Otherwise, a representation as meta type would be more preferable.
Note, however, that there may be cases where it is required to account for both levels of
abstraction (see 4).
Applying the suggested guidelines to a few examples produces an ambivalent result. On
the one hand, the criteria help to reduce the complexity of decisions related to the design
of DSML. On the other hand, they are not easy to apply. Instead, they require a thorough
and differentiated interpretation that may be regarded as arduous by some. This is
especially the case for those criteria that are marked as contingent. Therefore, the
guidelines need to be interpreted with respect to the intended scope and purpose of a
DSML. However, scope and purpose of a DSML are sometimes hard to determine in
advance. Our experience suggests that developing modelling scenarios helps to get a
better idea of scope and purpose and to dissolve the principle contingencies related to
some criteria.
4. Dealing with Conflicting Levels of Abstraction
The previous evaluation of the guidelines has shown that they do not always produce
clear results, which requires further analysis. Sometimes, however, the results may even
be partially contradictory with respect to the appropriate level of abstraction.
4.1 Local Types
There are terms that are not perfectly suited to be incorporated into a DSML, because
their semantics is not invariant within the intended language scope. Nevertheless they
may be important elements of respective technical languages - and may be useful for
particular applications of a DSML. For example: The semantics of the term 'Department'
is not invariant across a wider range of organisations. Therefore, it seems appropriate at
first to not include it in a DSML for organisation modelling. Instead, the DSML could
provide the more abstract concept 'Organisational Unit', which could be instantiated into
types such as Marketing Department, Finance Department etc. However, this would
not be satisfactory either, because it would not be possible to express that a marketing
102
department and a finance department (both defined on the M1 level) are two types of the
same kind. To deal with these contradictory requirements, a trade-off is recommended,
which is expressed in the following guideline:
Modelling Guideline R4: If a term is not used consistently throughout the set of
intended applications, it should not be incorporated into the language in order to
not violate modelling rule R2. However, if this term is an important part of the
respective technical language and its semantics is fairly invariant within a certain
application area, it can be represented as a "local type".
A local type is an auxiliary abstraction that allows for adapting a DSML to specific
needs. In the case of 'Department', it would enable a company to define its own
conception of department. Figure 2 illustrates the specification of a local type. The local
types Department and Division are associated to instances of OrganisationalUnit to
specify their semantics in a specific local scope.
Figure 2: Illustration of local types
In order to protect model integrity, the local type of an organisational unit can be defined
by an instance of an additional meta type (LocalUnitType), which can be associated
with an instance of OrganisationalUnit. LocalUnitType provides the option to specify
the hierarchical level and a characteristic designation. Furthermore, it would be possible
to assign a specific graphical notation. A level is defined by a positive Integer. The
lowest level, 0, is reserved for those organisational units that do not include further units
and/or that do not include any position that is in command of further organisational
units. If these levels are specified for all local types of organisational unit types, it is
possible to enforce the constraints that they imply, e.g. that no organisational unit on
level 2 can be part of a unit on level 1. The definition of a local terminology through
103
instantiations of LocalUnitType can be regarded as defining language concepts with a
limited scope, or as pseudo meta types. This approach can be further supported by
supplementing a DSML with instantiations of LocalUnitType for specific subdomains,
e.g. for industrial enterprises, for public organisations, universities etc.
4.2 Intrinsic Features
On the one hand, specifying a meta model requires reflecting upon the ontological
essence of a term. On the other hand, it recommends taking into account that instances of
a meta concept are types. Sometimes, this results in the problem that the essence of a
term includes features that do not apply directly to the type level. Instead, they apply to
the instances represented by a type. For example: A language for modelling business
processes includes the meta type BusinessProcess, which has attributes such as name,
averageExecutionTime and various associations to other meta types like Event,
OrganisationalUnit etc. Within a particular model, it is instantiated to a certain business
process type, e.g. Order Management, which includes the instantiation of attributes
from corresponding meta types. While we know that every business process instance has
a particular start and termination time, these materialized features do not apply to the
corresponding business process type. Since a meta type may only define features that can
be instantiated to describe features of a type, it is not possible to express features that
apply to the instances of this type only. However, assigning these features to every
instance would not only ignore an obvious abstraction, it would also result in
redundancy. This problem is well known in conceptual modelling and addressed by
concepts such as power types [Odel98] or clabjects [AtKü07]. A similar concept,
called "intrinsic feature" is proposed in [Fran11] to enrich a meta modelling language. A
(meta) type may have (regular) attributes that apply to its instances or intrinsic
attributes that can be instantiated only from the instances of its instances. The concept
can also be applied to an entire entity type (an instance of a meta type): An entity type
that must not be instantiated directly, but only on the level below the one it is presented
on, is called an intrinsic type. Figure 3 illustrates how to specify the example concepts
OrganisationalUnit and Location with the use of intrinsic features. Intrinsic features
are marked by an i printed white on black. Specifying e.g. Location with intrinsic
features would allow for representing both, the semantics of location types such as
City, Village etc., and respective instances. A corresponding model would include
instances, e.g. Hamburg, that are characterized by a type, e.g. City, which was
defined by instantiating the meta type Location. Hence, a respective model editor
would need to provide modellers with two levels of abstractions. Unfortunately not
every meta modelling language features concepts for specifying instance level features.
But even in these cases, it is a good idea to somehow add their description to the meta
model either by comments or by marking type level features.
Modelling Guideline R5: If the essential meaning of a meta type that is part of a
DSML includes aspects that apply to instances of the types of the meta type and
these aspects are relevant for the intended application of the DSML, then these
instance level features should be included in the specification. Preferably, this
should be done with meta modelling languages that feature corresponding
104
concepts. Alternatively, instance level features can be marked or added in
comments.
Figure 3: Example of Specification with Intrinsic Features
5. Concluding Remarks
In contrast to the remarkable attention generated by the idea of DSML, there has been
only little work on specific design decisions related to the conception of a DSML. While
there has been some work on differentiating instance and type level features (see 4.2),
questions that concern the distinction of a prospective DSML and its application have
hardly been addressed so far. This is the case, too, for the decision on how specific a
DSML should be. This paper has shown that these questions are essential for the
development and economics of DSMLs and very challenging at the same time. Due
to the remarkable contingency of modelling subjects and purposes, it may not come as a
surprise that the proposed guidelines do not serve as recipes. Instead, their application
requires a considerable effort. As a consequence, they are mainly intended to improve
the transparency of design decisions and draw the reflective language designers
attention to aspects that require further investigation. Also, they are meant to contribute
to the discourse on the quality of meta models and DSML in general (see, e.g., [Fran98],
[OpSe99], [FeLo03]). In addition to that, the guidelines suggest specific requirements for
meta modelling languages and corresponding meta model editors that are not satisfied by
most of todays solutions. Every guideline has been motivated with respect to certain
presuppositions and has been tested against a few examples. Nevertheless, there is still
need for further tests and discursive evaluations, since we are only at the beginning of
developing expressive methods for designing DSML.
105
6. References
[AtKü08]
Atkinson, C.; Kühne, T.: Reducing accidental complexity in domain models. In:
Software and Systems Modelling. Vol. 7, No. 3, 2008, pp. 345-359
[Brin96]
Brinkkemper, S.: Method engineering: engineering of information systems
development methods and tools. In: Information and Software Technology, Vol. 38,
No. 4, 1996, pp. 275280
[DaPi+02] Dahchour, M.; Pirotte, A.; Zimányi; E.: Materialization and Its Metaclass
Implementation. IEEE Transactions on Knowledge and Data Engineering Vol. 14,
No. 5, pp. 1078-1094
[FeLo03] Fettke, P.; Loos, P.: Ontological evaluation of reference models using the BungeWand-Weber-model. In: Proceedings of the Ninth Americas Conference on
Information Systems. Tampa, FL 2003, pp. 2944-2955
[Fran11]
Frank, U.: The MEMO Meta Modelling Language (MML) and Language
Architecture. 2nd Ed., ICB Research Report No. 43, University Duisburg-Essen 2011
[Fran10]
Frank, U.: Outline of a Method for Designing Domain-Specific Modelling Languages.
ICB-Research Report No. 42, Universität Duisburg-Essen 2010
[Fran98]
Frank, U.: Evaluating Modelling Languages: Relevant Issues, Epistemological
Challenges and a Preliminary Research Framework. Research Report No. 15, Institut
für Wirtschaftsinformatik, Universität Koblenz-Landau 1998
[HeRa10] Henderson-Sellers, B.; Ralyté, J.: Situational Method Engineering: State-of-the-Art
Review. In: Journal of Universal Computer Science, vol. 16, no. 3, 2010, pp. 424-478
[JaSz06]
Jackson, E.; Sztipanovits, J.: Towards A Formal Foundation For Domain-specific
Modelling Languages. In: Proceedings of the Sixth ACM International Conference on
Embedded Software (EMSOFT06), 2006, 53-63
[Jeus07]
Jeusfeld, M.A.: Partial Evaluation in Meta Modelling. Proceedings of the IFIP 8.1
Working Conference on Situational Method Engineering (ME-2007), Geneva, 2007,
Springer: Berlin, Heidelberg etc. 2007, pp. 115-129
[KeTo08] Kelly, S.; Tolvanen, J.-P.: Domain-Specific Modelling: Enabling full code generation.
Wiley-IEEE Computer Society Press 2008
[Klep08] Kleppe, A.: Software Language Engineering: Creating Domain-Specific Languages
Using Metamodels. Addison-Wesley 2008
[Odel98] Odell, J.: Power Types. In: Odell, J. (Ed.): Advanced Object-Oriented Analysis and
Design Using UML. Cambridge University Press: Cambridge 1998, pp. 23-33
(revised version of: Odell, J.: Power Types. In: Journal of Object-Oriented
Programming, Vol. 7, No. 2, 1994, pp. 8-12)
[OpHe99] Opdahl, A.L.; Henderson-Sellers, B.: Evaluating and Improving OO Modelling
Languages Using the BWW-Model. In Proceedings of the Information Systems
Foundations Workshop (Ontology, Semiotics and Practice), (digital publication),
Sydney 1999
[Roll09]
Rolland, C.: Method Engineering: Towards Methods as Services. In: Software Process
Improvement and Practice. Vol. 14, 2009, pp. 143164
[WaWe93] Wand, Y.; Weber, R.: On the Ontological Expressiveness of Information Systems
Analysis and Design Grammars. In: Journal of Information Systems, 1993, pp. 217237
106
Towards Process-oriented Information Logistics:
Why Quality Dimensions of Process Information Matter∗
Bernd Michelberger1 , Bela Mutschler1 , Manfred Reichert2
1
University of Applied Sciences Ravensburg-Weingarten, Germany
{bernd.michelberger; bela.mutschler}@hs-weingarten.de
2
Institute of Databases and Information Systems, University of Ulm, Germany
manfred.reichert@uni-ulm.de
Abstract: An increasing data overload makes it difficult to deliver needed information
to knowledge-workers and decision-makers in process-oriented enterprises. The main
problem is to identify information being relevant for process participants and their
activities. To cope with this challenge, enterprises crave for an intelligent and processoriented information logistics. The major challenge is to provide the right process
information in the right format and level of granularity at the right place and accurate
point in time to the right actors. When realizing such process-oriented information
logistics it becomes crucial to take into account quality dimensions of process information (e.g., completeness, topicality, punctuality). Reason is that these dimensions
determine process information quality and thus also the overall relevance of process
information for a particular process participant and his activities. This paper picks up
this issue and analyzes different quality dimensions of process information and their
impact on process-oriented information logistics.
1 Introduction
Market globalization has led to massive cost pressure and increased competition for enterprises. Products and services must be developed in ever-shorter cycles and new forms
of collaboration within and across organizations are continuously emerging. As examples consider product engineering processes [MHHR06] or the treatment of patients in an
integrated healthcare network [LR07]. To cope with these challenges, effective business
process management (BPM) [Wes07] becomes success-critical for enterprises.
BPM technology has focused on the modeling, analysis, and execution of processes (e.g.,
using BPM systems) [MRB08] in recent years. What has been neglected so far is the
support of knowledge-workers and decision-makers by providing personalized process information to them depending on their current work context. The latter determines the
information a process participant needs in order to perform current and newly upcoming
∗ This research was performed in the niPRO project. This project is funded by the German Federal Ministry of Education and Research (BMBF) under grant number 17102X10. More information can be found at
http://www.nipro-project.org.
107
activities. For example, to prepare his ward round a doctor needs information about his
patients. Besides process-related information (e.g., the current activity), the work context
also comprises device-related information (e.g., display size, bandwidth), location-based
information (e.g., GPS location), and user-related information (e.g., user name, role, department). Overall, an extensive amount of process information is provided within and
across organizations using techniques and tools such as e-mail, shared drives, Web 2.0
applications, and enterprise information systems [LL09].
Before characterizing the notion of process information, we first have to define the terms
data and information. In literature, respective definitions are broadly diversified [Row06].
In the context of our research we apply the definitions suggested in [BCGH06, RT08,
AG04]: data are raw facts or observations of things, events, activities, and transactions that
are recorded and stored, but are not organized and processed, and therefore do not convey
any specific meaning. Information, in turn, refers to data that has been organized and
processed for a specific purpose. Consequently, it has a meaning and provides some value
to the recipient. Generally, data turns into information, if someone is interested in this data.
For example, a doctor might be interested in the blood group of a particular patient or the
patient’s maximum and minimum body temperature during a day. Besides, information
can be also derived from data. As example consider the average body temperature that can
be calculated from the individual temperature data items. Consequently, the difference
between data and information is not structural, but functional [Ack89].
Analogously, we define process information as follows: process information refers to data
that has been processed to support process users in the modeling, execution, monitoring,
optimization, and design of processes. Hence, the data gets a meaning and has a value
with respect to the process users’ activities. Typical process information includes, for example, process descriptions, working guidelines, process models, operational instructions,
forms, checklists, and best practices (e.g., documented in text documents, spreadsheets,
presentations, and e-mails).
However, the mere availability of process information is not sufficient to adequately support knowledge-workers and decision-makers in their daily activities. What enterprises
need is an intelligent, process-oriented information logistics, i.e., the right process information must be provided in the right format and level of granularity at the right place
and accurate point in time. More precisely, process-oriented information logistics deals
with the planning, execution, and control of process information flows within or between
enterprises to support knowledge-intense business processes implying human interaction
and decision making [BD08]. In this paper, we address one first aspect of an effective
process-oriented information logistics: quality dimensions of process information.
Based on the question whether process information fulfills certain quality requirements,
the overall relevance of process information can be determined. Picking up our healthcare example, typically, it becomes necessary that patient information is up-to-date and
complete in order to be able to charge certain services through the accounting and billing
department. Therefore, patient information which is out-of-date or incomplete is not relevant, since it cannot be processed by the medical accounting. However, depending on a
specific work context, different quality dimensions might be more or less important than
others. For example, for a surgeon, patient information should be available punctual and
108
up-to-date. Conversely, for an employee being responsible for patient admission, information about the patient must be complete and error-free. Therefore, depending on the work
context, a different weighting of the individual quality dimensions becomes necessary.
In fact, the consideration of work context and quality dimensions of process information is
key to identify relevant process information. Figure 1 shows the relationship between work
context and process information quality. On the one hand, the work context determines
the process information a process participant needs to perform current activities. On the
other hand, the use of quality dimensions allows to determine process information quality.
Together, both aspects allow to determine the overall relevance of process information.
Work Context
task
Process
User
task
x
Goal: selection
of relevant
(overall relevance)
Process
Information
Quality
Quality Dimension #1
Quality Dimension #2
task
...
Process
Information
Quality Dimension #n
Figure 1: Determining the relevance of process information.
The presented research is performed in the niPRO project. In this project we apply semantic technology to integrate process information within intelligent, user-adequate process information portals. Our overall goal is to support knowledge-workers and decisionmakers with the needed process information depending on their current work context. So
far, both research and practice do not address how processes and related process information can be effectively merged. Currently, conventional methods of information retrieval or
enterprise search engines are mainly used for this purpose. Opposed to this, the niPRO process information portal aims at determining required information for knowledge-workers
and decision-makers dynamically and automatically. Key challenges include the personalized provision of process information, flexible visualization of process information, and
innovative design approaches for different levels of information granularity.
This paper is organized as follows. Section 2 presents a running example that will be used
throughout the paper. Section 3 investigates quality dimensions of process information.
Section 4 discusses why contextual quality dimensions are particularly important. Section
5 discusses related work. Section 6 concludes the paper with a summary and an outlook.
2 Running Example
To illustrate different quality dimensions of process information, we use a running example from the clinical domain. This example is based on experiences we gathered during
an exploratory case study at a large German university hospital [MMR11]. In this case
study we analyzed the process of an unplanned, stationary hospitalization, including patient admission, medical indication in the anesthesia, surgical intervention, post-surgery
treatment, patient discharge, and financial accounting and management.
109
Our running example (cf. Figure 2) specifically picks up the doctor’s ward round. First,
the ward round is prepared, i.e., the doctor has a look at patient information and current
medical instructions (e.g., endoscope, physical therapy). Next, the doctor communicates
with the patient and asks for information about his status. Afterwards, the patient is examined. This activity includes the analysis of blood values, vital values, and further follow-up
diagnosis. Finally, the doctor creates medical instructions and updates patient information.
...
Patient
Record
prepare ward
round
...
Notes
communicate
with patient
...
examine
patient
Notes
Instructions
create medical
instructions
Notes
Patient
Record
update patient
information
Figure 2: Our running example.
For each of these activities different process information is needed. For example, to perform activity ”create medical instructions” a doctor needs blood values, vital values, notes,
and current medical instructions. Conversely, to perform activity ”update patient information”, the doctor needs output information (e.g., notes, instructions) of other process activities and also access to the patient record. Note that the illustrated process information
(e.g., notes, instructions, patient record) from our running example should be viewed only
as a small part of all relevant process information.
To determine which process information is relevant in a specific work context, we now take
a closer look at quality dimensions of process information. This way, we can ensure that
process information meets necessary quality requirements (e.g., contextual relevance, freeof-error, objectivity, and believability), fits with the work context, is suitable for specific
activities, and can so be easily used by process users.
3 Process Information Quality
Apart from very broad characterizations such as ”fitness for use” [TB98] there exists no
commonly accepted understanding of the term information quality. In fact, giving a single
definition of information quality is difficult as this term is widely used in many areas. In
this paper we define process information quality as a set of quality dimensions.
Process information quality can be investigated from various viewpoints, e.g., integration,
transmission, security, storage, access, and representation. According to the goals of the
niPRO project, we focus on the viewpoints of integration, semantic processing, and provision. Integration concerns the collection of process information from different data sources
(e.g., databases, enterprise information systems, shared drives). The viewpoint semantic
processing implies semantic analysis, processing and linking of process information. The
provision viewpoint, finally, deals with the technical provision of process information.
110
Quality categories and quality dimensions described in the following were gathered based
on a literature study, two qualitative exploratory case studies, and an additional online
survey [MMR11, HMR11].
3.1
Quality Categories of Process Information
Quality dimensions of process information can be combined into different quality categories. Each category subsumes a set of dimensions. All dimensions belonging to the
same category are affected by the same influencing factors such as work context (e.g.,
process- and user-related information) or information systems characteristics (e.g., representation of information). Specifically, we apply the classification of Wang [WS96] and
Moore & Benbasat [MB91] and introduce four quality categories (cf. Figure 3):
• The Intrinsic quality category (QC1) integrates self-contained quality dimensions
of process information. Quality dimensions from this category are independent on
the work context. Examples include believability (e.g., to improve the believability
of a medical diagnosis several doctors have to approve it) and objectivity (e.g., to
guarantee the objectivity the health of patients must be determined by certain criteria
and not by estimation). Another example is free-of-error (e.g., to achieve error-free
patient lists, name and identification number of the patient must match).
Intrinsic
(QC1)
Accessible
(QC2)
Quality Categories of
Process Information
Representational
(QC3)
Contextual
(QC4)
independent on
work context
dependent on
work context
Figure 3: Quality categories of process information.
• The Accessible quality category (QC2) combines quality dimensions being important for the access to process information. These are mainly affected by the information systems providing process information. Examples of respective quality dimensions are accessibility (e.g., to treat a patient the doctor needs the patient record)
and security (e.g., ensure the security so that specific process information is only
accessible to authorized users).
• The Representational quality category (QC3) subsumes quality dimensions concerning the representation of process information. This quality category is again
mainly influenced by the information systems providing process information. As
examples of respective quality dimensions consider interpretability (e.g., the exact
unit of measurement is always indicated for the given values), understandability
111
(e.g., addresses should not be displayed as GPS coordinates), consistent representation (e.g., patient information should be display consistently), and concise representation (e.g., current diseases are displayed separately from pre-existing diseases).
• The Contextual quality category (QC4) integrates quality dimensions which are influenced by the work context of process users. Contextual quality dimensions are,
for example, contextual relevance (e.g., a doctor performing activity ”prepare ward
round” gets other process information than in activity ”create medical instructions”),
completeness (e.g., patient information must be completely available), and punctuality (e.g., blood values must be available when the doctor needs it). These quality
dimensions always depend on the current work context.
In the next section we restrict ourselves to the contextual quality category. We do this
because this quality category is particularly influenced by the work context and thus also
by process-related information.
3.2
Contextual Quality Dimensions of Process Information
We distinguish between nine contextual quality dimensions of process information (cf.
Figure 4). Each dimension is described in the following.
Contextual Quality
Category (QC4)
Punctuality
(QD1)
Topicality
(QD2)
Contextual
Relevance (QD3)
Completeness
(QD4)
Value-Added
(QD5)
Appropriate
Amount (QD6)
Granularity
(QD7)
Neighbourhood
(QD8)
Methods-of-Use
(QD9)
Figure 4: Contextual quality dimensions of process information.
3.2.1 Punctuality
The Punctuality quality dimension (QD1) indicates whether process information is provided punctually when the process participant needs it. Specifically, three different time
points (t) have to be distinguished: (a) the point in time at which the process user requests
the process information, (b) the point in time at which the process information is provided,
and (c) the point in time at which the process user applies the process information. Based
on this we can determine whether process information is punctual or not.
Additionally, it becomes necessary to distinguish between ad-hoc process information and
regular one. The former is requested spontaneously. For example, a doctor may request
blood values in order to be able to make decisions. Ad-hoc process information is accurate
112
in time if it is provided between the point in time it is requested and it is used (cf. Figure
5). The length of this period depends on the process participant.
request (a)
use (c)
time
t
t+1
t+2
t+3
t+4
punctual
t+5
t+6
unpunctual
Figure 5: Punctuality of ad-hoc process information.
Conversely, regular process information is provided at pre-defined points in time. For example, every morning a doctor may receive a patient list in order to know which patients
he has to visit. The punctuality of regular process information can be distinguished between two time points: (a) punctual in respect to the provision and (b) punctual in respect
to the use (cf. Figure 6).
provision (b)
use (c)
time
t
t+1
t+2
t+3
punctual to the provision
t+4
t+5
t+6
unpunctual
punctual to the use
Figure 6: Punctuality of regular process information.
3.2.2 Topicality
The Topicality quality dimension (QD2) indicates whether process information captures
the current characteristics (e.g., name, insurance agreement) of an object (e.g., patient) at
the current point in time (t). Process information is out-of-date if the characteristics of
the object have changed between the time point of capture and the time point at which
the process user applies the process information (cf. Figure 7). For example, a body
temperature of the patient measured two days ago is most likely obsolete. In practice,
the capture of characteristics is often time-consuming. Characteristics of an object may
continuously change (e.g., body temperature of a patient, health of patient). The capture
can be done either in real-time (e.g., using heart rate monitor) or at pre-defined time points
(e.g., during the ward round).
3.2.3 Contextual Relevance
The Contextual Relevance quality dimension (QD3) indicates whether process information
is relevant in a specific work context. Process information has a high contextual relevance
113
create
object
capture
t
t+1
t+2
use (1)
update
object
t+3
t+4
up-to-date
use (2)
t+5
t+6
time
out-of-date
Figure 7: Topicality of process information.
if it is needed to perform or support an activity. For example, for preparing the ward
round a doctor needs current diagnoses and medical instructions. The more precise a
work context can be defined the more accurate the contextual relevance can be determined.
Therefore it becomes necessary to consider not only process- and user-related information,
but also location-based, device-related and time information.
Unlike the overall relevance (cf. Figure 1 on page 3), the contextual relevance is not influenced by other quality dimensions. As an example consider again the preparation of the
ward round for which the doctor needs the patient record. Let us assume that the patient
record is punctually available. In this case the patient record has high contextual relevance
and high overall relevance. Let us assume that the patient record is not punctually available. In this case the contextual relevance is still high, but no overall relevance can be
identified since the quality dimension punctuality is not fulfilled.
3.2.4 Completeness
The quality dimension Completeness (QD4) indicates whether or not all parts of a complex process information (comprising several information parts) are available. In order
to perform the activity ”create medical instructions”, for example, different blood values
(together representing a process information ”blood values”) must be available. Process
information is incomplete if some parts of a process information are missing. It is important to mention that completeness thereby depends on the currently needed information,
i.e., it depends on the current work context. Picking up again our example, this does not
mean that all possible blood values must be available, but only those being needed for
current patient treatment.
3.2.5 Value-Added
The Value-Added quality dimension (QD5) indicates whether it is possible to increase
some ”value” (e.g., patient satisfaction, diagnostic accuracy) by using process information. For example, information about patient needs is value added because the fulfillment
of the needs increases patient satisfaction. The value-added amount is calculated as the difference between the value that can be realized without using specific process information
and the value that can be realized with specific process information. Figure 8 shows this
114
relationship. However it is quite difficult to determine the value-added quality dimension
since respective effects often cannot be exactly estimated.
without
use
50%
60%
use
70%
80%
patient
satisfaction
value-added
Figure 8: Value-Added of process information.
3.2.6 Appropriate Amount
The quality dimension Appropriate Amount (QD6) of process information indicates
whether the amount of available process information is sufficient. This is the case if
the amount meets the requirements of process participants. For example, a doctor needs
the name of the patient and pre-existing diseases. The appropriate amount of process
information is not sufficient if he gets the entire patient record. In practice, this problem
is solved by extracting process information via software (e.g., a document is divided into
individual information objects). In our case studies we analyzed the appropriate amount
of process information. Obviously, decision-makers are confronted with too much information. Knowledge-workers, by contrast, have the problem of being confronted with
insufficient information.
3.2.7 Granularity
a
f
c
methods for
aggregation
2xfindicates whether the aggregation of process
The Granularity quality dimension (QD7)
1x c
b meets
f
dthe requirements of process
information
users. Process information has the right
granularity if immediate use is possible (cf. Figure 9). For example, if a doctor needs to
know the average body temperature a patient had during the past week he should immediately get the calculated average value but not the individual values.
P. Miller
Day
01
02
03
04
05
Temp.
37.3
38.2
38.4
38.1
37.9
P. Miller
methods for
aggregation
Current Temp.:
37.9
Average Temp.:
38.0
Prognosis:
Figure 9: Granularity of process information.
According to [Jun06], three aggregation dimensions need to be distinguished: (a) time
dimension, (b) area-specific dimension, and (c) value and quantity dimension. As an example of the time dimensions consider the aggregation based on emergencies per day,
115
month, or year. Examples of the area-specific dimension include aggregation by organization units (e.g., number of patients on ward A) or patients (e.g., patients’ age and gender).
As an example for the value and quantity dimension consider aggregations relating to cost
centers (e.g., research and development, patient service).
Unlike the granularity, the appropriate amount (QD6) meets the requirements if nonaggregated information is provided. Let us assume that the doctor wants to know the
average body temperature of a patient. If individual data items are provided, only QD7
meets the requirements. If the average body temperature is provided, both QD6 and QD7
meet the requirements.
3.2.8 Neighbourhood
The Neighbourhood quality dimension (QD8) indicates how strong and how frequently
process information is linked to other process information. Process information which is
strongly and frequently linked tends to be more important. In addition, the semantic of the
relation is important. Examples include metadata-matching (e.g., author, keyword), text
similarity, and usage-pattern [WM09]. Figure 10 illustrates the relations between process
information along a simple example. Starting from a process information (doctor in our
example) we can identify the neighbourhood being relevant.
examination
document
patient
guideline
doctor
Jane Doe
project
planning
email
best practices
article
wiki
John Doe
expert
Figure 10: Neighbourhood of process information.
3.2.9 Methods-of-Use
The Methods-of-Use quality dimension (QD9) indicates how a process participant uses
the process information. Suitable use cases are, for example, to read, create, update, and
delete the process information. For example, a process user cannot use read-only process
information if he wants to edit process information.
4 Evaluating Quality Dimensions
Quality dimensions are an important means to determine process information quality and
the overall relevance of process information for a particular process participant. They are
thus also an important means to realize process-oriented information logistics.
116
Generally, there exist several approaches which can be used to decide whether process
information fulfills our defined quality dimensions: (a) algorithmic methods, (b) semantic technologies, (c) social methods, and (d) convergent methods. Algorithmic methods
are, for example, the vector space model, the term frequency algorithm, and methods
of clustering. The use of semantic technologies is another possibility to determine process information quality (e.g., via ontologies) [WM09]. Social methods include collaborative tagging or human-based rating of process information [WZY06]. Finally, convergent
methods improve the aforementioned methods through their combination (e.g., algorithmic detected relationships between process information are editable by process users).
Table 1 illustrates which of these methods can be used to determine our introduced quality
dimensions.
Table 1: Methods to determine process information quality.
Shortcut
Quality Dimension
Algorithmic
Semantic
Social
QD1
Punctuality
x
x
QD2
Topicality
x
x
QD3
Contextual Relevance
x
x
x
QD4
Completeness
x
x
x
QD5
Value-Added
QD6
Appropriate Amount
x
x
QD7
Granularity
x
x
QD8
Neighbourhood
QD9
Methods-of-Use
x
x
x
5 Related Work
Different authors investigate process-oriented information logistics in general. The stateof-the-art regarding information logistics is summarized in [DW09]. Empirical evidence
on benefits generated by information logistics concepts is provided in [BD08]. Contextawareness in particular is discussed by Haselhoff [Has05] and Meisen et. al [MPVW05].
Problems of data quality and solutions for data quality improvements are described in
[Red01]. Scannapieco et. Al [SVM+ 04] present an architecture for managing data quality
in cooperative information system.
Quality dimensions of information in general have been considered in literature. Jung
[Jun06], for example, investigates data integration architectures and also sketches quality
dimensions. Wang et. al [WW96, Wan98] identify aspects of data quality based on empirical research and integrate their findings into a data quality framework. Naumann et.
al [NLF99] describe a framework for multidatabase query processing taking information
quality into consideration. Table 2 compares our quality dimensions with the ones in the
aforementioned papers.
117
Table 2: Contextual Quality Dimensions from different viewpoints.
Shortcut
Quality Dimension
Our Paper
Jung
[Jun06]
Wang
[PLW02]
Naumann
[NLF99]
QD1
Punctuality
x
x
x
x
QD2
Topicality
x
x
x
x
QD3
Contextual Relevance
x
QD4
Completeness
x
x
x
x
QD5
Value-Added
x
QD6
Appropriate Amount
x
x
QD7
Granularity
x
x
QD8
Neighbourhood
x
QD9
Methods-of-Use
x
*
Relevance
x
*
Periodicity
x
*
Price
x
x
x
x
x
x
x
Wang subsume QD1 and QD2 under the term timeliness. In our paper, relevance is not a
separate quality dimension. Rather, relevance results from the combination of all quality
dimensions. Our QD3 is not influenced by other quality dimensions (cf. Section 3.2.3
on page 7). According to Jung, the relevance can be affected by other quality dimensions
(e.g., by QD1). In the paper of Wang, the influence towards the relevance is unclear. Wang
et. al [PLW02] define relevance as ”the extent to which information is applicable and
helpful for the task at hand”. Jung subsumed QD4 and QD6 under the term completeness.
In our research the quality dimension periodicity is based on the information sources and is
therefore not a quality dimension of process information. The quality dimension price can
be omitted because commercial data providers are not in focus. Naumann closely follows
Wang. Due to different perspectives some quality dimensions of Wang (e.g., value-added)
are omitted in the paper of Naumann.
6 Summary and Outlook
Enterprises are confronted with a continuously increasing data overload making it difficult
for them to provide the needed information to their employees. Thereby, relevant information is often closely related to the execution of business processes. Hence, the main
problem is to identify information being relevant for a process user and his activities in a
given work context. To solve this problem, enterprises crave for an intelligent and processoriented information logistics enabling them to provide the right process information, in
118
the right format and level of granularity, at the right place and accurate point of time to the
right people. To realize such process-oriented information logistics, quality dimensions
of process information adopt a key role. This paper picks up this issue and investigates
quality dimensions and discusses their role for process-oriented information logistics.
Future research includes an evaluation of the proposed contextual quality dimensions with
respect to the development of an intelligent, process-oriented information logistics. We
will investigate the factors determining the work context of process users. Only a precise
understanding of a work context allows us to accurately determine the overall relevance of
process information. Future research will also pick up again Table 1 and will address the
application of the discussed methods to determine our dimensions in detail.
References
[Ack89]
R. L. Ackoff. From Data to Wisdom. in: J. of Applied Systems Analysis, 16, pp. 3-9,
1989.
[AG04]
E. M. Awad and H. M. Ghaziri. Knowledge Management. Dorling Kindersley, 2004.
[BCGH06] P. Bocij, D. Chaffey, A. Greasley, and S. Hickie. Business Information Systems: Technology, Development and Management for the E-Business. Prentice Hall, 2006.
[BD08]
T. Bucher and B. Dinter. Process Orientation of Information Logistics - An Empirical
Analysis to Assess Benefits, Design Factors, and Realization Approaches. in: Proc. of
the 41st Annual Hawaii Int’l Conf. on System Sciences, pp. 392-402, 2008.
[DW09]
B. Dinter and R. Winter. Information Logistics Strategy - Analysis of Current Practices
and Proposal of a Framework. in: Proc. of the 42nd Hawaii Int’l Conf. on System
Sciences (HICSS-42), pp. 1-10, Hawaii, 2009.
[Has05]
S. Haseloff. Context Awareness in Information Logistics. PhD Thesis, Technical University of Berlin, 2005.
[HMR11]
M. Hipp, B. Mutschler, and M. Reichert. On the Context-aware, Personalized Delivery of Process Information: Viewpoints, Problems, and Requirements. in: Proc. of
the 6th Int’l Conf. on Availability, Reliability and Security (ARES’11),Vienna, 2011.
(accepted for publication).
[Jun06]
R. Jung. Architekturen zur Datenintegration. Deutscher Universitäts-Verlag/GWV
Fachverlage GmbH, 2006.
[LL09]
K. Laudon and J. Laudon. Management Information Systems: Managing the Digital
Firm. Pearson/Prentice Hall, 2009.
[LR07]
R. Lenz and M. Reichert. IT Support for Healthcare Processes - Premises, Challenges,
Perspectives. in: Data and Knowledge Engineering, 61(1), pp. 39-58, 2007.
[MB91]
G. C. Moore and I. Benbasat. Development of an Instrument to Measure the Perceptions of Adopting an Information Technology Innovation. in: Information Systems
Research, 2(3), pp. 192-222, 1991.
[MHHR06] D. Müller, J. Herbst, M. Hammori, and M. Reichert. IT Support for Release Management Processes in the Automotive Industry. in: Proc. of the 4th Int’l Conf. on Business
Process Management (BPM’06), pp. 368-377, Vienna, 2006.
119
[MMR11]
B. Michelberger, B. Mutschler, and M. Reichert. On Handling Process Information:
Results from Case Studies and a Survey. in: Proc. of the 2nd Int’l Workshop on Empirical Research in Business Process Management (ER-BPM 2011), Clermont-Ferrand,
2011. (accepted for publication).
[MPVW05] U. Meissen, S. Pfennigschmidt, A. Voisard, and T. Wahnfried. Context- and SituationAwareness in Information Logistics. in: Current Trends in Database Technology EDBT 2004 Workshops, pp. 335-344, 2005.
[MRB08]
B. Mutschler, M. Reichert, and J. Bumiller. Unleashing the Effectiveness of Processoriented Information Systems: Problem Analysis, Critical Success Factors and Implications. IEEE Transactions on Systems, Man, and Cybernetics (SMC) - Part C, 38(3),
pp. 280-291, 2008.
[NLF99]
F. Naumann, U. Leser, and J. C. Freytag. Quality-driven Integration of Heterogeneous Information Systems. in: Proc. of the 25th Int’l Conf. on Very Large Data Bases
(VLDB’99), pp. 447-458, Edinburgh, 1999.
[PLW02]
L. L. Pipino, Y. W. Lee, and R. Y. Wang. Data Quality Assessment. in: Commun. of the
ACM - Supporting community and building social capital CACM Homepage archive,
45 (4), pp. 211-218, 2002.
[Red01]
T. C. Redman. Data Quality: The Field Guide. Digital Press, 2001.
[Row06]
J. Rowley. The wisdom hierarchy: representations of the DIKW hierarchy. in: J. of
Information Science, 33(2), pp. 163-180, 2006.
[RT08]
R. K. Rainer and E. Turban. Introduction to Information Systems: Supporting and
Transforming Business. Wiley & Sons, 2008.
[SVM+ 04] M. Scannapieco, A. Virgillito, C. Marchetti, M. Mecella, and R. Baldoni. The DaQuinCIS Architecture: a Platform for Exchanging and Improving Data Quality in Cooperative Information Systems. in: Information Systems Journal, 29(7), pp. 551-582, 2004.
[TB98]
G. K. Tayi and D. P. Ballou. Examining data quality. in: Commun. ACM, 41(2), pp.
54-57, 1998.
[Wan98]
R. Y. Wang. A Product Perspective on Total Data Quality Management. in: Commun.
ACM, 41 (2), pp. 58-65, 1998.
[Wes07]
M. Weske. Business Process Management: Concepts, Languages, Architectures.
Springer, 2007.
[WM09]
J. Wurzer and B. Mutschler. Bringing Semantic Technology to Practice: The iQser
Approach and its Use Cases. in: Proc. of the 4th Int’l Workshop on Applications of
Semantic Technologies (AST ’09), pp. 3026-3040, Lübeck, 2009.
[WS96]
R. Y. Wang and D. M. Strong. Beyond Accuracy: What Data Quality Means to Data
Consumers. in: J. of Management Information Systems, 12(4), pp. 5-34, 1996.
[WW96]
Y. Wand and R. Y. Wang. Anchoring Data Quality Dimensions in Ontological Foundations. in: Commun. ACM, 39(11), pp. 86-95, 1996.
[WZY06]
X. Wu, L. Zhang, and Y. Yu. Exploring social annotations for the semantic web. in:
Proc. of the 15th Int’l Conf. on World Wide Web (WWW ’06), pp. 417-426, 2006.
120
Towards Automated Financial Process Auditing:
Aggregation and Visualization of Process Models
N. Mueller-Wickop, M. Schultz, N. Gehrke, M. Nüttgens
School of Business, Economics and Social Sciences
University of Hamburg
Max-Brauer-Allee 60
D-22765 Hamburg
niels.mueller-wickop@wiso.uni-hamburg.de
martin.schultz@wiso.uni-hamburg.de
nick.gehrke@nordakademie.de
markus.nuettgens@wiso.uni-hamburg.de
Abstract: Internal and external auditors face an enormous amount of financial
entries in accounting information systems. For many reasons - like legal
regulations - a process-oriented view of these entries is urgently needed in order to
understand the way financial entries are produced in accounting information
systems and to infer the underlying processes. Traditional modeling languages
focus on processes but pay no regard to the financial value-flows. Furthermore,
automated process retrieval approaches only reconstruct single process instances,
which need to be aggregated for reasons of comprehensibility, simplification and
clearness. The paper wants to close this gap and integrate the process with the
accounting perspective followed by an aggregation of single process instances. As
a result we present a visualization form capable of integrating the financial view
with process flows. In this way, auditors are able to trace how balance sheet items
have been produced in the system during the fiscal year.
1 Introduction
Current technological support for internal and external auditors is very limited. The last
decade painfully shows how weakly conducted audits result in unprecedented business
turbulences with corporate fraud and partly followed by collapse (years of incidents:
Enron 2001; MCI WorldCom 2002; Parmalat 2003; AIG 2004; Fannie Mae 2006;
Satyam 2009). Combined with the so-called Financial Crisis beginning in 2008 and
ongoing uncertainty for the global economy, political as well as scientific focus is on the
way audits are done nowadays [EU10]. Even though auditors are increasingly
recognized as playing a critical role within companies, their repertoire of supporting
tools and methods is out-dated. To (at least partly) remedy this deplorable state of affairs
is the main objective of the paper at hand.
Todays internal and external audits focus on processes [Be97][Ru03][Ru06]. Within the
bounds of process auditing an automated approach is to be developed. In this context,
different aspects need to be taken into account. First, an automated approach for a
121
retrograde reconstruction of process- and value-flows from Enterprise Resource
Planning (ERP) and Accounting Information Systems (AIS) must be developed. This has
been done by introducing Financial Process Mining [GM10]. The automated retrieval of
process instances (a process instance can be described as the representation of one
execution of a business process) from system data is followed by the presentation of
mined results. Modeling these results needs to fulfill different requirements for different
groups of stakeholders. This paper is focused on internal and external auditors as
stakeholders. Additional stakeholders include - but are not limited to - business process
managers, process owners, risk managers, the board of directors and the audit
committee. Both, internal and external auditors, have mainly overlapping interests.
In the case of process audits an integrated view of value-flows and process flows is
essential. While process flows are in the foreground, the corresponding value-flows also
have to be considered in terms of risk and materiality (for a definition see ISA 320.3
[IFAC10]). An adequate visualization integrating both views - has been developed and
is described in chapter 4 Related Work. In order to gain an overall-view of the process
flow including the financial flows, mined process instances need to be aggregated into
process models.
This paper was written to take an important step towards automated process audits. In a
first step the motivation for this research problem will be explained to introduce the
reader to this domain (section 2). For a further understanding the research method
(section 3) as well as related work (section 4) will be presented. Following this
introduction, an approach for the aggregation of financial process instances will be laid
out (section 5). We conclude this paper with a summary and an outlook on work of
future research in this field (section 6).
2 Motivation
An integral part of a modern approach for auditing companys financial statements is the
process audit [Ru06]. This is justified by the fact that conventional auditing methods
(substantive audit procedures performed for single business transactions, based on
Analytics and Test of Detail) seem more and more impracticable from an efficiency
point of view in a world where the number of business transactions is dramatically
increasing and all data is electronically available in ERP systems. The argument behind
process auditing is that all business transactions running through well designed and
controlled business processes will be properly represented in the financial statements of
the company [Be97] [Ru06]. Following this approach the design of the business
processes including relevant controls are reviewed. Furthermore the compliance with
defined processes throughout the year is checked for a sample of business transactions.
The audit results are then applied to all business transactions ran through the
corresponding process in the financial period under investigation.
Process audits utilize interviews as means to survey the business processes of a company
[Ru03]. A number of employees involved in a business process are interviewed. Based
on the information gained through these interviews auditors are modeling their
understanding of the actual processes including relevant controls with the help of
flowcharts and/ or narratives [Ru03] (ISA 315 [IFAC10]). Software supporting the task
122
of business process survey is rarely used in current audit approaches. There are several
severe issues in this approach. One is the influence of perceptions, i.e. involved
individuals are only able to express their subjective perception of reality or the gained
information is normative in the sense that it states what is expected to be done rather
than describing the actual process [Aa02]. Secondly an auditor has to rely on the
information he obtains by the employees regardless if it is potentially erroneous on
purpose or unintentionally. Thirdly interviews are time-consuming and therefore come
along with high costs. Last and from audit perspective the main point is that processes
are not derived from the business transactions stored in ERP systems itself although the
results of the process audit are applied to them later. This is the key problem our process
mining approach for financial audits is aiming at. Therefore the objective is the
development of an automated approach for a retrograde reconstruction of processes from
business transaction data stored in ERP and AIS.
Although motivated by efficiency arguments process audits are also obligate by
international standards1 as they are considered as important steps for understanding the
clients business and so forming the basis for a well-founded audit of the financial
statements. Starting point for this audit methodology was the introduction of the
Business Risk Audit (BRA) by Bell et al. in 1997 [Be97]. BRA can be characterized as a
top down audit approach starting with an analysis of the business strategy, significant
business transactions and business risks of a client. Subsequently the key processes of
the client are identified and analyzed regarding their conformance to business goals,
handling of significant business transactions and the coverage of identified business
risks. Detailed (transaction oriented) audit procedures are than effectively planned and
performed based on the comprehensive knowledge gained during the previous steps. The
BRA aims at focusing on audit procedures related to areas which have been identified as
being exposed to high audit risks. The core idea of BRA is that a better understanding of
the clients business significantly correlates with a better understanding of the audit risks
[Be97]. According to the International Standard on Auditing (ISA) 200 The risk that the
auditor expresses an inappropriate audit opinion when the financial statements are
materially misstated is defined as audit risk. Therefore an important method for
focusing the audit activities on areas with a high audit risk is the concept of materiality.
According to ISA 320 (Materiality in planning and performing an audit) materiality is
defined as follows Misstatements [in the financial statements], including omissions, are
considered to be material if they, individually or in the aggregate, could reasonably be
expected to influence the economic decisions of users taken on the basis of the financial
statements. [IFAC10]. Consequently for each audit a materiality (quantitative and/ or
qualitative measure) is to be determined by the auditor. The materiality is than applied in
planning and performing an audit of the financial statements (ISA 320 [IFAC10]).
The Big Four2 turned towards the BRA, not only for its higher efficiency (reducing the
amount of substantive audit procedures and test-of-details, as mentioned above), but also
1
For instance ISA 315 states that the auditor should obtain an understanding of the information system,
including the related business processes, relevant to financial reporting ( ) ISA 315.81 [IFAC10].
2
The Big Four are the four largest international accountancy and professional services firms: Ernst & Young
(E&Y), Deloitte Touche Tohmatsu (Deloitte), Klynveld Peat Marwick Goerdeler (KPMG) and
PricewaterhouseCoopers (PwC)
123
because of an increased value added for the client as well as a stronger link between risk
management and audit [Ru06].
To support the auditors planning activities especially the decision which processes/
process variants will be subject to detailed audit procedures in due consideration of
materiality our automated approach for a retrograde reconstruction of processes will not
only consider the process flow itself but also integrate the corresponding value-flows.
Two steps are necessary for this reconstruction task. First single business transactions
which can be seen as process instances need to be reconstructed from the data stored in
ERP systems. Secondly these process instances must be aggregated to infer the
underlying processes/ process variants. The first step is described in [GM10]. This paper
is dealing with the second step.
3 Research Method / Research problem and research methodology
Following the guidelines for Design Science Research in Information Systems
[He04][MS95], this paper is focused on developing a relevant IT artifact (constructs,
models, methods and implementations). This artifact represents a domain-specific
modeling notation and rules for aggregation of process instances constructed from
financial entries stored in ERP systems. Artifacts facilitate the analysis, design,
implementation and use of information systems [He04][De97].
The research idea came from the awareness of a problem that became apparent in current
audit practice (inductive reasoning) regarding process audits in the context of year-end
audits. Referred to [Pe07] a problem-centered approach is on hand. The research
problem can be described as follows. Process audits are in fact an integral part of the
audit of the financial statements of a company. However process audits are based on
information not directly derived from the financial entries constituting as a whole the
financial statements. Process audits are performed based on process models constructed
with the help of qualitative methods of collecting data (e.g. guided interviews).
Nonetheless the results of the process audit are used to develop the auditors opinion on
the financial statements. There is a gap between the data source the audit results are
based on (process audit) and the data source the audit results are applied to (financial
statements). Theoretical considerations based on a literature review (deductive
reasoning) helped to achieve a further understanding of the problem.
Starting from this research question following objectives were derived for an artifact to
be designed. There should be an automated approach for a retrograde reconstruction of
process instances and corresponding value-flows from financial entries stored in ERP
systems. Furthermore, there should be a modeling notation representing the
reconstructed process instances and aggregation rules to infer a generalized view on all
process instances constituting the financial statements.
In this paper a BPMN-based notation capable of integrating the financial view with the
process flows is proposed. In [GM10] an algorithm is described for reconstructing
process instances from financial entries. Based on this, aggregation rules for process
instances are developed. These rules were developed with individual sample process
instances from a standard SAP IDES system and are presented as narratives.
124
4 Related Work
4.1 Financial Process Mining
Financial Process Mining is a subset of the Process Mining family in which all kind of
process, control, data, organizational and social structures are discovered in log files or
other data stored in information systems [Aa11]. According to van Dongen and van der
Aalst the goal of Process Mining is the construction of a formal model describing a set
of real life executions [DA05]. By now, the research field of Process Mining as a whole
is a fairly well researched. The first research was done in 1995 by Cook and Wolf for the
software engineering process [CW95]. In 1998, process mining was adapted to
Workflow Systems by Agrawal and Gunopulos [AGL98]. A full overview until 2002 is
given by van der Aalst et al. in [Aa02]. Up to now most of the Process Mining
techniques have been developed for event log data stored in different kind of information
systems (e.g. workflow management systems, ERP or AIS) and therefore focusing on
process flow reconstruction [Ra06]. Most research in this domain focuses on heuristics
based algorithms searching for an order of relations between events in event logs
[Me03]. Recent approaches like decision mining and social network/ organizational
mining tend to broaden the considered data sources [RV06][VS04]. However, data of
financial entries stored in ERP systems instead of using event logs as data source is not
yet considered. In [GM10] an algorithm is described which mines financial entries and
reconstructs the corresponding process instances by using information from the open
item accounting. These process instances are the basis for the aggregation and inference
of the underlying process models described in this paper.
A number of approaches for aggregating resp. merging process instances are suggested
in literature. [Ro10] provide an algorithm that produces a single configurable process
model encompassing the behavior of the input models. [GVJ08] use an abstraction called
functional graph to merge event driven process chains. [SKY06] define four types of
merge (sequential, parallel, conditional, and iterative) and describe corresponding
algorithms for performing the merge operations. [LCW09] describe a heuristic algorithm
to construct a generic process model from process variants which minimizes the average
distance between the generic model and the process variants. [DDA06] offer three
algorithms for discovering process models from process runs which differ regarding the
information contained in these runs. In [DA05] a multi-step approach for inferring an
overall process model from single processes instances is suggested. Moreover
frameworks for comparing different merge approaches and for merging incomplete and
inconsistent graph-based views are presented in [Br06] and [SE06]. However, these
approaches do not fulfill the requirements from an audit perspective outlined in section
5.1 Data source, Requirements and Illustration.
4.2 Modeling Languages - BPMN(-Finance)
At present, there is a broad variety of different process modeling languages in research
and practice. The main representatives are the Business Process Modeling Notation
(BPMN) [BPMI04] and the Event Driven Process Chain (EPC) [SN95][SN00] [KNS92].
Other modeling languages include the Integrated Definition (IDEF) [MM06] [KBS10],
125
Petri Nets [Pe62] and the Unified M
Modeling Lan
nguage (UML
L) [OMG10] [Ru04]. In
uages, see e.g. [Za01] [KLL
L09].
addition thhere are further scientific moodeling langu
Considerinng the objectiv
ve of this papper only a mod
deling languag
ge capable off integrating
the financiial- and process-dimension is appropriatee. Subsequenttly a notation is proposed
to depict financial pro
ocess instancees including value flows using BPM
MN. For an
S
was chossen as test ER
RP System. Th
herefore all pro
roperties are
evaluation see [LK06]. SAP
w their SAP terms. The BPMN-Finan
nce notation described below
w will later
displayed with
be used forr an illustrativ
ve example off the aggregation algorithm.
B
BPMN-Finan
nce
BPMN activities aree business actiivities. In SAP
P these are im
mplemented
I this examp
ple the transaaction name
as transsactions or suubprograms. In
(Autom
matic Paymennt), the userr who carrieed out the transaction
(OLBE
ERT), the acttual transactiion code (F1
110) and thee document
numberr which wass created when
w
the transaction wass executed
(200000
00531) are shhown. It is posssible to add any
a data of thee document
header. Moreover, it is feasiblee to mark trransactions rrepresenting
control activities by ccolor coding them
t
.
Events are items onn accounts. The
T first num
mber represent
nts the item
ment (001), th
he number beelow indicatess the actual
numberr of the docum
amountt which was pposted to the account (3.14
41 ). In adddition to the
conventional BPMN
N meaning (staart-, intermed
diate- and endd-event) the
color off an event hass a meaning:
Blue
Yellow
The item has been pposted to a baalance sheet acccount on thee asset side.
m
an addittional asset haas been posteed or a liabilitty has been
This means
compen
nsated (= Debbit).
The iteem has been pposted to a balance
b
sheet account on tthe liability
side. Th
his means an additional liaability has beeen posted or aan asset has
been co
ompensated (=
= Credit).
Green
The item
m is an incom
me posting (= Debit,
D
Profit & Loss is affeccted).
Red
The item
m is an expennse entry (= Crredit, Profit & Loss is affeccted).
BPMN sequence floows connect activities,
a
gateeways and evvents. Their
n is to speccify the process flow in BPMN moddels. In the
function
proposeed notation thhe sequence flows
f
connectt transactions (activities)
and iteems on accouunts (events). The processs flow does not always
reflect the
t chronologgical order of the
t processing
g in an ERP syystem (here
in SAP
P). For a bettter understand
ding our notaation depicts the logical
order of events and aactivities. Ano
other advantag
ge of this apprroach is the
ndence of a pparticular ER
RP system, ass different syystems may
indepen
have deeviant internall processing lo
ogic.
Regardless of the pprocess flow
w, a connectio
on between an activity
(transacction) and an event (item) indicates thatt the item wass posted by
126
this traansaction. Acttivities within
n accounts (ssee below) coonstitute an
exception. They reprresent a clearring of open items.
i
Includiing them is
i technically possible posinng a risk of
necessaary as a manuual execution is
incorrecct assignmentt. Edges betw
ween events deepict the assiggnment of a
clearing
g item and thee correspondin
ng cleared item
m.
BPMN groups repre sent accountss in BPMN-Finance. In the top middle
the acccount numberr and the acccount name is shown. Thee design is
following conventioonal accountin
ng practice. By
B coloring tthe account
i
open-iteem managed aaccounts are indicated.
Orange
The fin
nancial accounnt is not involv
ved in open iteem accountingg.
Black
Financiial account is iinvolved in op
pen item accounting.
Table 1: BPMN
N-Finance mod
deling elements
5 The Fiinancial Process Data Aggregation (FPDA)) Algorithm
m
5.1 Data source, Requiirements and Illustration
Process innstances recon
nstructed from
m financial entries
e
can be
b described as directed
graphs. Nodes
N
are cllassified in two fundam
mentally diffeerent groups:: activities
(transactions in SAP) an
nd events (iteems in SAP). Items are gro
ouped into acccounts. In a
simplified representation of a processs instance acccounts are deepicted as a nnode, not a
m. These two groups are ddifferentiated by
b their propeerties. Each aaccount and
single item
each transaaction can be included more
re than once in
n the same pro
ocess instancee. A process
instance sttarts and endss with 1-n acccounts. Accounts are linkeed by transacctions. Each
transactionn can be linkeed to 2-n accoounts. Accoun
nts cannot be linked
l
directlyy with each
other. The linkage is dep
picted by edgges representin
ng the processs flow within the process
P
instan
nces include no cycles ressp. can be reepresented as an acyclic
instance. Process
directed grraph. Reason is
i that the linkkage between
n transactions and accounts is based on
financial entries.
e
A finaancial entry caannot be posteed twice, therre are uniquelyy identified
by their doocument numb
ber in the ERP
P systems.
From a daata perspective following ccan be stated:: for inferring
g the aggregat
ated process
model the data source is
i globally coomplete. Conssidering the domain
d
of finaancial audit
od (e.g. fiscal year) is relevaant. Hence theere are a finitee number of
only a certtain time perio
relevant finnancial entriees which can bbe determined
d in the ERP system and uused for the
aggregation. Provided that the used ERP system fulfills generaally accepted accounting
principles the
t source datta is free from
m noise.
For the agggregation alg
gorithm follow
wing requirem
ments can be specified froom an audit
perspectivee:
Completenness: Each pro
ocess sequencee being part of
o the process instances whiich are used
for the agggregation need
d to be includded in the aggrregated process model. In tthis context
sequence means
m
a possib
ble way in a pprocess instancce (graph).
127
Rationale: An auditor has to give an opinion on the financial statements as a whole.
Hence, the method of materiality would suggest reducing the aggregated processes
model to process sequences with a value flow on or above materiality level. But auditors
are obliged to apply also qualitative materiality measures and are forced to put specific
attention to sources for material misstatements (e.g. accounting errors, fraud) (ISA 240,
ISA 320, [IFAC10]). Therefore rare or abnormal process sequences are potentially of
high interest for the auditors opinion.
No new process sequences: Exclusively process sequences included in the process
instances which are used for the aggregation should be incorporated in the aggregated
process model.
Rationale: As described above only a certain time period (in most cases a fiscal year)
with finite number of financial entries is relevant from an audit perspective. The
aggregated process model should only contain process sequences at least one financial
entry can be associated with. Otherwise the process sequence was not executed in real
world and is therefore not relevant for the audit.
No cycles: The aggregation algorithm should not create cycles.
Rationale: As the process instances are acyclic itself the aggregated process model
should also be acyclic to reflect the characteristics of the underlying process instances
and consequently the characteristics of the underlying financial entries.
Given the requirements listed above algorithms mentioned in the literature can be
evaluated regarding their applicability. As said before the focus of most research in the
domain of process mining is on heuristics based algorithms. Statistical and heuristic
approaches do not necessarily ensure completeness as certain sequences in the process
instances may have a low probability and therefore remain undetected [Me03]. In
addition, as the data source in the domain of financial processes is globally complete
purely algorithmic approaches seem to be more appropriate [CW98][SKY06][Ro10]
[GVJ08][DA05][DDA06][Me04]. However, these approaches either create cycles in or
add additional behavior (new process sequences which are not included in the input
process instances) to the aggregated process models. As explained above these are
undesired properties in our use case.
Therefore we present a new aggregation algorithm. It operates with process instances
reconstructed according to the mining algorithm introduced in [GM10]. For
demonstration purposes a payment run instance and a manual payment instance were
chosen and simplified. The automatic payment run has three generic activities: 1.
Receive goods, 2. Receive invoice and 3. Pay invoice. The actual mined process
instance had 26 Post Goods Receipt for PO transactions and 26 corresponding Post
invoice document transactions. Out of these 26 concurrent process flows, two were
randomly chosen. The manual payment instance has the same three generic activities.
However, for the 3rd activity a different SAP transaction was used. The used accounts
are also the same besides two Accounts: 400100 - Purchase Raw M. and 305000 Packaging Material The basic process instances look like the following (please note the
sequence flow from the right to the left), see Figure 1:
128
Figure 1: Initial process instances (22 accounts;
a
10 transactions)
5.2 Definin
ng start and end
e events
Start- and end-events
e
(ittems) need to meet at least one
o of the folllowing two coonditions:
1. The account theey are posted to is not involved in open-iitem-accountinng
m
2. There is no cleaaring-documeent for this item
129
For all ideentified itemss (e.g. debit iitem) meeting
g one of thesse conditions it is to be
determinedd whether theey are start- or end-eventts. This is done by exam
mining their
offsetting item
i
(e.g. cred
dit item). If thhis offsetting item cleared an
a item the seelected item
is an end-eevent. On the contrary, if thhe offsetting item
i
got clearred the selecteed item is a
start-eventt. By applying
g this rule all start- and end-events are identified
i
andd marked as
such.
5.3 Definin
ng internal, start
s
and end
d nodes for the algorithm
All accounnts either exclusively contaaining items which are ideentified as staart- or endevents are marked as staart or end noddes depending on their items. All other acccounts and
a marked as
a process moodel internal nodes disting
guished by thheir names.
activities are
Accounts are
a named aft
fter their accouunt names, acctivities are named
n
after thheir activity
names.
For demonnstration reasons a genericc model of th
he payment run
r instance ddepicted in
Figure 1 iss created. Accounts are reppresented by nodes named
d with capital letters and
activities / SAP transacttions are repreesented by nodes named wiith numbers. T
The generic
F
2 repressents the abovve presented model.
m
Green nodes
n
are start
rt nodes, red
model in Figure
nodes are end
e nodes and
d yellow node s are internal nodes.
Figure 22: Generic proceess model
5.4 Perforrm the aggreg
gation
Apply rule no1: All iden
ntical start andd end nodes are merged.
Merging accounts
a
is do
one by insertiing all items to one combiined account.. In general
activities are
a also unitted straight aaway. Please note that so
ome transactioon specific
information could be lost
l
when meerging activitties, e.g. the user who exxecuted the
ocess model inn Figure 3 illu
ustrates this steep:
activity. Thhe generic pro
130
Figure 3: Generic
G
processs model after merging
m
start and
d end nodes
After mergging all start and end noddes the algoriithm identifies all possiblee sequences
from start to end nodes. When referrring to sequen
nces a possiblle way from a start to an
end node is meant.
Apply rulee no2: All ideentical sequennces (includin
ng the predecessor sequencces of each
node) begiinning from th
he start nodes are merged.
Each identtified sequencce starting from
m a start nodee is compared
d to all other ssequences the longest possible maatch between each pair is marked for merger
m
if all ppredecessor
a identical. T
The generic
sequences of each node included in thhe two selecteed sequences are
process moodel in Figure 4 illustrates tthis step:
Figurre 4: Generic prrocess model affter applying rule no2
v
node sequences annd therefore
In theory, a complex prrocess model would have various
c
to be comparred would be enormous.
e
Forr real world
the possiblle number of combinations
data we did not experience this as a problem. Sophisticated
S
process instannces which
result in a great number of comparisoons were very rare.
Apply rule no3: All iden
ntical sequencees (including the successor sequences off each node)
f
the startt and end nodees are merged
d.
beginning from
First the allgorithm comp
pares all identtified sequencces beginning with a start nnode to each
other. The longest matcch between eaach sequence is marked forr merger if alll successor
a identical. This step is
sequences of each node included in thhe two selecteed sequences are
del in Figure 5 illustrates thhis step:
repeated foor all end notees. The genericc process mod
Figurre 5: Generic prrocess model affter applying rule no3
131
Transform
ming the abovee shown geneeric process model
m
back in
n the originall format we
obtain thee process mo
odel presentedd in Figure 6. As you can see a cconsiderable
simplificattion of the orriginal processs instances is achieved. Never
N
the lesss all initial
process innstances are fully includeed and no new
n
process sequences aare created.
Furthermore, it is evideent that a purrchase of raw
w material (acccount 4001000 Purchase
ment (activity
y/ SAP transaaction FB01
Raw M.) is always acccompanied byy manual paym
M Schultz. O
On the other hand, a purch
hase of invenntory goods
"Post docuument") by Mr.
always com
mes along with
h an automatiic payment run
n.
Fig
gure 6: Aggregaated process moodel (11 accoun
nts; 6 activities / SAP transactioons)
6 Summary and Fu
uture Work
k
In this papper we have provided an approach for the visualizaation and agggregation of
process innstances extraacted from E
ERP systems or AIS. The developmen
ent and the
demonstrattion were based on SAP ass a well-know
wn and widely
y-used ERP syystem. With
this approaach two goalss have been acchieved. Firsttly, process flows and valuue-flows are
integrated in a notation (BPMN-Finannce). Secondlly, rules for th
he aggregationn of process
w
provided
d to infer a leess complex process
p
model representing all process
instances were
instances considering
c
au
udit requirem
ments. The visu
ualization of financial proccesses shall
help externnal and intern
nal auditors too shift their viiew towards financial
f
proccesses: from
financial statements being a set of iteems on accou
unts to a process-oriented pperspective.
ns which proccesses and pro
ocedures produ
uced the finanncial entries
This perspective explain
fore the financcial statementts throughout the fiscal year. Using an appropriate
and therefo
notation annd aggregation this processs view aims at
a a better and
d faster underrstanding of
the interdeependencies and
a procedurees within the accounting function
f
by aallowing an
easier cognnitive processsing. Future w
work will inclu
ude research in the area off rule based
abstractionn and documen
nt property orriented visualization.
The underllying project for
f this articlee is funded by
y the BMBF (G
German Federral Ministry
of Educattion and Research) andd supervised by PT-DL
LR (project references:
01IS100411B).
132
References
[Aa02]
Van der Aalst, W. M. P. et al.: Workflow Mining: a Survey of Issues and Approaches,
Beta: Research School for Operations Management/ Logistics, Working Paper 74, 2002
[Aa11] Van der Aalst, W. M. P.: Process Mining: Discovery, Conformance and Enhancement of
Business Processes, Springer Verlag, Berlin, 2011
[AGL98] Agrawal, R.; Gunopulos, D.; Leymann, F.: Mining Process Models from Workflow
Logs, Advances in Database Technology - EDBT'98. Springer Berlin, Volume
1377/1998.page 469, 1998.
[Be97] Bell, T. B.et al.: Auditing organizations through a strategic-systems lens, University of
Illinois at Urbana-Champaign, KPMG Peat Marwick LLP, 1997
[BPMI04] BPMI (Business Process Management Initiative): Business Process Modeling Notation
(BPMN) http://www.bpmn.org/Documents/BPMN_V1-0_May_3_2004.pdf .retrieved
25.02.2010, 2004
[Br06] Brunet, G. et. al.: A manifesto for model merging. Proceedings of the 2006 international
workshop on Global integrated model management, ACM (2006), 512.
[CW95] Cook, J. E.; Wolf, A. L.: Automating Process Discovery through Event-Data Analysis,
Proceedings of the 17thinternational conference on Software engineering, Seattle, pp. 73
82, 1995
[CW98] Cook, Jonathan E.; Wolf, Alexander L.: Discovering models of software processes from
event-based data, ACM Transactions on Software Engineering Methodology, 7 (3), pp.
215-249, 1998
[DA05] Van Dongen, B.F.; van der Aalst, W.M.P.:Multi-phase Process mining: Aggregating
Instance Graphs into EPCs and Petri Nets, 2nd International Workshop on Applications
of Petri Nets to Coordination. Workflow and Business Process Management (PNCWB).
ICATPN 2005
[DDA06]Van Dongen, B.F.; Desel, J.; van der Aalst, W.M.P.: Aggregating Causal Runs into
Workflow nets, BETA Working Paper Series, WP 173, Eindhoven University of
Technology, 2006
[De97] Denning, P. J.: A new social contract for research. Communications of the ACM.
Volume 40 Issue 2. Feb. 1997, 1997
[EU10] European commission, Green Paper - Audit Policy: Lessons from the Crisis http://eurlex.europa.eu/LexUriServ/LexUriServ.do?uri=COM:2010:0561:FIN:EN:PDF retrieved
12.06.2011, 2010
[GM10] Gehrke, N.; Mueller-Wickop, N.: Basic Principles of Financial Process Mining: A
Journey through Financial Data in Accounting Information Systems. In: Proceedings of
the 16th Americas Conference on Information Systems, Lima, Peru, 2010
[GVJ08] Gottschalk, F.; Aalst, W.M.P.; Jansen-Vullers, M.H.: Merging Event-Driven Process
Chains. In: (R. Meersman, Z. Tari) On the Move to Meaningful Internet Systems: OTM
2008. Springer Berlin Heidelberg, Berlin, Heidelberg, 2008, pp. 418-426.
[He04] Hevner, A. R. et al.: Design Science in Information Systems Research. In: MIS Quarterly
28(1), pp. 75-105, 2004
[IFAC10] International Federation of Accountants (IFAC): 2010 Handbook of International
Quality Control, Auditing, Review, Other Assurance, and Related Services
Pronouncements Part I, ISBN 978-1-60815-052-6, 2010
[KBS10] Knowledge Based Systems, Inc (KBSI): All IDEF Method Reports (KBSI)
http://www.idef.com/Downloads.htm retrieved 02.August 2010, 2010
[KLL09] Ko, R.K.L.; Lee, S.S.G; Lee, E. W.: Business process management (BPM) standards: a
survey, Business Process Management Journal. Vol. 15.Iss: 5. pp.744 791, 2009
[KNS92] Keller, G.; Nüttgens, M.; Scheer, A.-W.: Semantische Prozessmodellierung auf der
Grundlage
Ereignisgesteuerter
Prozessketten
(EPK).
Technical
Report.
Veröffentlichungen des Instituts für Wirtschaftsinformatik (IWi), 1992
133
[LCW09]Li, C.; Reichert, M.; Wombacher, A.: Discovering Reference Models by Mining Process
Variants Using a Heuristic Approach. In, (U. Dayal, J. Eder, J. Koehler and H.A.
Reijers): Business Process Management. Springer Berlin Heidelberg, 2009, pp. 344-362.
[LK06] List, B.; Korherr, B.: An evaluation of conceptual business process modeling languages.
Proceeding SAC 06. Proceedings of the ACM symposium on Applied computing, 2006
[Me03] Medeiros, A.K.A. de; van der Aalst, W.M.P.; Weijters, A.J.M.M.: Workflow Mining:
Current Status and Future Directions In R. Meersman et al. (Eds.):
CoopIS/DOA/ODBASE 2003, LNCS 2888, pp. 389406, 2003. Springer-Verlag Berlin
Heidelberg 2003
[Me04] Medeiros, A.K.A de; van Dongen, B.F.; van der Aalst, W.M.P.; Weijters, A.J.M.M.:
Process Mining: Extending the alpha-algorithm to Mine Short Loops. BETA Working
Paper Series, WP 113, Eindhoven University of Technology, Eindhoven, 2004
[MM06] Menzel, C.; Mayer, R.J.: The IDEF Family of Languages. Springer. Berlin, 2006
[MS95] March, S.; Smith, G. F.: Design and natural science research on information technology.
Decision Support Systems. Volume 15. Issue 4. December 1995. pp. 251-266, 1995
[OMG10] Object Management Group (OMG). OMG Unified Modeling Language (OMG UML),
Infrastructure. Version 2.3, http://www.omg.org/spec/UML/2.3/Infrastructure/PDF/
retrieved 02.August 2010, 2010
[Pe07] Peffers, K; Tuunanen, Z.; Rothenberger, M. A.; Chatterjee, S.: A Design Science
Research Methodology for Information Systems Research, Journal of Management
Information Systems, Issue: Volume 24, Number 3, Pages: 45 77, 2007
[Pe62] Petri, C. A.: Kommunikation mit Automaten, Darmstadt, T. H., F. f. Math. u. Physik,
Diss. v. 20. Juni 1962 (Nicht f. d. Aust.) III, p. 128, 1962
[Ra06] Ramesh, A. (2006). Process Mining in PeopleSoft. Master's Thesis. Technische
Universiteit
Eindhoven.
Eindhoven.
The
Netherlands
http://prom.win.tue.nl/research/wiki/blogs/pub2006/process_mining_in_peoplesoft
[Ro10] Rosa, M. et. al.: Merging Business Process Models. In (R. Meersman, T. Dillon und P.
Herrero): On the Move to Meaningful Internet Systems: OTM 2010. Springer Berlin
Heidelberg, Berlin, Heidelberg, 2010, pp. 96-113.
[Ru03] Russell, J.P.: The Process Auditing Techniques Guide: A Pocket Guide. ASQ Quality
Press, Milwaukee, 2003
[Ru04] Rumpe, B.: Modellierung mit UML. Berlin/Heidelberg. Springer. XIII, 2004
[Ru06] Ruhnke, Klaus: Business Risk Audits: State of the Art und Entwicklungsperspektiven,
Journal für Betriebswirtschaft, 56 (4), pp. 189-218, 2006
[RV06] Rozinat, A.; van der Aalst, W.M.P.: Decision Mining in Business Processes in BETA
Working Paper Series, WP 164, Eindhoven University of Technology, Eindhoven, 2006
[SE06] Sabetzadeh, M.; Easterbrook, S.: View merging in the presence of incompleteness and
inconsistency. Requirements Engineering 11, 3 (2006), pp. 174-193.
[SKY06] Sun, S.; Kumar, A.; Yen, J.: Merging workflows: A new perspective on connecting
business processes. Decision Support Systems 42, 2 (2006), pp. 844-858.
[SN00] Scheer, A.-W.; Nüttgens, M.: ARIS Architecture and Reference Models for Business
Process Management. Business Process Management. Springer. Berlin / Heidelberg, pp.
301-304, 2000
[SN95] Scheer, A.-W.; Nüttgens, M.: Business Process (Re-) Engineering: Architecture,
Reference Models and Toolset, Kasprzak, T. (Ed.) Information Technologies in Business
Reengineering, Uniwerytet Warszawski, Studia Informatyki Gospodarczej. Warschau.
pp. 31-39, 1995
[VS04] van der Aalst, W.; Song, M.: Mining Social Networks: Uncovering Interaction Patterns
in Business Processes. In Business Process Management, Lecture Notes in Computer
Science, pp. 244-260, Springer Berlin / Heidelberg, 2004
[Za01] Zamli, K. Z.: Process Modeling Languages: A Literature Review, Malaysian Journal of
Computer Science, 14 (2), pp. 26-37, 2001
134
A Reference Architecture for
Semantic Content Management Systems
Fabian Christ, Benjamin Nagel
s-lab - Software Quality Lab
University of Paderborn
Warburger Str. 100
D-33098 Paderborn
fchrist@s-lab.upb.de
bnagel@s-lab.upb.de
Abstract: Content Management Systems (CMS) lack the ability of managing semantic information that is part of the content stored in a CMS. On the other hand, a lot
of research has been done in the field of Information Extraction (IE) and Information Retrieval (IR), respectively. Additionally, the vision of the Semantic Web yields
to new software components that make semantic technology usable for application
developers. In this paper, we combine IE/IR concepts with the technologies of the
Semantic Web and propose a new family of CMS, called Semantic CMS (SCMS),
with advanced semantic capabilities. We provide a reference architecture for SCMS
and prove its value along two implementations. One implementation was created as
part of the Interactive Knowledge Stack research project and another one in a one-year
student project exploring the design of an SCMS for the software engineering domain.
1 Introduction
The growing importance of an efficient content management is omnipresent in nearly each
aspect of daily business. The challenges posed by the raising amount of content are wellknown [AL99] and are becoming greater due to the high availability of content from various sources like the Internet. Through this, more and more unstructured content needs to
be handled and therefore aggravate the management of content by Content Management
Systems (CMS).
Addressing these challenges, the “Semantic Web Wave” [LHL01] brought up several technologies that enable the definition of semantics as machine-readable metadata, termed
“semantic metadata”. Recent approaches [Sar08], like named-entity recognition, clustering and classification use the foundations and facilitate the automatic enrichment of
content with semantic metadata. The semantic metadata allow the automatic structuring
This work has been part-funded by the European Commission under grant agreement FP7-ICT-2007-3/ No.
231527 (IKS - Interactive Knowledge Stack).
135
Presentation & Interaction
Semantic Lifting
Knowledge Representation
& Reasoning
Persistence
Figure 1: The four layers of an SCMS
and thereby the further processing like searching or filtering. In addition, these metadata
enable the reasoning of new knowledge based on the content.
Summarizing the research trends of the last years, several technological foundations and
algorithms have been developed to face the challenges of efficient content and knowledge
management. On the other hand, only little attention has been paid to the integration of
these approaches from a software engineering perspective. The integration of semantic
functionalities has significant impact on the architecture. The CMS architecture has to
be extended in comparison to traditional CMS, because the storage and processing of
semantic metadata in a semantic CMS (SCMS) needs to be considered. Furthermore,
concepts for reasoning and knowledge extraction need to be integrated and the controlling
of their processing must be handled.
The authors participate in the Interactive Knowledge Stack (IKS)1 project, that is focused
on building an open and flexible technology platform for SCMS. As a working hypothesis,
when starting the project in 2009, we thought of four layers, shown in Figure 1, that would
be required to have in an SCMS architecture. The top layer called Presentation & Interaction for presenting knowledge to the end user and support a direct interaction with this
knowledge. To extract knowledge from content the second layer called Semantic Lifting
is used. In this layer, a given content is lifted to semantic level by extracting knowledge
from it. To generate new knowledge based on existing knowledge and to represent this
knowledge within an SMCS we need the third layer, called Knowledge Representation
and Reasoning. The last layer persists the new knowledge in the Persistence layer.
Starting with the initial four layer architecture we began our research on requirements.
We also implemented semantic components in an iterative development process. The requirements were gathered by asking industrial CMS vendors for their needs and by doing
research on possible semantic features [CES+ 09]. The requirements were consolidated
and implemented in prototypical semantic components in an agile development process.
The architecture was refined and adapted with each iteration of research and development [CEN+ 10].
In this paper, we present the resulting reference architecture for SCMS that addresses the
1 http://www.iks-project.eu
(July 27, 2011)
136
domain-specific characteristics of such systems. This reference architecture has been implemented in two software development projects that are described as case studies. The
reference architecture can be adapted in two different ways. On the one hand, the architecture can be used to enhance a traditional CMS with semantic features. On the other hand,
the reference architecture provides support for the specification of SCMS from scratch.
The remainder of this paper is structured as follows. At first, an overview about the stateof-the-art in engineering CMS is given in Section 2. In Section 3, traditional CMS architectures and their characteristics are discussed. Section 4 introduces the SCMS reference
architecture that is applied in two case studies described in Section 5. Finally, the presented
results are concluded and an outlook about future work is given in Section 6.
2 Related Work
Recent research provides only few reference models for the development of SCMS. Based
on the technologies and foundations developed for the Semantic Web, a stack architecture,
termed the “Semantic Web Stack” as described in [Ber00], has been designed. This architecture has been extended in [HPPH05] with a rule layer. Both approaches provide a
conceptual architecture by defining layers and assigning appropriate semantic technologies
to these layers. Though, both approaches do not provide architectures that can be operationalized in the sense of development of software. They neither specify functionalities
provided by layers or components nor the interaction between the layers.
In [BKvH02] Sesame is introduced, a generic architecture model that focuses on storing
and querying metadata in RDF and RDF schema. The architecture includes functional
components that can be deployed. As described, this approach is restricted on a specific
technology and thus limited to RDF and RDF schema. It mainly addresses persistence and
data access components, whereas other functionalities like knowledge extraction or user
interaction are not considered sufficiently.
A server architecture for XML document management is proposed in [AMFK+ 00]. The
architecture consists of a set of high-level layers that focus on the processing of textual
documents. The approach is limited to XML-based content and does not provide solutions
for the storing or extraction of knowledge from existing content.
The “Context Broker Architecture” (CoBrA) described in [CFJ03] uses semantic web technologies in an architecture for context-aware information systems. The storage and access
of knowledge and context reasoning is explicitly considered. The architecture is designed
on a high level of abstraction and does not describe the different components and their
interfaces in detail. In addition, dynamic processing of knowledge extraction is not in the
focus of the approach.
Summarizing the state-of-the-art in architecturing of SCMS, existing approaches do not
provide a generic model that is independent from concrete technologies and considers
all aspects of knowledge extraction and management. Another shortcoming of existing
architectures is the lack of user interaction with knowledge providing an actual value of
semantic technologies for the end user.
137
User Interface
Content Access
Content
Administration
Content
Management
Content
Data Model
Content
Repository
Figure 2: CMS Server Architecture
3 Traditional CMS Architecture
CMS architectures are built upon the concept of a 3-tier architecture with client, CMS
server, and database backend. Figure 2 shows the internal server architecture of a CMS.
The main difference of CMSs compared to other information systems is to focus on flexible
content modeling and storage. Content data models as the representation of content and
their persistence need to be highly adaptable to any domain or customer scenario.
A CMS User Interface at the top layer in Figure 2 presents the content and offers editorial
features to create, modify, and manage content within its lifecycle. Access to the content
itself is provided by a Content Access layer. This layer is used by the User Interface to get
access to the content and the content management features of the CMS. Additionally, the
Content Access layer can be used by third party software that may want to integrate the
CMS into other applications.
The core management features are implemented in the Content Management layer. This
layer provides functionalities for the definition of the domain or application specific Content Data Model. Access control and content lifecycle definitions are further typical management features implemented in this layer. The Content Data Model layer is conceptually
placed below the Content Management layer that has the necessary features to manipulate
the model. The Content Data Model is the application specific model based on the underlying Content Repository. The Content Repository defines the fundamental concepts and
persistence mechanisms for any Content Data Model that is defined on top. The Content
Management features are tightly related to the Content Administration layer to administer
the CMS stack.
4 Reference Architecture
In an SCMS the content is stored in a traditional Content Repository and the knowledge
about that content is additionally stored in a Knowledge Repository. Our proposal of an
138
SCMS Semantic Content Management System
Semantic User Interface
Content
Knowledge
User Interface
Semantic User Interaction
Content Access
Knowledge Access
Presentation &
Interaction
Knowledge
Extraction Pipelines
Reasoning
Content
Data Model
Knowledge
Models
Content
Repository
Knowledge
Repository
Knowledge Administration
Content Administration
Content
Management
Semantic Lifting
Knowledge
Representation
and Reasoning
Persistence
Figure 3: SCMS Reference Architecture
SCMS reference architecture, as shown in Figure 3, is designed in such a way that any
existing CMS, that has an architecture similar to the one depicted in Figure 2, can be
extended to become an SCMS. This ensures that a CMS can be semantified without any
major changes to the existing CMS. To create an SCMS out of a CMS the traditional CMS
content column is extended by a second knowledge column for the semantic features in
parallel.
To interact with each other both columns are connected at the level of Content Access and
Knowledge Access. The content of an SCMS is stored in the content column. Traditional
content repositories are best prepared for this task. The content is transported from the
content column to the knowledge column at the Content/Knowledge Access layer. For a
loosely coupled solution, this could be done via some service implementation, e.g. RESTful Web Services [Fie00]. The delivered content can be analyzed by the components of
the knowledge column. The obtained knowledge is stored in the knowledge column. Once
the content column needs additional knowledge regarding some content it can query the
content column. On the other hand the knowledge column can search for new or changed
content in the content column.
In summary, a SCMS is a CMS with the capability of interacting with, extracting, managing, and storing semantic metadata about content. In the following, we will describe
the SCMS reference architecture along our coarse grained four layer concept shown in
139
Figure 1 that consists of Presentation & Interaction, Semantic Lifting, Knowledge Representation & Reasoning and Persistence.
Presentation & Interaction In a traditional CMS, the user is able to edit and consume
content through a user interface. When dealing with knowledge in SCMS we need an additional layer at the user interface level that allows a user to interact with content, called
Semantic User Interaction. For example, a user writes an article and the SCMS recognizes the name of a person in that article. An SCMS includes a reference to an object
representing that person – not only the person’s name. The user can interact with the person object and see, e.g. its birthday. The person is a knowledge object that is part of a
Semantic User Interaction. Access to knowledge is encapsulated through a Knowledge
Access layer similar to the Content Access layer. Whereas the content column of Figure 3
provides access to the content from the User Interface, provides the knowledge column
access to knowledge for Semantic User Interaction. By combining existing features of a
CMS User Interface layer with new feature from the Semantic User Interaction layer, an
SCMS provides a Semantic User Interface layer on top of both.
Semantic Lifting One problem for traditional CMS is the missing ability to extract
knowledge in terms of semantic metadata from the stored content. Therefore, an SCMS
defines a layer for Knowledge Extraction Pipelines that encapsulate algorithms for semantic metadata extraction. Typically, knowledge extraction is a multistage process [FL04]
by applying algorithms known from the research field of Information Extraction and Retrieval. A Knowledge Extraction Pipeline defines that each stage of the pipeline produces
results that are an additional input for the next stage [ELB07]. For example, a typical step
in a Knowledge Extraction Pipeline is to identify all entities in a given content.
Knowledge Representation & Reasoning After lifting content to a semantic level this
extracted information may be used as inputs for reasoning techniques in the Reasoning
layer. Logical reasoning is a well-known artificial intelligence technique that uses semantic relations to retrieve knowledge about the content that was not explicitly known before.
To handle knowledge within the system we use Knowledge (representation) Models that
define the semantic metadata used to express knowledge. These metadata are often defined
along some ontology that specifies so-called concepts and their semantic relations. For example, persons and organizations are concepts and a semantic relation between these concepts may define that persons can be employees of organizations. Using this definition, one
can express that a concrete person is an employee of an organization and this knowledge
may have been extracted from a given content through a Knowledge Extraction Pipeline.
In the same way the content column provides an extra orthogonal layer for Content Administration there is a need for a corresponding construct to administer knowledge. Knowledge Administration leads from the management of Semantic User Interaction templates,
over Knowledge Extraction Pipeline and Reasoning management to the administration of
Knowledge Models and Repositories.
140
Persistence Knowledge is stored in a Knowledge Repository that defines the fundamental data structure for knowledge. State-of-the-art knowledge repositories implement a
triple store where a triple is formed by a subject, a predicate, and an object. Influenced
by the ideas of the semantic web a triple can be used to express any relation between a
subject and an object. To make this a semantic relation one has to define the Knowledge
Models on top of the Knowledge Repository, e.g. to specify the semantic meaning of a
certain predicate.
In the following section, two case studies are introduced that demonstrate the applicability
of our reference architecture in concrete project contexts.
5 Case Studies
The concepts introduced by the SCMS reference architecture are validated through two
distinct implementation projects. The first project is the IKS reference implementation
(IKS-RI) project with the goal to deliver an implementation of the Knowledge column
of the presented SCMS architecture in Section 4. The IKS-RI is validated through an
industrial early adopters program where CMS vendors are invited to integrate the IKS-RI
technology into their existing CMS and by this they create an SCMS. The major part of the
IKS-RI is implemented as its own open source sub-project hosted at the Apache Software
Foundation. This IKS subproject is called Apache Stanbol and was founded to create an
independent open source community around the presented concepts of an SCMS. In this
paper we reflect on the current work in progress status plus the planned features of the
IKS-RI. The IKS-RI 1.0 release is scheduled for the end of 2011 and the Apache Stanbol
community is supposed to push the development further on after 2011.
The second project is a one-year student project that lasted from April 2010 to April 2011.
The project was dedicated to develop an SCMS for the software engineering domain and
is called Information-Driven Software Engineering – abbreviated as ID|SE. The idea is to
create an SCMS that can handle large unstructured software specification documents and
helps to extract the semantic information that is hidden in these documents. Therefore, the
ID|SE platform tries to, e.g., identify pieces of content within a specification that describe
a requirement.
5.1
IKS Reference Implementation
The IKS-RI2 is used as a proof of concept implementation of the SCMS reference architecture. The implementation of IKS-RI and the development of the SCMS reference
architecture is an intertwined process driven by open-source developers for the implementation and researchers for the reference architecture. Distinguished roles and viewpoints
on the problem to support the development of SCMS ensured that all decisions made on
2 http://code.google.com/p/iks-project/
(July 27, 2011)
141
Reference Architecture
IKS Reference Implementation
Semantic User Interface
IKS VIE Widgets
jQuery
User Interface
Semantic User
Interaction
jQuery
UI
IKS VIE
rdf
Query
Backbone.js
Figure 4: Reference Architecture compared to VIE Architecture
the implementation or the conceptual design were critically questioned.
The IKS-RI implements the Knowledge column of the SCMS reference architecture (see
Figure 3). Its implementation is divided into different open-source projects. The IKS-RI
developers, amongst others the authors of this paper, use and support several existing opensource projects for their work. Examples are the Apache Clerezza3 project to manipulate
semantically linked (RDF) data or the Apache OpenNLP4 project for natural language
processing. Additionally, the developers are actively involved in the development of a
newly founded open-source project, called Apache Stanbol5 , to implement the required
knowledge features. Before describing the work done at Apache Stanbol we will first
focus on the Semantic User Interaction layer of the reference architecture.
Today, most CMS are web-based CMS whereas clients are primarily web browsers. Thus,
the User Interface is implemented using modern Java Script browser technology. Along
the traditional User Interface, the new Semantic User Interaction features are also required
to work inside web browsers. The IKS/VIE6 sub-project makes use of a set of Java Script
libraries, namely jQuery, Backbone.js, and rdfQuery and creates new semantic interaction
widgets on top. The VIE architecture is depicted in Figure 4 in comparison to the reference
architecture. VIE provides a framework at the Semantic User Interaction layer for user
interaction widgets that can be used to create user interface components in the Semantic
User Interface layer. Such widgets realize user interaction with content plus knowledge. A
simple scenario is to edit a text and insert not just the literal name of the person but instead
the entity person with the associated knowledge about that person.
The Apache Stanbol architecture is built upon the OSGi [All11] component model implemented by the Apache Felix7 project. The OSGi model supports elegant separation of different components required by the Knowledge column. Each component is based on a Resource Oriented Architecture [Ove07] exposing their interfaces in terms of a REST [Fie00]
API. The aggregation of the components’ interfaces forms an implementation of the Knowledge Access layer in Figure 3. This means that all components of Apache Stanbol address
functionality within this layer and below of the reference architecture. Apache Stanbol
3 http://incubator.apache.org/clerezza/
(July 27, 2011)
(July 27, 2011)
5 http://incubator.apache.org/stanbol/ (July 27, 2011)
6 http://github.com/IKS/VIE (July 27, 2011)
7 http://felix.apache.org/ (July 27, 2011)
4 http://incubator.apache.org/opennlp/
142
Reference Architecture
IKS Reference Implementation
Knowledge Access
Stanbol REST Service API
Knowledge
Extraction Pipelines
Stanbol Enhancer
Knowledge
Repository
Stanbol
Reasoners
Stanbol
ContentHub
Stanbol
EntityHub
Stanbol
Ontology Mgr
Apache
Clerezza
Apache
Solr
Apache
Jena
OSGi System Console
Knowledge
Models
Knowledge Administration
Reasoning
Enhancement Engines
Figure 5: Reference Architecture compared to IKS-RI Architecture
does not address the top layers. The IKS-RI architecture for the SCMS layers starting with
Knowledge Access is depicted in Figure 5.
Apache Stanbol defines a sub-framework, called Stanbol Enhancer, to implement Knowledge Extraction Pipelines. Such a pipeline consists of Enhancement Engines in which
each engine is responsible to automatically extract one piece of information. Subsequent
engines can use information extracted by previously executed engines. At the end of this
process a given content is enhanced with the extracted metadata that become knowledge.
The most important knowledge that can be extracted with Apache Stanbol is the identification of certain types of entities within the content. Supported entities are e.g. persons and locations. The extracted knowledge is represented in a triple-graph structure
provided by Apache Clerezza and stored with the use of the Stanbol ContentHub. The
ContentHub provides access to previously enhanced content and allows to retrieve already
stored knowledge for a given content.
The EntityHub is used to retrieve semantic information about entities available through
accessible Linked Open Data sources. A prominent example of a public Linked Open
Data source is the DBPedia8 archive which provides access to the data available through
Wikipedia9 . The EntityHub is able to search for entities in such data sources, to cache the
information in an Apache Solr database, and to publish this information within Stanbol.
Reasoning about knowledge is prepared through the Stanbol Reasoner component which
is able to integrate existing reasoning engines. This reasoning implementation requires
knowledge that is structured according to an ontology that can be managed by the Stanbol
Ontology Manager. The Ontology Manager provides a management API for ontologies
expressed in the Web Ontology Language (OWL) [W3C09]. Clients can query the Stanbol
8 http://dbpedia.org/
(July 27, 2011)
(July 27, 2011)
9 http://www.wikipedia.org/
143
Reasoner to gain new knowledge that is only implicitly present in the stored knowledge.
The Stanbol Reasoner processes existing knowledge represented in an OWL ontology and
is able to retrieve new knowledge, e.g. by using the semantic relations between entities
defined through the ontology. Ontologies can be stored using Apache Jena which supports
semantic web standards like the used OWL.
The administration of all Apache Stanbol components is done via the OSGi System Console that is provided by the underlying Apache Felix OSGi implementation. Each component defines its own configuration parameters that can be manipulated via the System
Console. For example, the used Linked Open Data source for the EntityHub is configurable through this mechanism.
5.2
Information-Driven Software Engineering
The ID|SE project10 was a one-year student project with the goal to create an SCMS for
the software engineering domain. Imagine a CMS in which one can upload unstructured
requirements and specification documents and the system is able to identify specified requirements, actors, use cases, and components of the system plus the ability to reason
about relations, e.g. between actors and use cases. Such an SCMS would help to structure
the large amount of information written in plain text specifications. The ID|SE project
used the publicly available specification of the German health card system for evaluation.
The specification11 consits of more than 10,000 pages of text written in German including
some tables plus a few informal figures and is available in Word or PDF format.
In this paper, we focus on the ID|SE architecture and leave out the discussion about semantic features of the ID|SE platform. In Figure 6, we compare the reference architecture
with the ID|SE version of an SCMS architecture. The ID|SE project realizes the Content
column in Figure 3 on the basis of OpenCMS12 , an open-source CMS. In consequence,
the ID|SE User Interface is implemented as an OpenCMS module. The focus of the ID|SE
project was on features to analyze a given software specification document. The creation
of sophisticated Semantic User Interaction components was not part of the project. That is
the reason why the ID|SE architecture does not reflect Semantic User Interaction but only
provides a User Interface without extra semantic capabilities.
The ID|SE Service Platform API provides access through web services to the platform
services that are realized by the IE/IR Service Orchestrators. To connect the Knowledge
column with the Content column provided by OpenCMS the ID|SE project implemented
an adapter in OpenCMS that informs the ID|SE platform about newly uploaded content to
trigger the semantic lifting machinery and to transfer the content from OpenCMS to the
ID|SE platform.
The core of the ID|SE platform is a set of semantic lifting components that implement
features from the research field of Information Extraction (IE) and Information Retrieval
10 http://is.uni-paderborn.de/fachgebiete/fg-engels/lehre/ss10/pg-idse/pg-idse.html
11 http://www.gematik.de/
(July 27, 2011)
12 http://www.opencms.org/ (July 27, 2011)
144
(July 27, 2011)
Reference Architecture
ID|SE Implementation
User Interface
ID|SE Presentation-Platform
based on OpenCMS UI
Knowledge Access
ID|SE Service Platform API
Knowledge
Extraction Pipelines
Knowledge
Models
Knowledge Administration
Reasoning
IE/IR Service Orchestrators
IE/IR Services
Information Aggregators
Metadata Model
Knowledge
Repository
Metadata Storage
Figure 6: Reference Architecture compared to ID|SE Architecture
(IR), respectively. This IE/IR services are executed in a pipeline by the IE/IR Service
Orchestrators. In the following, we present a list of the IE/IR services used by the platform
to give an impression of how this Knowledge Extraction Pipeline depicted in Figure 7
works.
• Content Extraction
Features to convert a given content in DOC or PDF format into a format that is
suitable for being further processed by the pipeline.
• Preprocessors
Features to pre-process the content by applying techniques like tokenization of text,
sentence detection, part-of-speech tagging, and word stemming.
• Classifier
A set of algorithms to classify pre-processed content according some classification
schema. It is used for example to classify content as a text containing requirements
or as a use case description.
• Clusterer
Features to partition large sets of documents into clusters of similar documents. This
is used, e.g., to identify similar requirements as they are partitioned into the same
cluster with these techniques.
• Named Entity Recognizer
Features to recognize named entities like actors, components, and use cases within
the content.
145
Content
Extraction
PreClassifier
processors
Clusterer
Named Entity
Recognizer
Figure 7: ID|SE IE/IR Services Pipeline
After processing the documents by the IE/IR Services, the ID|SE platform provides additional features that are aligned to the reasoning capabilities of the reference architecture,
called Information Aggregators. The Information Aggregators layer consists of components that collect semantic metadata that were extracted by the IE/IR Services and generates new metadata on that basis. The ID|SE platform provides three aggregators.
• Entity Cover Calculator
The Entity Cover Calculator computes the ratio between the coverage of named entities between two documents. This feature can be used, e.g., to search for documents
that may be duplicates of each other like duplicate requirements.
• Faceted Data Creator
The Faceted Data Creator generates metadata for a faceted search user interface
component. A faceted search lets users browse through large sets of documents by
selecting different document categories. Each selection of a new category reduces
the number of matching documents and allows an easy way to find a searched document.
• Use-Case Model Recognizer
Searches for use cases and actors in documents and creates UML [OMG10] use-case
diagrams based on this information.
The knowledge extracted by the IE/IR Services and the Information Aggregators is represented using the ID|SE specific Metadata Model. This model defines that each document consists of a number of artifacts in which each artifact is stored as a character
sequence. This character sequence can be annotated with metadata defining that a certain
sub-sequence contains a specific semantic information. Additionally, character sequences
form tokens and sentences. The Metadata Model is implemented using the Java Persistence API (JPA) [Gro06] for storing these data into the Metadata Storage. The ID|SE platform uses the open-source relational database MySQL13 as its Metadata Storage which
implements the Knowledge Repository layer of the reference architecture.
13 http://www.mysql.com/
(July 27, 2011)
146
6 Discussion & Future Work
In this paper, we have motivated the need for a new generation of CMS that improve the
handling of the growing amount of content by using semantic web technologies and IE/IR
concepts. For these systems, termed Semantic Content Management Systems (SCMS)
we proposed a reference architecture that supports the development of such systems by
considering domain-specific requirements like the pipeline processing of knowledge extraction technologies. The reference model is intended to be used in different ways. The
architecture can be adapted to develop an SCMS from scratch or to enhance existing, traditional CMS with semantic functionalities.
The applicability of our approach in these two different ways has been evaluated by the
implementation of the reference architecture in two software development projects. Applying the reference model to concrete implementation projects, the architecture provided
a starting point for integrating semantic web and IE/IR technologies. The proposed structure ensures the consideration of all relevant aspects of semantic content and knowledge
management by appropriate concepts. Hereby, the presented approach attains benefits for
the design and implementation of SCMS.
Addressing the experiences from the performed case studies, we identified three directions
for the further work on our reference architecture. As a first step, the reference architecture
will be evaluated as part of an overall evaluation in the IKS project. The applicability is
measured by the expertises from the project consortium including partners from seven
research institutions and six industrial CMS providers.
Secondly, we will use best practices and feedback given by architects and developers from
our case studies to concretize the different layers and their functionality to fine-granular
components. In addition, conceptual interfaces for the interaction between these components will be defined. Through this, the applicability of the reference architecture in the
technical design will be improved.
Finally, we still see the potential for further improvement in the presentation and interaction layer of our architecture. The investigation of useful concepts for semantic user
interaction and corresponding user interfaces is still part of our current work. The impact of these concepts on an architecture level needs to be analyzed in order to identify
alignments for the reference architecture.
References
[AL99]
Maryam Alavi and Dorothy E. Leidner. Knowledge management systems: issues,
challenges, and benefits. Commun. AIS, 1999.
[All11]
OSGi Alliance. OSGi Service Platform - Core Service Specification Version 4.3,
2011. http://www.osgi.org/Release4/HomePage (July 27, 2011).
[AMFK+ 00] Tim Arnold-Moore, Michael Fuller, Alan Kent, Ron Sacks-Davis, and Neil Sharman.
Architecture of a Content Management Server for XML Document Applications. In
147
1st International Conference on Web Information Systems Engineering (WISE 2000),
2000.
[Ber00]
Tim
Berners-Lee.
Semantic
Web
XML2000,
2000.
http://www.w3.org/2000/Talks/1206-xml2k-tbl/slide10-0.html (July 27, 2011).
[BKvH02]
Jeen Broekstra, Arjohn Kampman, and Frank van Harmelen. Sesame: A Generic
Architecture for Storing and Querying RDF and RDF Schema. In Proceedings of
the first Int’l Semantic Web Conference (ISWC 2002), Lecture Notes in Computer
Science, pages 54–68. Springer, 2002.
[CEN+ 10]
Fabian Christ, Gregor Engels, Benjamin Nagel, Stefan Sauer, Sebastian Germesin,
Enrico Daga, and Ozgur Kilic. IKS Alpha Development. Deliverable, 2010.
http://www.iks-project.eu/resources/iks-alpha-development (July 27, 2011).
[CES+ 09]
Fabian Christ, Gregor Engels, Stefan Sauer, Gokce B. Laleci, Erdem Alpay,
Tuncay Namli, Ali Anil Sinaci, and Fulya Tuncer. Requirements Specification for the Horizontal Industrial Case. Deliverable, 2009. http://www.iksproject.eu/resources/requirements-capture-through-use-cases (July 27, 2011).
[CFJ03]
Harry Chen, Tim Finin, and Anupam Joshi. Semantic Web in a Pervasive ContextAware Architecture. IN ARTIFICIAL INTELLIGENCE IN MOBILE SYSTEM 2003
(AIMS 2003), IN CONJUCTION WITH UBICOMP, pages 33–40, 2003.
[ELB07]
Michael Thomas Egner, Markus Lorch, and Edd Biddle. UIMA GRID: Distributed
Large-scale Text Analysis. In Seventh IEEE International Symposium on Cluster
Computing and the Grid, CCGRID 2007., pages 317–326, 2007.
[Fie00]
Roy Thomas Fielding.
Architectural Styles and the Design of Networkbased Software Architectures.
PhD thesis, University of California, 2000.
http://www.ics.uci.edu/ fielding/pubs/dissertation/top.htm (July 27, 2011).
[FL04]
David Angelo Ferrucci and Adam Lally. Building an example application with the
unstructured information management architecture. IBM Systems Journal, pages 455–
475, 2004.
[Gro06]
EJB 3.0 Expert Group. JSR 220: Enterprise JavaBeans,Version 3.0 Java Persistence
API, 2006. http://jcp.org/aboutJava/communityprocess/final/jsr220/index.html (July
27, 2011).
[HPPH05]
Ian Horrocks, Bijan Parsia, Peter Patel-Schneider, and James Hendler. Semantic Web
Architecture: Stack or Two Towers? In Principles and Practice of Semantic Web
Reasoning, pages 37–41. Springer, 2005.
[LHL01]
Berners Lee, J Hendler, and O Lassila. The Semantic Web. Scientific American, 2001.
[OMG10]
OMG. Unified Modeling Language (OMG UML), Superstructure, V2.3, 2010.
http://www.omg.org/spec/UML/2.3/Superstructure/PDF (July 27, 2011).
[Ove07]
Hagen Overdick. The Resource-Oriented Architecture. IEEE Congress on Services,
pages 340–347, 2007.
[Sar08]
Sunita Sarawagi. Information Extraction. Foundations and Trends in Databases,
pages 261–377, 2008.
[W3C09]
W3C. OWL 2 Web Ontology Language Structural Specification and Functional-Style
Syntax, 2009. http://www.w3.org/TR/2009/REC-owl2-syntax-20091027/ (July 27,
2011).
148
Enriched Service Descriptions Using
Business Process Configurations
Mohammed AbuJarour, Ahmed Awad
Hasso-Plattner-Institut
University of Potsdam, Germany
{firstname.lastname}@hpi.uni-potsdam.de
Abstract: Service descriptions play a crucial role in Service-oriented Computing (SOC),
e.g., for service discovery, service selection, service composition, etc. However, it has
been observed that service providers – the main source of service descriptions – typically release poor service descriptions, i.e., mostly technical-oriented descriptions.
Several approaches have been proposed to tackle this problem by enriching poor service descriptions with additional information from other sources, e.g., communities or
domain experts. In this work, we propose a novel approach to generate additional information about web services based on the configurations of their consuming business
processes. For instance, we can extract annotations and context information for web
services based on the configurations of their consuming business processes. In this paper, we introduce our proposed approach and its architecture based on the open-source
online modeling platform Oryx and our public service registry Depot.
1 Introduction: Service Descriptions
Service-oriented Architecture (SOA) incorporates the main three roles of service-provider,
-consumer, and -registry. Typically, service providers register their offered web services
in one or more service registries that act as brokers between service providers and consumers. Such registries are then used by service consumers to discover web services that
satisfy their (business) needs. Selected web services are then used as part of business
processes (BP) running inside organizations, such as making a health insurance contract,
selling items online, booking flights and hotels for journeys, etc.
Typically, service consumers maintain repositories of business process models (BPM) for
their daily activities, e.g., add new client, offer special promotions, etc. Each business
process is composed of a set of tasks. Some tasks are manual tasks that are performed by
employees, whereas others are service tasks performed through web services. Building a
new BPM incorporates three steps: 1) creating and labeling its tasks 2) determining the
interaction between them, i.e., data and control flow 3) configuring the created model, i.e.,
selecting web services to perform service tasks. The created BPM is then stored in the
consumer’s repository for similar future requests.
Using the aforementioned scenario in practice involves several challenges, e.g., service
discovery, service selection, BP configuration. Service discovery has been one of the main
challenges in Service-oriented Computing (SOC). Several approaches have been proposed
149
to tackle this challenge, e.g., semantic web service discovery [KLKR07]. However, such
approaches assume the existence of rich and correct service descriptions. In practice, this
assumption is not realistic because it has been observed that service providers release poor
service descriptions [SWGM05]. Additionally, the assumption that service consumers’
queries during service discovery can be expressed formally using a query language is not
realistic due to the increasing complexity of business needs and web services [Sun05].
During service discovery, a list of candidate web services are returned to the service consumer as a result for her submitted query. Choosing a particular web service to invoke
from this list of candidates is known as service selection [SS04]. Service selection has
become a complex task due to several factors, such as lack of rich service descriptions,
increasing number of web services, increasing complexity of web services and modern
business needs, etc. However, it has been shown that additional information about web
services, e.g., annotations, helps meet this challenge [BTR10].
Business process configuration is the service selection step for service tasks in BPMs.
In this step, each service task is assigned a web service to execute it. In practice, process engineers configure tasks manually using static lists of services they know already.
To perform this task dynamically, process engineers require sufficient information about
web services to configure their business process models accordingly. Technical-oriented
information only is not sufficient, because it does not fit their backgrounds and knowledge [SBE07]. The “Internet of Services” initiative introduced the concept of “Business
Services” to emphasize this requirement, namely, service descriptions should be understandable by business people, as well [SAP09].
In this work, we introduce a novel approach to use business process configurations to
learn additional information about web services used in these processes. We have already
introduced an approach to discover relationships among web services based on such configurations [AA11]. In particular, we consider extracting annotations for web services
based on business processes that use them.
1.1
Research Context: Lack of Rich Service Descriptions
Services are typically parts of one or more distributed business processes. Identifying tasks
in a BP that should be implemented as web services is a system design decision. In general,
system designers follow a bottom-up or top-down approach. In the bottom-up approach,
existing web services are composed to achieve more complex tasks. This composition is
repeated at higher levels until the required BP is achieved. On the other hand, the topdown approach means decomposing the complex task of the BP in question into sub-tasks
iteratively until simple tasks are reached. Each simple service task is then implemented
as a web service. The best practice is to follow a mix of both approaches to avoid high
costs of the top-down approach and inflexibility of the bottom-up approach [Jos07]. In our
approach, we follow this best practice, where we allow BP designers to get an overview
of existing, relevant web services, and give them the chance to add more tasks that reflect
their requirements.
The fact that web services are typically parts of one or more distributed business processes
150
brings the challenges of service discovery and selection to business processes. Business
processes and web services are usually separated and each of them is investigated individually. In this work, we introduce a novel approach to use business process configurations
to derive additional annotations for the used web services from the service consumers’
perspective (business view) to enrich their technical service descriptions released by their
providers.
Several limitations can be identified in the current, traditional settings used in modern
distributed information systems. In this work, we target the following limitations:
• Lack of rich service descriptions: This limitation has been identified in both industry and research [BTR10, KTSW08]. Semantic web services tackle this problem by
means of ontologies, but such approaches are not so common in practice [Bos08].
• Insufficient criteria for service selection: Service consumers take several factors
into account when they select web service(s) they want to use, e.g., service quality,
provider’s reputation, context in which the service is typically used, etc. Such information is usually provided by third parties, namely, service brokers (registries).
However, context information is typically ignored.
• Complex configuration of business processes: Creating business process models
is labor-intensive due to the increasing complexity of modern business needs. Additionally, the configuration of business process models is error-prone due to the
increasing complexity of web services and because it is usually a manual task .
1.2
Contributions
The contributions of this work are:
1. A collaborative, incremental, and practical approach to enrich web service descriptions with annotations generated from the configurations of business processes that
consume these web services.
2. Smooth configuration of business process models by enabling context-aware service
selection.
3. Enhanced service discovery using both technical and business perspectives of service descriptions that are gathered from service providers and business process configurations, respectively.
The rest of this paper is organized as follows: In Sec. 2, we discuss potential application
domains for our approach. An overview of our approach is presented in Sec. 3. Further
implementation details and application scenarios are given in Sec. 4. Related work is
summarized in Sec. 5. We close this paper with a summary and future work in Sec. 6.
151
2 Application Domains
Sharing information about business processes and web services used to achieve them is not
usually desired by service consumers. This information might be considered one aspect of
their competitive advantage. Nevertheless, our approach can be valuable in several scenarios and application domains, in particular, where there is a high potential for collaboration
and a low potential for competition among service consumers, such as:
1. Government services: The number of governmental web services has been increasing steadily. For instance, the municipalities of the Netherlands have decided to use
a common infrastructure to run their business processes [GWJV+ 09]. Several similar cases have also evolved in other European countries, such as Germany and UK.
We give a detailed use case from this domain in Sec. 4.3.
2. Education: Web services are usually used in education to share information and
offer computationally-intensive tasks. For instance in Life Sciences, web services
are used to provide information about Genes, and extract protein encodings from
natural text 1 . As collaboration between educational institutions is expected instead
of competition, our approach can be applied in this application domain.
3. Online modeling platforms: Due to the increasing popularity of the Software
-as-a-Service trend, several online modeling platforms are available on the
Internet, e.g., Oryx 2 . Such platforms enable model sharing and collaborative modeling among communities, e.g., distributed teams. In such cases, our approach can
be applied inside each community.
4. Quality-based service brokers: Several service brokers (registries) provide qualitybased service discovery and recommendation, e.g., www.asperado.com. Such
brokers gather statistics about web services and measure their quality. These measures are then used to answer consumers’ requests to find relevant web services with
specific quality constraints. In this scenario, our approach can be applied directly
because the involved service consumers agree already to provide information about
web services they use to service brokers.
Due to policy and privacy issues, our approach cannot be applied to some application
domains, such as banking. However, partial information about the used web services can
be provided by service consumers, without leaking information about the internal structure
of their business processes.
1 Several
examples available on: http://www.biocatalogue.org
2 http://oryx-project.org
152
3 Annotations from Business Process Configurations
Several approaches have been proposed to enrich poor service descriptions with additional
information, such as websites of service providers [ANC10], invocation analysis [AN10],
ontologies [SWGM05], etc. In [ANC10] we have introduced an automatic approach to
annotate public web services with additional information that are extracted by means of
crawling from the websites of their providers. However, this recent approach is limited
to the content provided by service providers on their websites. Therefore, we introduced
another approach to generate dynamic tags for data web services based on the analysis of
their invocations [AN10]. This recent approach is not applicable to functional and business
web services. Hence, additional sources and types of information about web services are
still required. In this work, we introduce our approach to use BP configurations to gather
additional information about web services. These three sources of additional information
about web services are integrated together, indexed, and used to enhance service discovery.
This architecture is depicted in Figure 1.
Depot
Service Discovery
Service Recommendation
<?xml v
<ref:
<r
<?xml
v
<gr
gr
<ref:
<gr
XML
WS
WS
WS
A
at nno
ion t
s
Crawling
A
at nno
ion t
s
Web
A
at nno
ion t
s
Index
BP Repository
A
+
+
B
HTML
C
Invocation
Analysis
BP Configurations
+
+
D
Figure 1: Multiple sources of information about web services to enhance service discovery
3.1 Approach Overview
The main goal of our approach is to enrich web service descriptions with annotations
generated from the configuration of business processes that consume these services. To
achieve this goal, we link business processes and web services. We give an overview of
our approach to achieve this linkage in Figure 2. The scenario starts when a business
process designer creates a new BPM. At that point, the designer gives a descriptive name
and summary for the new process (Step 1), e.g., establish a company in Germany. Behind
the scene, a request is sent to a service registry to find relevant web services that have
been used in similar models (Step 2), e.g., establish a company in UK. The returned set of
services is made available to the designer to select from. A by-product of this step is to
accelerate the process designing step (Cf. Sec. 1.1).
However, the process designer might not use all web services suggested by the service
registry in the new model. She introduces new tasks to express their particular business
153
Service
Consumers
1
3
New Business
Process Model
Model Name
WS
WS
5
Build Business
Process Model
Task Label
Anno
tation
2
WS
WS
Configure Business
Process Model
BPM+
Annotations
Anno
tation
4
Find Relevant
Web Services
Find Web Services
for Tasks
Task-WS
Assignment
Service
Providers
Business Process Repository
Web Service Registry
Task1
WS
WS
S
WS
S
WS
WS
WS
W
WS
WS
W
W
WS
WS
Task2
Task3
Provided By
WS
WS
WS
WS
Task4
Figure 2: A high-level overview of our approach to link BPs and web services
needs (Step 3), where each task is given a label to identify it. This label is passed to the
service registry to find potential web services that can perform the required functionality
(Step 4). For each new task in the model, a list of candidate web services is returned to
the process designer. Each web service is associated with a set of annotations that explain
its functionality. These annotations are extracted and generated automatically from the
websites of corresponding service providers [ANC10]. The new BPM is finalized when
each service task is configured by assigning a web service to execute it (Step 5). With
the finalized BPM, two types of information can be generated, namely, task-web service
assignment and annotated BPM.
The task-web service assignment is passed from the modeling framework to the service
registry. Task labels and documentations are then extracted and their assigned web services
are annotated with this extracted information. These labels and documentations are created
by service consumers that represent the application level, i.e., business people. Whereas,
the initial annotations are extracted and generated from the content released by service
providers, i.e., technical people. Additionally, this assignment and the title of the created
BPM are used to determine the context(s) where these web services are typically used.
This information is then used to enable context-aware service selection for similar cases
in future BPMs.
The tasks in the created BPM are automatically annotated with the annotations of the web
services they are bound with. The result is an annotated BPM. These annotations are
crucial for BPM lookup because not only task labels are used to index and find tasks, but
enriched annotations are also used to achieve this goal [Awa07].
This approach to enrich web service descriptions is collaborative and incremental. With
each newly-created and configured business process model, the service registry gains addi-
154
tional annotations (e.g., labels) and information (e.g., context) about the used web services.
These annotations and information are used to provide enhanced service discovery and enable smooth business process configurations due to context-aware service selection in later
business process models. These advantages speed up the process design. Our approach
is beneficial for both service registries and service consumers. Moreover, we believe that
this is a practical approach to build rich semantic descriptions of web services, which can
be considered as an additional step towards semi-automatic ontology construction. With
such enriched service descriptions, more sophisticated approaches for service composition
can be applied [AAMZM04].
It is worth mentioning that our approach does not suffer from a cold start. Available web
services are already enriched with annotations that are collected automatically from the
websites of their providers. The goal of our approach is to use these annotations to smooth
BP configurations and generate additional annotations for them from the configurations of
BPs that use them to reflect the service consumers’ (business) perspective, as well.
3.2 Formal Model
A service registry usually contains a collection of web services. Each web service, W Si ,
has one or more operations. An operation, OPj , is given a name by its developer, e.g.,
GetWeather, and a list of annotations, A (by service registry). An annotation of an
operation, a, can be a short sentence or several paragraphs describing that operation. These
definitions can be expressed formally as follows:
• Web service: W Si = {OP1 , · · · , OPn }; n ≥ 1
• Operation: OPj =< name, A >; A : annotations
• Annotations: A = {a1 , · · · , am }; m ≥ 1, a : annotation
A business process, BPk , has a list of tasks. Each task, Tl , has a label, documentation,
and an optional operation, op, that executes that task if it is a service task. Formally, these
definitions can be expressed as follows:
• Business Proces: BPk = {T1 , · · · , Ts }; s ≥ 1, T : task
• Task: Tl =< label, doc, op >; doc : documentation, op : operation
Configuring a BP can be expressed as a function that takes a BP and assigns an operation
to each service task in that BP. This function is defined formally in Equaion 13 .
Conf : BP
! → {OP1 , · · · , OPy }; ∀Tb ∈ BP :
OPc if Tb service task; OPc ∈ {OP1 , · · · , OPy }
Tb .op =
none otherwise
(1)
The newly-introduced effect of this configuration function is expressed formally in Equation 2. The label and documentation of each configures service task are appended to
3 We
use the dot notation T.op to refer to the operation property of task
155
the annotation set of its assigned operation. These added annotations usually reflect the
consumer’s (business) perspective of these operation. This perspective complements the
providers’ (technical) perspective.
Tx .op.A = Tx .op.A ∪ Tx .label ∪ Tx .doc
(2)
4 Implementation and Application Scenarios
We have implemented a prototype that realizes our approach to enrich service descriptions
using business process configurations. In this section, we give details about the implementation of this prototype that integrates Oryx – an open-source business process modeling
platform and repository – and Depot – a public web service registry. Our prototype is
available at: https://www.hpi.uni-potsdam.de/naumann/sites/bpws/.
The front-end of our prototype is Oryx, whereas Depot represents the backend. An end
user, such as business process designer, interacts with an Oryx front-end to model her
business processes. Behind the scene, Oryx contacts Depot to find relevant web services
for each particular business process and its individual tasks.
4.1
Oryx: A Modeling Platform
Oryx is an open-source, web-based graphical modeling tool and repository for business
process models [DOW08]. Using a standard web browser, users can login to Oryx and
create, modify and share business process models. Oryx supports a plethora of modeling
languages, e.g, BPMN, Petri nets, UML, etc.
Business process modeling languages are defined within Oryx by means of stencil sets. A
stencil set describes the visual appearance of a model element, either node or edge, the set
of properties it has and connectivity rules to other model elements. The user can define
a stencil set extension to adapt a specific stencil set to a certain modeling domain. We
use this feature to populate a process model with the set of web services returned from
the service registry. In Oryx, a task has several properties, such as label, documentation,
type, implementation, status, etc. During the configuration task, the implementation of
each service task should be specified.
4.2
Depot: A Public Service Registry
Depot is a service registry with multiple sources of information about web services, e.g.,
service providers, invocation analysis, service consumers[ANC10]. The information that
Depot gathers is integrated into unified service descriptions that are used to provide enhanced service discovery. Depot uses an automatic approach to collect public web services
156
Figure 3: Example process model of establishing a UK limited company
from the websites of their providers and annotate them [ANC10].
Traditionally, service consumers provide ratings, feedback, and other quality measures
about web services. Although such information is valuable, it requires additional effort and
is provided by service consumers explicitly. The assumption that the majority of service
consumers give this information is not realistic. In our approach, two types of information
are provided by service providers implicitly, namely, task labels and documentations as
annotations for their assigned operations, and titles of BPMs as context for web services.
4.3
Use Case: Establishing a Limited Company in UK
The European Union (EU) has been one of the main service markets. In 2006, the European Parliament and Council introduced the “Directive 2006/123/EC on services in the
internal market” [Eur06], that aims to create a genuine internal market for services. Therefore, the directive seeks to remove legal and administrative barriers in order to make it
easier for businesses to set up in other member states of the EU and to provide services
cross-borders or on a temporary basis. The services offered by the UK Companies House 4
and other private service providers to establish a limited company in UK is an effort to apply this directive in UK.
Figure 3 shows a process model using BPMN for establishing a UK limited company.
The first six activities of the process are services of the UK Companies House. In the
beginning, it has to be checked whether the desired company name is not already in use
and the directors are not disqualified from leading a company. For the remaining steps,
electronic forms are provided in the Portable Document Format (PDF). Last two activities
in the model are web services offered by private service providers, e.g., Buy domain name
can be executed using the whois web service (http://www.webservicex.net/
whois.asmx).
Using our approach, a process engineer can configure her BPM to establish a company
in UK using the existing annotations for web services and their operations used in such a
model. These annotations are generated from the websites of their providers and through
invocation analysis. Although, they might be not rich enough, such annotation can be helpful in some cases. Generating additional annotations for such web services and operations
4 http://companieshouse.gov.uk
157
from this BPM enrich their descriptions. In this particular use case, all used operations
are associated with the context “establish a company in UK”. Additionally, each operation
is annotated with the label and documentation of each task that uses it. For instance, the
whois web service is annotated with “buy domain name”.
4.4
Our Approach in Action: Establishing a Company in Germany
According to the German Federal Ministry of Economics and Technology, establishing a
company in Germany incorporates 9 major steps 5 . For instance, check the company name,
notarize the articles of association and foundation agreement, notify the Office of Business
and Standards, register at the Trade Office (involves check manager’s qualifications), etc.
Some of these steps are similar to the ones involved in establishing a limited company
in UK (Cf. Sec. 4.3), such as check the company name, check qualified managers, rent
office, buy domain name. Figure 4 shows a screenshot of using our prototype to design a
business process for establishing a company in Germany.
B
A
C
Figure 4: A screenshot of our prototype used to model the process of establishing a company in
Germany. A: suggested web services, B: a pre-configured task from A, C: task properties
To create a new model for this process in Oryx, a proper title, such as “Establishing a
company in Germany” is given by process designer. The area labeled with A in Figure 4
shows a list of web services discovered in Depot that have been used is similar contexts and
are relevant to this process. For instance, the whois web services used in the UK example
can be shown in this example despite the fact that there is no high similarity between terms
appearing in “establish a company in Germany” and “whois”. Each of these web services
is already configured and can be simply dragged-and-dropped to the design area. Indeed,
5 http://www.existenzgruender.de/englisch/
158
the task labeled with B is an example of such pre-configured tasks.
Saving this model adds additional information to Depot about the considered web services,
such as the new labels and the current context. This additional information helps Depot
provide better results in future similar cases, such as “Establishing a company in France”.
5 Related Work
The problem of web service discovery is similar to looking for a needle in a haystack [GPST04].
Seeking the right service based on user’s search criteria in which the user may be interested is still one of the main challenges in SOC [HLV07]. Several factors exacerbate this
challenge [Bos08]:
• The increasing number of available web services on the web: With the increasing
popularity of Software-as-a-Service and Cloud Computing, the number of
public web services has been increasing steadily. For instance, Seekda (http:
//webservices.seekda.com) reported more than 28,000 web services;
• Keyword-based search: This factor is two-folded; from one side service providers
typically release poor service descriptions that are not suitable for keywords matching. Additionally, today’s complex business needs make it difficult for consumers
to even choose proper keywords;
• Syntax-based search that ignores semantics of the terms used by service consumers
(query) and service providers (service description).
The lack of rich service descriptions is highlighted by several researchers in the community [KTSW08]. Therefore, researchers have proposed several approaches to gather additional information about web services to handle the problem of poor service descriptions
[BCPS05, DJZB05, HK03, MB05]. In [ANC10] we introduced an approach to extract and
generate annotations for public web services from the content that their providers publish
on their websites. We showed also that additional annotations can be generated for web
services based on the analysis of the their invocations, such as tags [AN10].
Tremendous work has been proposed to meet the challenge of service discovery. This
work can be grouped into four categories:
1. Semantic approaches: These approaches treat the problem of service discovery as
a matchmaking problem [LPC06]. Service consumer’s needs (goals) are captured in
rich semantic expressions. These semantic expressions are then matched against the
semantic descriptions of the considered web services.
2. Information Retrieval approaches: These approach investigate the semantic relationships between the term used by service consumers in their queries and those
between the terms used to advertise the considered web services [MCZ07].
3. Data mining approaches: These approaches transform the service discovery task
into a constraint satisfaction problem, where the submitted query is matched against
159
the collection of web services. Graph traversal and clustering methods are typically
used to solve the constraint satisfaction problem, where the relationships between
the considered web services are used to increase the accuracy of the returned result
list [YSZX07].
4. Linking approaches: These approaches follow an integration approach to perform
web service discovery. Typically, a set of web services that collectively satisfy a
service consumer’s need is presented to the consumer. Service composition is then
used to consume these web services [RS04].
In this work, we target two of the main limitations in existing service discovery approaches, such as the ones mentioned above: (i) The unrealistic assumptions of the availability of complete and correct service descriptions [FKL+ 05], and (ii) the ability of service consumers to express their sophisticated real business needs so that service discovery
can be achieved [Sun05]. In the Adaptive Service Grid (ASG) project 6 , the authors assume that today’s complex business needs can be expressed (reasonably) by service consumers using ontologies to achieve a semantic service discovery. Our approach uses actual
service usages in BPs to extract additional annotations to enrich their poor service descriptions that are used to provide enhanced service discovery, context-aware service selection,
and smooth configuration of business process models.
6 Summary and Future Work
We tackle the problem of lack of rich service descriptions and its effects on the configuration of BPs. Several approaches have been proposed to enrich poor service descriptions
with additional sources of information, such as community annotations, domain experts,
etc. In this paper, we introduced a novel approach to enrich poor service descriptions with
annotations extracted from the configurations of BPs that consume them. Two types of information can be extracted from such configurations, namely, annotations from task-web
service assignment and context(s) of web services. the extracted annotations help enhance
service discovery in future cases. Context information enables context-ware service selection, and smooth BP configuration.
We used Oryx and Depot to implement a prototype that realizes our approach. Oryx represents the front-end and Depot represents the back-end of our prototype. We showed how
our approach can help create BPMs in several domains and introduced a detailed example
to to model a process for establishing a company in Germany.
Our approach is collaborative, incremental, and practical. The more we can ship back from
process models to service registries, the better service discovery we can provide. In future,
we will consider not only task-service assignments but also global behavioral relationships
imposed on services by process models via their tasks. We believe that this is of crucial
importance to achieve service composition in a practical way.
6 www.asg-platform.org
160
References
[AA11]
Mohammed AbuJarour and Ahmed Awad. Discovering Linkage Patterns among Web
Services using Business Process Knowledge. In Proceeding of the International Conference on Services Computing, Washington, DC, USA, 2011. IEEE Computer Society.
[AAMZM04] I. Budak Arpinar, Boanerges Aleman-Meza, Ruoyan Zhang, and Angela Maduko.
Ontology-Driven Web Services Composition Platform. In Proceedings of the IEEE
International Conference on E-Commerce Technology, pages 146–152, Washington,
DC, USA, 2004. IEEE Computer Society.
[AN10]
Mohammed AbuJarour and Felix Naumann. Dynamic Tags For Dynamic Data Web
Services. In Proceedings of WEWST’10, Ayia Napa, Cyprus, 2010. ACM.
[ANC10]
Mohammed AbuJarour, Felix Naumann, and Mircea Craculeac. Collecting, Annotating, and Classifying Public Web Services . In OTM 2010 Conferences, Crete, Greece,
2010. Springer.
[Awa07]
Ahmed Awad. BPMN-Q: A Language to Query Business Processes. In Proceedings
of EMISA’07, volume 119, pages 115–128, St. Goar, Germany, 2007.
[BCPS05]
Marcello Bruno, Gerardo Canfora, Massimiliano Di Penta, and Rita Scognamiglio.
An Approach to Support Web Service Classification and Annotation. In Proceedings
of EEE’05, pages 138–143, Washington, DC, USA, 2005. IEEE Computer Society.
[Bos08]
Aishwarya Bose. Effective Web Service Discovery using a Combination of a Semantic Model and a Data Mining Technique. Master’s thesis, Queensland University of
Technology, Queensland, Australia, 2008.
[BTR10]
Stephan Buchwald, Julian Tiedeken, and Manfred Reichert. Anforderungen an ein
Metamodell für SOA-Repositories. In Christian Gierds and Jan Sürmeli, editors,
ZEUS, volume 563, pages 17–24. CEUR-WS.org, 2010.
[DJZB05]
Zhang Duo, Li Juan-Zi, and Xu Bin. Web Service Annotation Using Ontology Mapping. In Proceedings of SOSE ’05, pages 243–250, Washington, DC, USA, 2005.
IEEE Computer Society.
[DOW08]
Gero Decker, Hagen Overdick, and Mathias Weske. Oryx - An Open Modeling
Platform for the BPM Community. In BPM, volume 5240 of LNCS, pages 382–385.
Springer, 2008.
[Eur06]
European Parliament and Council. Directive 2006/123/EC of the European Parliament and of the Council of 12 December 2006 on services in the internal market.
ISSN 1725-2555, December 2006.
[FKL+ 05]
Dieter Fensel, Uwe Keller, Holger Lausen, Axel Polleres, and Ioans Toma. WWW
or what is wrong with web service discovery? In Proceedings of the W3C Workshop
on Frameworks for Semantics in Web Services, 2005.
[GPST04]
John Garofalakis, Yannis Panagis, Evangelos Sakkopoulos, and Athanasios Tsakalidis. Web Service Discovery Mechanisms: Looking for a Needle in a Haystack? In
International Workshop on Web Engineering, 2004.
[GWJV+ 09] Florian Gottschalk, Teun A. Wagemakers, Monique H. Jansen-Vullers, Wil M. Aalst,
and Marcello Rosa. Configurable Process Models: Experiences from a Municipality
Case Study. In Proceedings of CAiSE’09, pages 486–500, Berlin, Heidelberg, 2009.
Springer-Verlag.
161
[HK03]
Andreas Heß and Nicholas Kushmerick. Learning to Attach Semantic Metadata to
Web Services. In International Semantic Web Conference, pages 258–273. Springer,
2003.
[HLV07]
Stephan Hagemann, Carolin Letz, and Gottfried Vossen. Web Service Discovery Reality Check 2.0. In Proceedings of NWESP ’07, pages 113–118, Washington, DC,
USA, 2007. IEEE Computer Society.
[Jos07]
Nicolai M. Josuttis. SOA in Practice: The Art of Distributed System Design. O’Reilly,
2007.
[KLKR07]
Ulrich Küster, Holger Lausen, and Birgitta König-Ries. Evaluation of Semantic Service Discovery- A Survey and Directions for Future Research. In Postproceedings of
WEWST’07, 2007.
[KTSW08]
Dominik Kuropka, Peter Tröger, Steffen Staab, and Matthias Weske. Semantic Service Provisioning. Springer, Germany, 2008.
[LPC06]
Chuanchang Liu, Yong Peng, and Junliang Chen. Web Services Description
Ontology-Based Service Discovery Model. In Proceedings of WI ’06, pages 633–
636, Washington, DC, USA, 2006. IEEE Computer Society.
[MB05]
Brahim Medjahed and Athman Bouguettaya. A Dynamic Foundational Architecture
for Semantic Web Services. Distrib. Parallel Databases, 17(2):179–206, 2005.
[MCZ07]
Jiangang Ma, Jinli Cao, and Yanchun Zhang. A Probabilistic Semantic Approach
for Discovering Web Services. In Proceedings of WWW ’07, pages 1221–1222, New
York, NY, USA, 2007. ACM.
[RS04]
Jinghai Rao and Xiaomeng Su. A Survey of Automated Web Service Composition
Methods. In Proceedings of SWSWPC’04, pages 43–54, 2004.
[SAP09]
SAP Research.
Unified Service Description Language.
internet-of-services.com, 2009.
[SBE07]
Sebastian Stein, Katja Barchewitz, and Marwane El Kharbili. Enabling Business
Experts to Discover Web Services for Business Process Automation. In Proceedings
of WEWST, Halle (Saale), Germany, 2007.
[SS04]
Raghuram M. Sreenath and Munindar P. Singh. Agent-based Service Selection. J.
Web Sem., 1(3):261–279, 2004.
[Sun05]
Sun Microsystems. Effective SOA Deployment using an SOA Registry Repository,
A Practical Guide, 2005. White paper.
[SWGM05]
Marta Sabou, Chris Wroe, Carole Goble, and Gilad Mishne. Learning Domain Ontologies for Web Service Descriptions: An Experiment in Bioinformatics. In Proceedings of WWW’05, 2005.
[YSZX07]
Jianjun Yu, Hao Su, Gang Zhou, and Ke Xu. SNet: skip graph based semantic web
services discovery. In Proceedings of SAC ’07, pages 1393–1397, New York, NY,
USA, 2007. ACM.
162
http://www.
Developing a Cloud Provider Selection Model
Jonas Repschläger1, Stefan Wind2, Rüdiger Zarnekow1, Klaus Turowski2
1
Chair of Information and Communication Management
Technical University Berlin, Sekr. H 93, Straße des 17. Juni 135, 10623 Berlin
jonas.repschlaeger@ikm.tu-berlin.de, ruediger.zarnekow@ikm.tu-berlin.de
2
Chair of Business Information Systems I
Otto-von-Guericke University Magdeburg, Universitätsplatz 2, 39106 Magdeburg
stefan.wind@ovgu.de, klaus.turowski@ovgu.de
Abstract. Recently, a growing development and use of Cloud computing
services has been observed. Especially modeling cross-organizational
cooperation and the respective provider comparison are gaining importance.
Despite initial positive results, it is challenging in theory and practice to find an
appropriate provider matching the individual requirements. Moreover, the
comparison process is complicated by a number of new entrants as well as
offers of non-transparent services, which sometimes differ significantly.
Due to the lack of adequate possibilities to compare and classify Cloud
providers we are presenting in this paper a provider independent classification
model for Infrastructure as a Service (IaaS). For this purpose, the target
dimensions for Cloud Computing from a customer perspective were defined,
based on expert interviews, an international literature review and a Cloud
provider market (IaaS and hosting) analysis.
1
Motivation
For several years Cloud Computing has been influencing the IT landscape and has
been an important economic factor. Gartner predicts a strong growth of Cloud
services with a worldwide revenue of 146.8 billion US dollars in 2014 [Pr09]. In this
context, Cloud Computing has become a fast growing and non-transparent market
with many small and large providers, each of them having their specific service model
[RZ11a], [HK10], [VJ10]. That makes it difficult to compare the providers as such
and their service offerings. In the majority of cases the service portfolios are
heterogeneous and combined with complex pricing models and service features.
Furthermore, the fact that interoperability between providers hasnt been achieved
makes a provider selection often irreversible or requires much effort [HK10],
[RZ11b]. This difficulty, known as provider lock-in, is discussed extensively and is
an important topic in many companies and international research activity e.g. Open
Grid Forum (OGF) [CH09], [Ar09]. Consequently, the customer is confronted with
the situation to select an appropriate provider to realize his specific requirements
163
mostly based on diffuse classification criteria. Due to the lack of adequate
possibilities of comparing a Cloud provider, especially on the infrastructure level, this
paper focuses on developing a classification model for IaaS providers. In this context
the following research question need to be answered:
What are appropriate classification characteristics for Cloud providers and what
should an Infrastructure as a Service (IaaS) classification model look like?
To reduce entry barriers and support the migration into the Cloud this paper starts
with examining Cloud Computing characteristics and the IaaS providers market.
Based on a literature review and interviews conducted with experts we have derived
six customer target dimensions valid for Cloud Computing (see chapter 4). These
target dimensions serve as strategic objectives regarding Cloud Computing and
provide a structure for Cloud characteristics in the first place. Next we gathered all
requirements available and appropriate classification criteria for Cloud providers and
reviewed them in cooperation with the experts. On this basis we have developed a
classification model by assigning relevant provider and service requirements to the
target dimensions using a four level hierarchy (see chapter 5). We defined 17
classification criteria on the 2nd level and 51 criteria on the 3rd level.
2
Cloud Computing introduction
With Cloud Computing a paradigm shift to standardization and service orientation in
the information and communication industry emerges and marks the industrialization
of IT [Le10]. Cloud Computing allows companies to rent IT services on demand to
support their business processes. As a new part of IT sourcing, the effort of operation
and maintenance is completely managed by the provider. The customer only rents the
service in a pay-per-use manner like a commodity similar to the energy or water
market.
Considering the extensive usage of on demand services, mobile applications and
interactive elements, transactions and workloads will rise significantly and require
scalable IT structures [La10]. This growing acceptance will result in an increasing
demand for IT resources. At this point the Cloud Computing model is essential. It
makes on demand network access to a shared pool of configurable computing
resources possible, that can be rapidly provisioned and released with minimal
management effort or service provider interaction [GM09]. The resources (e.g.
networks, servers, storage, applications and services) are offered in a scalable way via
the Internet without the need for any long-term capital expenditures and specific IT
knowledge on the customer side. It is possible to obtain complete software
applications or the underlying IT infrastructure in the form of virtual machine images.
Basically Cloud Computing is composed of the characteristics above described and
consists of three levels, software as a service (SaaS), platform as a service (PaaS) and
infrastructure as a service (IaaS) [GM09].
164
On the infrastructure level customers have the possibility to obtain on demand
resources from a Cloud provider with no need to operate the required IT
infrastructure. On the one hand, the customer can extend his existing IT infrastructure
by renting additional capacity from the Cloud to compensate load peaks, preventing
the customer from having much capacity available. On the other hand, the complete
infrastructure and respective services can be obtained from the Cloud. This decreases
the IT maintenance effort and creates value especially for small businesses with
fluctuating demand, e.g. due to seasonal activities. Alongside the benefits of Cloud
Computing, several challenges emerge on the management level regarding different
technological components and modules, integration of heterogeneous interfaces,
usage dependent provisioning and billing of resources as well as ensuring data
privacy and data security [IH10], [Bi10], [Cl09].
3
Research Methodology
Cloud Computing is characterized by various factors and a common definition of this
term doesnt exist [We09], [Wa08], [Va09]. Thus Cloud Computing was examined
from different perspectives (technological, business issues, applications and general
aspects) [YT09]. According to this, we attempted to develop an IaaS classification
model which would be valid for all research fields without limitation.
Prior to the literature review five experts were interviewed on common objectives in
Cloud Computing. As a result four target dimensions for the Cloud could be derived.
The expert interviews were conducted with five experts from four companies, all
holding different positions within their companies. Care was taken that those
respondents were representative of all perspectives (provider, customer and
mediator/consultant) being important for the selection process (see Table 1). The
interviews with the experts were structured and conducted referring to Glaeser and
Laudel [GL10].
Table 1. Type of experts interviewed
(Expert from)
Company type
Position within
company
Company data
170.000 employees
Global IT service offerings
IT service provider
Software provider
Software provider
Consulting company
Customer / Client
10-15% revenue based on Cloud
Computing
Innovative solutions in IaaS
SME software company
11 employees
Development of standardized
components for web-based
services
SME software company
11 employees
Development of standardized
components for web-based
services
International consulting company
500 consultants worldwide
Cloud Computing as one
consultancy topic
Automotive sector
ca. 95.000 employees
Director of IaaS
division
CIO
Software
architect
Partner
Divisional
director IT
165
Cloud experience
Deep understanding of
IaaS services and
infrastructure
Expert know-how in SaaS
and general Cloud
approaches
Expert knowledge in IaaS
and PaaS especially in the
implementation
Current consulting focus;
Cloud market appreciation
Experience in selecting,
implementing and
operating IaaS
Next, in order to describe, synthesize, evaluate and integrate the results of existing
scientific work in comparison of Cloud providers and distinguishing criteria of Cloud
offerings, we conducted a systematic literature review following the approach of
Webster and Watson [WW02]. This research method ensures that an extensive
number of relevant papers considered. During the literature review we attempted to
match each criterion gathered to a representative target dimension. The necessity
emerged to define two more target dimensions in order to allocate all relevant criteria.
Afterwards the additional dimensions were discussed and evaluated with the experts
as well. The outcomes were six target dimensions, each including a group of relevant
classification criteria.
The first step in the literature selection process was conducted to identify a
comprehensive list of literature sources. We started off by taking the top journals
based on the VHB-JOURQUAL2 [SH09] and Saunderss journal ranking [Sa11b]. To
complete the analysis, publications of renowned national and international
organizations and associations (e.g. Bitkom) were included. Table 2 lists all literature
sources that were examined to identify relevant papers.
Table 2. Journals and conferences investigated for the literature review
Publication
type
Publisher
Journals
ACMSIG, CACM, CAIS, CompDcsn, DATABASE, DSI, DSS,
EJIS, I&M, I&O, IBMSJ, IEEEComp, IEEESw, IEEEIC,
IEEETC, IEEETKDE, IEEETrans, IEEETSE, IEEETSMC, IJEC,
IJHCS, InfoS ys, ISF, ISJ, ISM, ISR, IT&M, IT&P, JACM, JAIS,
JCIS, JComp, JCSS, JIM, JITTA, JMIS, JSIS, KBS, MISQ,
MS, SMR, WIRT
Conferences
Associations,
Organizations,
Companies
AMCIS, ECIS, ICIS, HICSS, IEEE Conferences, LNI, LNCS,
MKWI, PACIS, WI
Cloud Security Alliance (CS A), EuroCloud, Bi tkom,
Bundesamt für Sicherheit in der Inform ationstechnik (BSI),
Securing Europe's Information Society (ENISA), Center for
Experimental Researchin Computer Systems (CE RCS),
Fraunhofer SIT, Distributed Managem ent Task Force
(DMTF), The European Telecommunications Standards
Institute (ETSI), National Institute of Standards and
Technology (NIST), Open Grid Forum (OGF), Object
Managem ent Group (OMG), Open Cloud Consortium (OCC),
Organization
for
the
Advancement
of
Structured
Information Standards (OASIS), Storage Networking
Industry Association (SNIA), The Open Group, TM Forum,
SaaS EcoSystem, OpenCloudMani festo, Experton Group, TSystems
In a subsequent step, we chose topic related papers from the selected literature
sources. An initial list of papers was generated by using key words such as Cloud
Provider, Cloud Vendor, Cloud Characteristics, Cloud Classification, Cloud
Selection, Cloud Taxonomy, IaaS, Infrastructure as a Service, Platform as a
Service, Software as a Service, PaaS and SaaS to search for titles, abstracts
and keywords. We only scanned the directories of the journals and conference
proceedings manually if no electronic search was possible. Furthermore, we expanded
our scientific foundation by reviewing the citations in the papers identified in the first
cycle of literature exploration to determine previous papers that should be considered
for an analysis in a subsequent cycle of literature exploration.
166
We identified 55 papers all dealing with the comparison of Cloud providers or at least
containing related keywords. In order to identify the final set of publications we
subjected these papers to a detailed (content-related) review. Therefore, we manually
reviewed the papers of the initial list and selected only those papers which primarily
dealt with the comparison of Cloud providers. Thus, 38 articles were selected which
dealt primarily with the classification or selection of Cloud providers and
distinguishing criteria of Cloud offerings. It is surprising that almost the entire set of
finally selected papers consists of conference papers and there is just a small amount
of high-quality journal papers available. This probably shows that there is a lack of
research regarding the classification of Cloud services and distinguishing criteria of
Cloud offerings.
Complementary to the literature review the provider market of IaaS and hosting were
investigated. This analysis was based on an extensive internet research where the
websites of relevant companies were examined regarding their pricing model, IaaS
service offering, company data and customer segment. By means of market studies
and business publications on the Cloud market we detected over 60 relevant providers
in the IaaS and hosting business [VJ10], [LC10], [HF09]. Based on this analysis we
compiled a feature catalog for IaaS providers.
In order to develop a detailed proposal for an IaaS classification model the target
dimensions including the related criteria from the literature review were matched with
the features of the provider catalog supported by the experts. Thereby, a designoriented approach was used.
4
Target Dimensions in Cloud Computing
In this contribution we define six target dimensions - such as cost savings or
increasing flexibility - to group and structure the Cloud characteristics. Each target
dimension represents a general objective which the customer pursues and which
characterizes his Cloud or IT strategy. Four target dimensions (costs, IT security &
privacy, scope & performance, reliability & trustworthiness) were defined together
with the experts prior to our analysis. Through our literature review and market
research, we validated these four dimensions and simultaneously discovered two
additional dimensions (flexibility and service & Cloud management), which were
evaluated subsequently by the experts as well. Table 3 shows the relevant sources
assigned to the six target dimensions. In this context we discovered that practitioners
mainly deal with questions about security, reliability and manageability of Cloud
Computing. Largely unnoticed, however, remain so far the performance as well as the
cost / price models. Science is exploring the topic covering all target dimensions and
attaches significantly more importance to the performance dimension.
167
Table 3. Results of the literature review
IT Security &
Compliance
Service & Cloud
Management
Source
Reliability &
Trusttworthines
Scope &
Performance
Costs
Flexibility
Target
Dimension
30
29
14
27
18
Academic Publications
Günther et al. (2001)
Hilley (2009)
Hoefer and
Karagiannis (2010)
Li et al. (2010)
Prodan and Ostermann
(2009)
Annecy (2010)
Vaquero et al. (2009)
Peng et al. (2009)
Weinhardt et al. (2009)
Hay et al. (2011)
Martens et al. (2010)
(2011)
Christmann et al.
(2010)
Tsvihun et al. (2010)
Armbrust et al. (2010)
Iyer und Henderson
(2010)
Anandasivam and
Premm (2009)
Lehmann et al. (2009)
Rimal et al. (2009)
Schwarz et al. (2009)
Talukder et al. (2010)
Koehler et al. (2010)
(2010)
Saya et al. (2010)
Narasimhan et al.
(2011)
Russell et al. (2010)
Popularity
36
41
35
Industry Publications
BITKOM (2010)
BSI (2010)
EuroCloud (2010)
ENISA (2009)
CSA (2009)
SaaS EcoSystem
(2011)
DMTF (2009) (2010)
OpenCloudManifesto
(2009)
T-Systems (2008)
Experton Group (2010)
The Open Group
(2009)
Popularity
Low priority
21
6
1
Average priority
168
24
High priority
4.1
Target Dimension: Flexibility
An associated advantage of Cloud Computing, identified in science and industrial
practice, is the gain in flexibility compared to traditional solutions [We09].
Resources, for example, can be allocated and de-allocated as required, whereas
requirements can sometimes vary greatly. The provisioning time is shorter compared
to traditional outsourcing such as Application System Provider (ASP) and with a very
small overall commitment period to the vendor [RZ11a]. Also other aspects such as
standardization (e.g. APIs), the traceability of data, the short-term contracts or a
demand-driven and scalable resource recovery have to be considered regarding an
appropriate supplier selection.
4.2
Target Dimension: Costs
The decision to choose Cloud Computing and a particular provider is often guided by
monetary considerations [Hi09] and linked with the slogan "pay-as-you-use".
Customers who decide to use Cloud services benefit mostly from small capital
commitment, low acquisition costs for required servers, licenses or necessary
hardware space and the reduced complexity of IT operations. Despite similar services
on the IaaS level the pricing and billing models often differentiate between each
provider [RZ11a].
4.3
Target Dimension: Scope & Performance
With this target dimension the scope of services and the performance of a Cloud
provider are described. To select the Cloud provider which best meets the
requirements, knowledge about their service and performance is of crucial importance
[Ar09]. Here it is essential to consider features regarding performance (latency, or
transaction speed), capacity limits (e.g. maximum number of accounts or storage
space), service complexity (how many functions are available) and degree of
customization (how far the service can be adapted).
4.4
Target Dimension: IT Security & Privacy
The decision on selecting a provider in the Cloud is very often influenced by
company requirements in the areas of security, compliance and privacy [CH09],
[Ts10], [Ja08], [Cl09]. Companies have to be sure that their data and applications,
even operated in the Cloud, meet the required compliance guidelines of both and are
adequately protected against unauthorized access. Here, the decision criteria are rather
referring to the infrastructure of the provider itself than to the provided service.
169
4.5
Target Dimension: Reliability and Trustworthiness
This target dimension describes how certain the customer can be that the service from
the Cloud has the guaranteed availability [Ja08]. It is important to know what
commitment the provider makes, mostly in form of Service Level Agreements
(SLAs). Moreover, the reliability with which these commitments are kept is of great
importance. In contrast to the commitment, the trustworthiness describes the
provider's infrastructural features, which may be evidence of a high reliability. These
include disaster recovery, redundant sites or certifications.
4.6
Target Dimension: Service & Cloud Management
The Service & Cloud Management includes features of the provider that are
substantial for convenient Cloud service operations. These include the offered support
and functions for controlling and monitoring as well as the individualization of the
web interface [YT09]. The manageability (usability) of services, especially in a
distributed IT architecture, and the Cloud governance, dealing with requirements and
responsibilities by the customer, are essential features of this target dimension.
5
Classification Model for Infrastructure as a Service (IaaS)
The aim of this paper is to develop a classification model for supporting a Cloud
service selection on the infrastructure level. The model may help companies in their
selection process and creates greater transparency in the Cloud market. For that
purpose, different target dimensions were elaborated (chapter 4). These dimensions
cover the Cloud Computing in its entirety and arent limited to one level (SaaS, PaaS,
IaaS). Next, these dimensions have to be broken down into selection criteria that are
measureable and comparable. Based on the fact that all three Cloud levels are
targeting different customer needs, it wasnt possible to define selection criteria valid
for all levels at once. At this point, we limited our examination to the infrastructure
level (IaaS) because in our opinion the services are particularly convenient due to a
certain degree of consistency. With the purpose of developing a classification model
we summarized and mapped similar characteristics and requirements regarding their
target dimension into three hierarchical levels (see Figure 1). Thereby abstract and
operational selection criteria were identified. The abstract selection criteria are used
for further structuring and differentiation purposes. On the third level of the
classification scheme, criteria have been operationalized, so that they can be weighted
and compared (e.g. pricing options, delivery time). The level below finally defines
figures and measurable requirements (key performance indicator, KPI). Each
operative criterion (3. level) consists of various level 4 requirements. Furthermore,
requirements can be divided into provider requirements and service requirements (see
Figure 1). Provider requirements describe the features of the Cloud provider
independently from any service, e.g. existing certifications, IT infrastructure
characteristics or key figures of the company. Service requirements in contrast deal
170
with characteristics directly referring to the usage of a service, e.g. service
availability, scalability or interface features.
Classification scheme example: The target dimension Flexibility (1st level) consists
of service dynamics among other things, an abstract classification criterion (2 nd
level) which is characterized through the provisioning time (3rd level). The
provisioning time is measured among other things by the required time to start up
an instance (4th level; KPI). For instance, if the deployment time is less than five
minutes, the provisioning time is rated as low, assuming the other requirements will
be rated similarly.
1st level
general objectives of
Cloud Computing
target dimensions
2nd level
specific classification of
Infrastructure as a Service
abstract classification
criteria
3rd level
operative
classification criteria
service requirements
4th level
requirements / KPI
provider requirements
Fig. 1: Classification scheme
The result of this paper is a classification model with six target dimensions, 17
abstract selection criteria (2. level) and 51 operative selection criteria (3. level). This
model is illustrated in Figure 2. For reasons of clarity the requirements on level four
are not shown.
171
Costs
Flexibility
interoperability & portability
standardisation
price model
data portability
network
configuration
Scope & Performance
technology
price class
load balancing
virtualization
price resilience
multi-tenancy
network access
price options
automatisation degree
software
price transparency
resource
scalablity
changes / updates
system
management
payment options
payment method
service dynamics
service charging
provisioning time
instance
customizing
contract length
charging
granularity
scalability
charging type
set-up time
IT Security & Privacy
datacenter security
hardware security
performance
computing time
reliability
incident und service management
redundancy
(network)
trustworthiness
support
contact and
consulting services
service operations
data protection
provider reporting
auditing
system
management
datacenter location
service
transparency
provider
profile
Controlling
network security
service level agreements
availability
usability
connection
opportunities
liability and
penalties
customizing
options
abstract selection criteria
reporting
and
monitoring
web portal
connection security
target dimension
instance capacity
connection
bandwidth
Service & Cloud
Management
redundancy
(datacenter)
data protection &
compliance
add-on services
Reliability and Trustworthiness
disaster recovery
software security
instance type
storage service
operative selection criteria
Fig 2: Classification model for IaaS (without KPIs)
Next, the selection criteria and target dimensions will be briefly explained. The
classical target dimension is the flexibility, which describes the ability to respond
quickly to changing capacity requirements. Here, the topics interoperability &
portability, automation and service dynamics are particularly noteworthy. But there
are far more important aspects in supplier selection. These include, for example, price
stability (duration of the agreed price) and the granularity of allocation (e.g. 1 MB,
100 MB or 1 GB steps). Between IaaS providers in particular, the pricing (price
range, pricing options, price transparency, payment options etc.) and billing model
differentiate clearly. The performance is regarded as a relevant criterion for selection
by the experts, although it was marginally discussed so far in practice (see Table 3).
The efficiency and the scope of services can be divided into performance, technology
and software. Criteria like the max. CPU, RAM, hard drive, transfer volumes and
transfer speed belong to the topic of performance. Features such as the type of
virtualization, the load balancing and so on will be subsumed under the term
172
technology. This target dimension is completed by software characteristics, such as
the selection of different operating systems. The target dimension of IT security is
attracting particular attention in practice and refers to the data center or network.
There are requirements for privacy (encryption of data) and compliance (e.g. location
of data center) as well. Using external services from the Cloud, trustworthiness and
reliability are very important. In this field they can be grouped into commitment
(which efforts are guaranteed by the provider, service SLAs), reliability (the
probability that service pledges can be met) and trustworthiness (e.g. performance
transparency, market experience). The target dimension Cloud & Service
Management can be differentiated according to incident and service management
(support and customer service), service operation (e.g. monitoring of services and
volume control via APIs) and web portal (usability and adaptability of the surface).
6
Conclusion & Future Work
Based on our literature and provider review, complemented by interviews with
specialists in the respective field, we introduced in this contribution a classification
model enabling a provider selection on the IaaS level. Our view is that via the model,
companies are able to support their decision process and simplify the provider
comparison. In this way, heterogeneous IaaS offerings in the Cloud can be compared
a lot easier and the time of the selection process can be accelerated as well. It is
conceivable that via middleware or standardized interfaces different IaaS offerings
can be automatically selected and offered based on the customer preferences (target
dimensions). Cloud actors like mediators or aggregators are becoming more
important. The consequence is a shift from a subjective provider assessment to a
mostly fact -based performance selection where the realization of service
requirements is gaining importance. There is no need for customers to know from
which provider they obtain the service as long as the requirements are met.
As further research approach, the classification model will be evaluated and refined.
Therefore IT managers will be surveyed and various case studies will be conducted to
ensure applicability of the presented classification model.
7
References
[AP09] Anandasivam, A., and Premm, M., Bid price control and dynamic pricing in
Clouds, in Proceedings of the European Conference on Information Systems,
Verona, Italy, 2009.
[An10] Annecy, M.: XaaS Check 2010 Status Quo und Trends im Cloud Computing.
S.A.R.L. Martin, Forschungsgruppe Service-oriented Computing,Technische
Universität Darmstadt und IT Research, Sauerlach. 2010.
[Ar09] Armbrust, M., Fox, A., Griffith, R., Joseph, A.D., Katz, R.H., Konwinski, A., Lee,
G., Patterson, D.A., Rabkin, A., Stoica, I. and Zaharia, M., Above the Clouds: A
173
[Bi10]
[Bs10]
[CH09]
[Ch10]
[Cl09]
[Di09]
[Di10]
[Eu10]
[GL10]
[GM09]
[Gü01]
[Hi09]
[HK10]
[Ha11]
[HF09]
[IH10]
[Ja08]
[Ko10a]
Berkeley View of Cloud Computing, UC Berkeley Reliable Adaptive Distributed
Systems Laboratory, 2009.
BITKOM: Cloud Computing Was Entscheider wissen müssen. BITKOM Leitfaden,
2010.
BSI: BSI-Mindestsicherheitsanforderungen an Cloud-Computing-Anbieter.
Bundesamt für Sicherheit in der Informationstechnik, Bonn, September 2010.
Catteddu, D.; Hogben, G.: Cloud Computing - Benefits, risks and recommendations
for information security. European Network and Information Security Agency
(ENISA), 2009.
Christmann,S., Hilpert, H., Thöne, M., and Hagenhoff, S., Datensicherheit und
Datenschutz im Cloud Computing Risiken und Kriterien zur Anbieterauswahl,
HMD- Praxis der Wirtschaftsinformatik, Heft 275, Oktober 2010, S.62-70.
Cloud Security Alliance: Security Guidance for Critical Areas of Focus in Cloud
Computing V2.1. Cloud Security Alliance (CSA), Dezember 2009.
Distributed Management Task Force (DMTF) Interoperable Clouds, Whitepaper
from the Open Cloud Standards Incubator, 2009.
Distributed Management Task Force (DMTF) Architecture of Managing Clouds,
Whitepaper from the Open Cloud Standards Incubator, 2010.
EuroCloud: Cloud Computing, Recht, Datenschutz & Compliance. Leitfaden des
EuroCloud Deutschland_eco e. V., Köln, November 2010.
Glaeser, L., Laudel. G., Experteninterviews und qualitative Inhaltsanalyse: als
Instrumente rekonstruierender Untersuchungen Vs Verlag; Auflage: 4. Auflage, 2010.
Grance, T.; Mell, P.: The NIST definition of Cloud Computing. National Institute of
Standards and Technology (NIST), Juli 2009, http://csrc.nist.gov/groups/SNS/cloudcomputing/cloud-def-v15.doc, abgerufen am: 31.12.2010.
Günther, O et. al.: Application Service Providers: Angebot, Nachfrage und
langfristige Perspektiven. In. WIRTSCHAFTSINFORMATIK, 2001 Ausgabe Nr.:
2001-06.
Hilley, D.: Cloud Computing: A Taxonomy of Platform and Infrastructure-level
Offerings. CERCS Technical Report, April 2009.
Hoefer, C.N; Karagiannis, G.: Taxonomy of cloud computing services. IEEE
Globecom 2010 Workshop on Enabling the Future Service-Oriented Internet.
Hay, B.; Nance, K.; Bishop, M.: Storm Clouds Rising: Security Challenges for IaaS
Cloud Computing, Proceedings of the 44th Hawaii International Conference on
System Sciences 2011.
Hoch, D.J., and Freking, U., Cloud Computing, SaaS und SOA als Bausteine im
Outsourcing der Zukunft, McKinsey on the 7. Entscheiderforum Outsourcing Bad
Homburg v.d.Höhe, November 2009.
Iyer, B.; Henderson, J.C.: Preparing for the future: Understanding the seven
capabilities of Cloud Computing. MIS Quarterly Executive Vol. 9, No. 2, 2010.
Jaeger, P.T.; Lin, J.; Grimes, J.M.: Cloud Computing and Information Policy:
Computing in a Policy Cloud?. Journal of Information Technology & Politics, Vol.
5(3), 2008.
Koehler, P., Anandasivam, A., Dan, M.A., Weinhardt, C., Customer heterogeneity
and tariffs biases in Cloud Computing, ICIS, 2010.
174
[Ko10b] Koehler, P., Anandasivam, A., Dan, M.A., Weinhardt, C., Cloud Services from a
Consumer Perspective, AMCIS, 2010.
[La10] Lackermair, G., Strahringer, S. and Mandl, P., Dynamically Scalable Architectures
for E-Commerce; MKWI 2010 E-Commerce und E-Business, 2010.
[Le09] Lehmann, S., and Buxmann, P., Pricing Strategies of Software Vendors, Business
Information Systems Engineering (1:6), pp. 452-462, 2009.
[Le10] Leimeister, S.; Böhm, M.; Riedl, C.; Krcmar, H.: The Business Perspective of Cloud
Computing: Actors, Roles, And Value Networks. 18th European Conference on
Information Systems ECIS, 2010.
[LC10] Leong, L., and Chamberlin, T., Magic Quadrant for Cloud Infrastructure as a
Service and Web Hosting. Gartner RAS Core Research, December 2010.
[Li10]
Li, A., Yang, X., Kandula, S., and Zhang. M., CloudCmp: Comparing Public Cloud
Providers, Internet Measurement Conference, Nov 2010.
[Ma10] Martens, B., Teuteberg, F., and Gräuler, M., Datenbank und Reifegradmodell für die
Auswahl und Bewertung von Cloud-Computing-Services, HMD-Praxis der
Wirtschaftsinformatik, Heft 275, S.52-61, 2010.
[Ma11] Martens, B., Teuteberg, F. and Gräuler, M., Design and Implementation of a
Community Platform for the Evaluation and Selection of Cloud Computing Services:
A Market Analysis, in Proceedings of the 19th European Conference on Information
Systems, Helsinki, 2011.
[Op09] Opencloudmanifesto.org Open Cloud Manifesto -Dedicated to the belief that the
cloud should be open, www.opencloudmanifesto.org, 2009.
[Pe09] Peng J. et. al.: Comparison of Several Cloud Computing Platforms. Second
International Symposium on Information Science and Engineering, IEEE 2009.
[Pr09] Pring, B., Brown, R. H., Frank, A., Hayward, S. and Leong, L., Forecast: Sizing the
Cloud; Understanding the Opportunities in Cloud Services, 2009.
[PO09] Prodan, R; Ostermann, S.: A Survey and Taxonomy of Infrastructure as a Service and
Web Hosting Cloud Providers. IEEE/ACM International Conference on Grid
Computing, 2009.
[Ri09] Rimal, B. et. al.: A Taxonomy and Survey of Cloud Computing Systems. Fifth
International Joint Conference on INC, IMS and IDC, IEEE, 2009.
[RZ11a] Repschläger, J.; Zarnekow, R.: Cloud Computing in der IKT-Branche: Status-quo
und Entwicklung des Cloud Sourcing von KMUs in der Informations- und
Kommunikationsbranche in der Region Berlin Brandenburg, Studie des IKMLehrstuhls der TU-Berlin und dem Verband der Software-, Informations- und
Kommunikations-Industrie in Berlin und Brandenburg (SIBB e.V.), 2011.
[RZ11b] Repschlaeger, J., and Zarnekow, R., Umfrage zur Anbieterauswahl und
Markttransparenz in der Cloud, survey from the technical university Berlin within
the IT Operations Day, 2011.
[Ru10] Russell, S., Yoon, V., and Forgionne, G., Cloud-based decision support systems and
availability context: the probability of successful decision outcomes, Information
Systems and eBusiness Management, Vol.8, Iss. 3, p.189-205, 2010.
[Sa11a] SaaS-EcoSystem: SaaS-EcoSystem Check-Liste. SaaS-EcoSystem e. V., 2011.
http://www.saasecosystem.org/trust-in-cloud/checkliste/, aufgerufen am 19.04.2011.
175
[Sa11b] Saunders, C.: MIS Journal rankings.
http://ais.affiniscape.com/displaycommon.cfm?an=1&subarticlenbr=432.
[Sa10] Saya, S., Pee, L.G., and Kankanhalli, A.,The impact of insitutional influrences on
perceived technological characteristics and real options in Cloud Computing
adoption, ICIS, 2010.
[Sc10] Schwarz, A. et. al.: A Conjoint Approach to Understanding IT Application Services
Outsourcing. Journal of the Association for Information Systems, Vol. 10: Iss. 10,
Article 1, 2010.
[SH09] Schrader, U.; Hennig-Thurau, T.: VHBJOURQUAL2. Business Research 2(2):180
204, 2009.
[Ta10] Talukder, A.K., Zimmerman, L., and Prahalad, H.A., Cloud Economics: Principles,
Costs, and Benefits, in Cloud Computing: Principles, Systems and Applications, N.
Antonopoulos and L. Gillam (eds.), Springer-Verlag London, pp. 343-360, 2010.
[Th09] The Open Group, Cloud Computing Business Scenario Workshop, T. Blevins, D.
Lounsbury, M. Kirk, and C. Harding, Report published by the Open Group, August
2009.
[Ts10] Tsvihun, I.; Stephanow, P.; Streitberger, W.: Vergleich der Sicherheit traditioneller
IT-Systeme und Public Cloud Computing Systeme. Fraunhofer-Institut für sichere
Informationstechnologie (SIT), Juli 2010.
[Ts08] T-Systems White Paper Cloud Computing. Alternative sourcing strategy for
business ICT, T-Systems Enterprise Services GmbH, 2008.
[Va09] Vaquero, L.M.; Merino, L.M.; Caceres, J.; Lindner, M.: A Break in the Clouds:
Towards a Cloud Definition. ACM SIGCOMM Computer Communication Review
Volume 39, Nr. 1, 2009.
[VJ10] Velton, C.; Janata, S.: Cloud Vendor Benchmark 2010 Cloud Computing Anbieter im
Vergleich Deutschland. Experton Group AG, 2010.
[Wa08] Wang, L., and Laszewski, G.v., Scientific Cloud Computing: Early Definition and
Experience, in IEEE International Conference on High Performance Computing and
Communications, Dalian, China, 2008.
[We09] Weinhardt, C.; Anandasivam, A.; Blau, B.; Borissov, N.; Meinl, T.; Michalk, W.;
Stößer, J.: Cloud-Computing Eine Abgrenzung, Geschäftsmodelle und
Forschungsgebiete. In: Wirtschaftsinformatik, Nr. 5, 2009, S.453-462.
[WW02] Webster, J.; Watson, R.T.: Analyzing the past to prepare for the future: Writing a
literature review. MIS Quart 26(2):1323, 2002.
[YT09] Yang, H.; Tate, M.: Where are we at with Cloud Computing?: A Descriptive
Literature Review. 20. Australasian Conference on Information Systems, Melbourne,
2009.
176
Assessing Process Models with Cognitive Psychology
Stefan Zugal, Jakob Pinggera, Barbara Weber
{stefan.zugal, jakob.pinggera, barbara.weber}@uibk.ac.at
Abstract: The importance of a business process model to be understandable to its
reader is widely acknowledged. In this vein, several approaches to assess and improve
understandability exist, such as theoretical quality frameworks, modeling guidelines
and process model metrics. In this paper we propose to investigate the issue of understandability from the angle of cognitive psychology. To this end, we discuss how the
cognitive process of inference acts as a central process of problem solving. In particular, we illustrate in how far chunking, computational offloading and external memory
might have an impact on the understandability of process models. Our propositions
are theory-based so far and will serve as basis for planned empirical investigations, as
discussed in the research agenda.
Keywords: Business Process Modeling, Model Understandability, Cognitive Psychology.
1 Introduction
In order to support the analysis and design of, for example, process-aware information systems, service-oriented architectures and web services, business process models or process
models for short, are employed [RM11]. Especially process models used in project phases
that are concerned with requirement documentation are needed to be intuitive and easy
to understand [WW02]. As consequence, various process modeling languages have been
proposed, each one of them claiming superiority, cf. [FMR+ 09]. To contribute to an objective discussion, we employ cognitive psychology as tool for assessing and discussing the
understandability of process models. More specifically, we embed the concept of inference
as central process for problem solving in the context of business process modeling.
The remainder of this paper is structured as follows. Section 2 discusses related work.
Then, Section 3 introduces concepts from cognitive psychology and puts them in the context of business process modeling. Finally, Section 4 summarizes the findings and concludes the paper with a research agenda.
2 Related Work
The understandability of process models is approached from various angles. For instance, in [BRU00], a theoretical point of view is chosen. Complementary, several studies report from empirical investigations. Most works thereby employ the concept of
177
metrics computed on structural aspects of the process model to assess understandability,
e.g., [MRC07, VRvdA08, RM11, Car06, MMRS09]. An overview of metrics from software engineering, adapted to business process models can be found in [CMNR06, GL06].
Another interesting work to be mentioned is [MRvdA10], in which empirically validated
guidelines for process modeling with the aim to improve understandability are presented.
Common to all these approaches is that the model is in the focus of investigation. In our
work, however, we use cognitive psychology to focus on the interplay of model and human
to examine understandability.
3 Cognitive Concepts for Understanding Process Models
Basically, three different problem-solving “programs” or “processes” are known from
cognitive psychology: search, recognition and inference [LS87]. Search and recognition
allow for the identification of information of rather low complexity, i.e., locating an object or the recognition of patterns. Most conceptual models, however, go well beyond
complexity that can be handled by search and recognition. Here, the human brain as
“truly generic problem solver” [Tra79] comes into play. Thereby, cognitive psychology
differentiates between working memory that contains information that is currently being
processed, as well as long-term memory in which information can be stored for a long
period of time [PTTG03]. Most severe, and thus of high interest and relevance, are the
limitations of the working memory. As reported in [Mil56], the working memory can not
hold more than 7±2 items at the same time. In addition, information held in the working
memory decays after 18–30 seconds if not rehearsed [Tra79]. To illustrate how severe
these limitations are, consider the sequence A-G-K-O-M-L-J. The average human mind is
just capable of keeping this sequence in working memory, and, after 18–30 seconds the
sequence will be forgotten.
The importance of the working memory has been recognized and led to the development
and establishment of Cognitive Load Theory (CLT), meanwhile widespread and empirically validated in numerous studies [Ban02]. Theories of CLT revolve around the limitations of the working memory and how these limitations can be overcome—especially the
second question is of interest for the understandability of process models. Hence, subsequently, we will discuss three strategies for dealing with the working memory’s limits.
First, we will discuss the chunking of information. Second, we will show how process
models can support inference through computational offloading. Third, we will introduce
external memory, which allows to free resources in the working memory.
Chunking and Schema Acquisition. Information is believed to be memorized in interconnected schemata rather than in isolation [Gra07]. Those schemata, which are stored in
long-time memory, can be used to aggregate information and handle it as a “chunk” of information in the working memory. Each schemata requires only one slot in working memory when used, hence mental effort is effectively reduced, cf. [New90, CWWW00, Gra07].
The process of aggregating information to a chunk using a schema, in turn, is referred to
178
as “chunking” [Gra07].
To illustrate how chunking actually influences the understandability of business process
models, an example is provided in Figure 1. An unexperienced reader may, as shown
on the left hand side, use three chunks to store the process fragment: one for each XORgateway and one for activity A. In contrast, an expert may, as shown on the right hand side,
recognize this process fragment as a pattern for making activity A optional. In other words,
in her long-time memory a schema for optional activities is present, thereby allowing her
to store the entire process fragment in one slot in the working memory.
A
A
or
in
Memory
Figure 1: Chunking of an Optional Activity
Computational Offloading. In contrast to chunking, which is highly dependent on the
internal representation of information, i.e., how the reader organizes information in his
mind, computational offloading highly depends on the exact external presentation of the
business process model, i.e., visualization of the process model. In particular, computational offloading “refers the extent to which differential external representations reduce the
amount of cognitive effort required to solve information equivalent problems” [SR96]. In
other words, an external representation may provide features that help the reader to extract
information. Instead of computing and inferencing respective information in the modeler’s
mind, information can, as in the process of recognition, more or less be “read-off” [SR96].
To put computational offloading into the context of business process modeling, a simple
example illustrating the described phenomenon is shown in Figure 2. The illustrated process models are information equivalent, i.e., the same execution traces can be produced
based on both Model A and Model B. However, Model A is modeled in BPMN, whereas
for Model B the declarative modeling language ConDec [Pes08] was used (for a detailed
explanation of declarative process models we refer to [Pes08, MPvdA+ 10]). Consider the
task of listing all execution traces that can be inferred from the process model. A reader
familiar with BPMN will probably infer within a few seconds that Model A supports execution traces <A, B, D> and <A, C, D>. Such information is easy to extract as BPMN
provides an explicit concept for the representation of execution sequences, namely sequence flows. Thus, for identifying all possible execution traces, the reader simply follows
the process model’s sequence flows—the computation of all execution traces is offloaded
to the process model. In contrast, for Model B, no explicit representation of sequences is
present. Rather, constraints define the interplay of actions and do not necessarily specify
sequential behavior. Thus, the reader cannot directly read off execution traces, but has
to interpret the constraints in her mind to infer execution traces. In other words, process
model B, while information equivalent to Model A, does not provide computational of-
179
floading for extracting execution traces. Consequently, even for a reader experienced with
ConDec, listing all supported traces is far from trivial.
) Model B
a) Model A
0 1
B
B
A
D
1
1
A
D
0 1
C
C
Figure 2: Computational Offloading
External Memory. Finally, we would like to introduce another mechanism that is known
for reducing mental effort, i.e., the amount of working memory slots in use. In particular,
we will discuss the concept of external memory in the context of business process models.
An external memory is referred to any information storage outside the human cognitive
system, e.g., pencil and paper or a black board [Swe88, Tra79, SR96, LS87]. Information
that is taken from the working memory and stored in an external memory is then referred
to as cognitive trace. In the context of a diagram, a cognitive trace would be to, e.g., mark,
update and highlight information [SR96]. Likewise, in the context of process modeling,
the model itself may serve as external memory. When interpreting a process model, marking an activity as executed while checking whether an execution trace is supported, can be
seen as leaving a cognitive trace.
For the illustration of external memory and cognitive traces, consider the process model
shown in Figure 3. Assuming the reader wants to verify, whether execution trace <A, D,
E, F, G, H> is supported by the process model. So far, as indicated by the bold path and
the position of the token, she has “mentally executed” the activities A, D, E, F and G.
Without the aid of external memory, she will have to keep in the working memory, which
activities have been executed, i.e., sub-trace <A, D, E, F, G>, as well as the position of
the token within the process instance. By writing down the activities executed so far, i.e.,
transferring this information from working memory to external memory (e.g., piece of
paper), load on working memory is reduced. In addition, the process model even allows
to store the “mental token”—either by simply putting a finger on the respective part of the
process model or by marking the location of the token, as shown in Figure 3.
4 Summary and Research Agenda
In this paper we discussed the concept of inference from cognitive psychology in the
context of business process modeling. In particular, we described how chunking, computational offloading and external memory can lower mental effort and, in turn, support
inference and facilitate the understanding of a process model.
We acknowledge that this approach is still work in progress and conclude with identifying
180
B
A
C
F
D
E
G
Figure 3: External Memory
a research agenda guiding our future research. First and foremost, we are planning to validate the described concepts for process models, i.e., to verify whether the expected effects
can be observed in practice. It is assumed that effects are highly dependent on the exact representation of information, e.g., computational offloading may work differently in
BPMN models than in ConDec [Pes08] models. Hence, we will start validation using wellknown languages like BPMN and then investigate in how far findings can be generalized
to other modeling languages like ConDec. Second, after concepts have been validated,
we envision two major directions. On the one hand, as motivated in the discussion, we
aim at contributing to an objective discussion about the superiority of process modeling
languages. Validated concepts will give us then a tool to assess or predict the understandability of a process model. Such insights may be then used to guide the developement
and improvement of process modeling languages. On the other hand, we plan to use the
introduced concepts to systematically develop software that supports the understandability
of process models. More specifically, as soon as the cognitive processes of understanding
a process model are well enough understood, they can be used to systematically analyze
shortcomings of process modeling languages and to provide appropriate computer-based
support for compensating those shortcomings.
References
[Ban02]
Maria Bannert. Managing cognitive load—recent trends in cognitive load theory.
Learning and Instruction, 12(1):139–146, 2002.
[BRU00]
Jörg Becker, Michael Rosemann, and Christoph Uthmann. Guidelines of Business
Process Modeling. In Business Process Management, Models, Techniques, and Empirical Studies, pages 30–49, London, UK, 2000. Springer-Verlag.
[Car06]
Jorge Cardoso. Process control-flow complexity metric: An empirical validation. In
Proc. IEEE SCC ’06, pages 167–173, 2006.
[CMNR06]
Jorge Cardoso, Jan Mendling, Gustav Neumann, and Hajo A. Reijers. A Discourse
on Complexity of Process Models. In BPM’06, pages 117–128, 2006.
[CWWW00] Andrew W. Crapo, Laurie B. Waisel, William A. Wallace, and Thomas R. Willemain.
Visualization and the process of modeling: a cognitive-theoretic view. In Proc. KDD
’00, pages 218–226, 2000.
181
[FMR+ 09]
Dirk Fahland, Jan Mendling, Hajo A. Reijers, Barbara Weber, Mathias Weidlich, and
Stefan Zugal. Declarative versus Imperative Process Modeling Languages: The Issue
of Understandability. In Proc. EMMSAD ’09, pages 353–366, 2009.
[GL06]
Volker Gruhn and Ralf Laue. Complexity metrics for business process models. In
Proc. BIS ’06, pages 1–12, 2006.
[Gra07]
Peter Gray. Psychology. Worth Publishers, 2007.
[LS87]
Jill H. Larkin and Herbert A. Simon. Why a Diagram is (Sometimes) Worth Ten
Thousand Words. Cognitive Science, 11(1):65–100, 1987.
[Mil56]
George Miller. The Magical Number Seven, Plus or Minus Two: Some Limits on
Our Capacity for Processing Information. The Psychological Review, 63(2):81–97,
1956.
[MMRS09]
Joachim Melcher, Jan Mendling, Hajo A. Reijers, and Detlef Seese. On Measuring
the Understandability of Process Models. In Proc. BPM Workshops ’09, pages 465–
476, 2009.
[MPvdA+ 10] Marco Montali, Maja Pesic, W.M.P. van der Aalst, Federico Chesani, Paola Mello,
and Sergio Storari. Declarative Specification and Verification of Service Choreographies. ACM Trans. Web, 4(1):1–62, 2010.
[MRC07]
Jan Mendling, Hajo A. Reijers, and Jorge Cardoso. What Makes Process Models
Understandable? In Proc. BPM ’07, pages 48–63, 2007.
[MRvdA10]
Jan Mendling, Hajo A. Reijers, and Wil M. P. van der Aalst. Seven process modeling
guidelines (7PMG). Information & Software Technology, 52(2):127–136, 2010.
[New90]
Allen Newell. Unified Theories of Cognition. Harvard University Press, 1990.
[Pes08]
Maja Pesic. Constraint-Based Workflow Management Systems: Shifting Control to
Users. PhD thesis, TU Eindhoven, 2008.
[PTTG03]
Fred Paas, Juhani E. Tuovinen, Huib Tabbers, and Pascal W. M. Van Gerven. Cognitive Load Measurement as a Means to Advance Cognitive Load Theory. Educational
Psychologist, 38(1):63–71, 2003.
[RM11]
Hajo A. Reijers and Jan Mendling. A Study into the Factors that Influence the Understandability of Business Process Models. SMCA, 41(3):449–462, 2011.
[SR96]
Mike Scaife and Yvonne Rogers. External cognition: how do graphical representations work? Int.J. Human-Computer Studies, 45(2):185–213, 1996.
[Swe88]
John Sweller. Cognitive load during problem solving: Effects on learning. Cognitive
Science, 12(2):257–285, 1988.
[Tra79]
William J. Tracz. Computer programming and the human thought process. Software:
Practice and Experience, 9(2):127–137, 1979.
[VRvdA08]
Irene Vanderfeesten, Hajo A. Reijers, and Wil M. P. van der Aalst. Evaluating workflow process designs using cohesion and coupling metrics. Comput. Ind., 59(5):420–
437, 2008.
[WW02]
Yair Wand and Ron Weber. Research Commentary: Information Systems and Conceptual Modeling—A Research Agenda. ISR, 13(4):363–376, 2002.
182
Eye Tracking Experiments in Business Process Modeling:
Agenda Setting and Proof of Concept
Frank Hogrebe1, Nick Gehrke2, Markus Nüttgens3
1
Hessische Hochschule für Polizei und Verwaltung,
Fachbereich Verwaltung, D- 65197 Wiesbaden
2
3
Nordakademie, Hochschule der Wirtschaft,
Köllner Chaussee 11, D-25337 Elmshor
Universität Hamburg, Lehrstuhl für Wirtschaftsinformatik,
Von-Melle-Park 5, D-20146 Hamburg
Abstract. For almost all applications there is a need to judge business process
models by user´s perspective. This paper presents first facts and findings how
the eye tracking method can contribute to a deeper empirical understanding and
evaluation of user satisfaction in business process modeling. The method of eye
tracking is used in order to check subjective perceptions of users through objective measurement. The experimental investigation is done using two variants of
the widespread business process modelling notation Event-driven Process
Chain (EPC).
Keywords: eye tracking, user satisfaction, event-driven process chain (EPC)
1 Introduction
User satisfaction is one of the most widely researched topics in the information systems field [ANC08], [A03]. There are strong relationships between user satisfaction
and intended use or actual use of information systems [ANC08, p.44]. The paper
focuses on user-based requirements in business process modeling. In order to be able
to judge subjective appraisals of test users and real-life users through objective measurements, the method of eye tracking is used in our experiment. For the evaluation of
requirements of user satisfaction in business process modeling we focused on, a research framework is developed and used as part of an experiment. Due to the fact, that
event-driven process chains (EPC) are still establish as a widespread notation in business process modeling, we used two variants of the EPC to show the prospects and
limitations of our approach.
2 Eye Tracking Technique
Eye tracking is a technique with which the course of a person's vision when viewing
an object, model or an application (investigation objects) can be measured [RP07]. In
the taxonomy of research methods, eye tracking is assigned to the user-oriented
183
methods [Y03, p.125]. In the combination of eye tracking and surveys, survey results
are supplemented through quantitative eye tracking data. The test person comments
on his or her activities, feelings and thoughts in the use of the object of investigation
(qualitative aspect). This methodological approach is to be transferred in this
investigation to the measurement and analysis of the user satisfaction in business
process modeling.
3 User Satisfaction Eye Tracking Experiment
3.1 Research Framework and Hypotheses
The research framework has the aim, to create a framework for designing experiments
measuring the user satisfaction by combining surveys and eye tracking methods. Our
aim in this experiment is to compare the user satisfaction regarding two semiformal
modeling notations as the test case. We developed the following basic hypothesis (H1
H3) for this comparison:
H1: Seven user-related requirements deducted from our literature review are acknowledged by users (editors and recipients).
H2: Seven requirements are appropriate criteria to check subjective assessments of
users with objective measurements of experiments.
H3: Eye tracking is an appropriate method for measuring and judging user-related
requirements related to the user satisfaction in business process modeling.
Table 1. Research Framework for the empirical study
Requirements
Understanding
Completeness
Easy to use
Usefulness
Timeliness
Flexibility
Accuracy
Subjective Perception (Assessment of
users in the survey)
Users have been asked: the variant of
the modeling language is easy to
understand?
Users have been asked: the variant of
the modeling language is complete?
Users have been asked: the variant of
the modeling language is easy to use?
Users have been asked: the variant of
the modeling language is useful?
Users have been asked: the variant of
the modeling language is time saving?
Users have been asked: the variant of
the modeling language is flexible?
Users have been asked: the variant of
the modeling language is accurate?
Measurement
(Experiment with users)
A graphical representation of a process is shown to the
users and specific questions are asked. Number of false
answers is counted.
Users have to model a pre-defined process. It is noted if
users are capable to model the process (Editors)
Users have to model a pre-defined process. Number of
modeling errors is counted (editors).
In a given graphical representation of a process users
need to find a pre-defined object. The following is
measured:
*Numbers of fixations with eye tracking methods
*Length of fixation with eye tracking methods.
The more and the longer fixations are, the more the
effort and the worst the balance between effort and
utility.
Users have to model a pre-defined process. The time
needed is recorded (editors).
No measurement since users considered this requirement to be less important (rank 6 of 7).
No measurement since users considered this requirement to be less important (rank 7 of 7).
Due to the fact, that event-driven process chains (EPC) are still establish as a widespread notation in business process modeling, we used two variants of the EPC to
show the prospects and limitations of our approach. In our experiment two variants of
the event-driven process Chain (EPC) are used for validating the seven requirements:
184
the extended EPC (eEPC) and the object oriented EPC (oEPC) [SNZ97]. Our investigation is based on 24 users (12 editors and 12 recipients of models). In each group 6
users have been students and 6 have been employees of a public administration. We
use as our test case graphical representation of public administration processes, because there are good experiences about the notations in this area (see survey, [anonymous]). The users have been asked to subjectively assess the modeling languages
regarding the seven requirements. Furthermore, we did experiments with the users
like measuring time to model a process (for editing users) or counting the number of
errors a user was able to find in a model (for recipient users). Another example is
measuring the length of fixations during solving a specific task by using eye tracking
(editors and recipients). All requirements have been mapped in order to measure how
the requirements are achieved by the two variants of the modeling language. After
that we compare the results of the subjective assessment and the objective experiment
for both, requirements and variants. As a result of our investigation, the most important requirements were understanding (editors: 8, recipients: 11), completeness
(editors: 7, recipients: 6) and easy to use (editors: 5, recipients: 7). In this context,
we asked if a user missed an important requirement but none of the 24 users answered
yes. Thus, we conclude that the seven requirements used in our investigation are
sufficient. Hypothesis 1 can be considered to be confirmed in the context of our experiment. Table 4 depicts the research framework for our empirical study of the userrelated requirements. The column measurement describes the experiment used for
the specific requirement.
3.2 Statistical Analysis
In the following section we analyze the data gathered during our survey regarding the
requirements. For all requirements interviewees could answer with the following
options: I fully agree (=4), I agree partially (=3), I doubt to agree (=2), I do not agree
at all (=1) for each variant of the modeling language (eEPC and oEPC). Due to missing answers the sample size could be less than 12. Table 2 depicts basic statistic ratios
regarding the assessment of the requirements by the interviewees during the survey.
Obviously, all requirements except the requirement accuracy got a better average
assessment regarding the oEPC. The requirement accuracy has been assessed equally. The correlation of the assessments between eEPC and oEPC (right table) only
reveal a significant (negative) relationship [> (-) 0.5] for the requirement time saving. That means interviewees assessing one modeling language as time saving tend to
assess the other language as less time saving. In the following we want to verify if
differences in the assessments of the interviewees are significant regarding each requirement. For this reason we conducted a t-Test testing for different means in dependent samples. Table 3 shows the results of the statistic test.
The column Sig. (2-tailed) shows the significance level of the difference of means
regarding the assessment of the interviewees for each requirement (meaning H0:
eEPC=oEPC; H1: oEPC#eEPC). But we want to focus on the hypothesis that one
modeling language is better than the other one. Thus, we also test H0: eEPC=oEPC;
H1: oEPC>eEPC to validate if the oEPC language dominates the eEPC variant (see
185
column Sig. (1-tailed)). Since we test H0: eEPC=oEPC; H1: oEPC>eEPC the confidence level is always half the confidence level of column Sig. (2-tailed). As a result,
we can see that the requirement flexibility has been considered to be better in respect to the oEPC language on a 95% confidence level whereas the requirements
easy to use and usefulness are significant on a 90% level. In the same manner we
analyzed the data of the survey we now want to consider the results of the experiments as depicted in table 4. Please note that Pair 4 and 5 both are linked to the requirement usefulness since the data represents on the one hand the number of fixations and on the other hand the length of the fixations.
Table 2. Statistic ratios regarding the assessment of requirements by interviewees
Paired Samples Statistics
Mean N
air 1 E C easy-to-use
oE C easy-to-use
air 2 E C Flexi ility
oE C Flexi ility
air
air
air
air
air
2
Std
De iation
00 12
12
11
Std Error
Mean
000
8
8
000
22
2
Paired Samples Correlations
N
11
E C accuracy
oE C accuracy
E C understandin
oE C understandin
E C completeness
oE C completeness
E C timeliness
oE C timeliness
E C usefulness
oE C usefulness
20
2
2
11
11
2 12
12
1 12
12
2
11
18 11
2 82 11
11
8
1
92
8
8
1 009
1
982
2
19
1 9
1 2
2 1
22
0
22
29
20
C
C
C
C
Corr
Si
air 1
air 2
air
air
E
E
E
E
easy-to-use oE C easy-to-use
flexi ility oE C flexi ility
accuracy oE C accuracy
understandin
oE C understandin
12
11
11
0
12
- 120
air
air
air
E C completeness oE C completeness
E C timeliness oE C timeliness
E C usefulness oE C usefulness
12
11
11
9
2 0
91
- 88
2 1
11
12
0
Table 3. t-tests of the assessments of the interviewees for each requirement (different means)
Paired Samples Test
air 1
air 2
air
air
air
air
air
E
E
E
E
E
E
E
aired Differences
90
Std
Std Error
o er
pper
Mean De iation
Mean
8
22
0 0
809
2
-1 0 8 - 19
000
1 000
02 -2 0
218 - 1
1 1
-1
8
2 1 - 99
2
1 2
-1 1
0
10
12 -1 111
021
easy-to-use - oE
easy-to-use
flexi ility - oE
flexi ility
accuracy - oE
accuracy
understandin - oE
understandin
completeness - oE
completeness
timeliness - oE
timeliness
usefulness - oE
usefulness
t
-1 8
-2 09
000
-1 1 9
- 92
-9 9
-1
Si (2- Si (1df tailed) tailed)
11
1
08
10
02
01
10 1 000
00
11
2
1
11
0
2 2
10
0
180
10
111
0
inner
oE
oE
NA
oE
oE
oE
oE
Table 4. Statistical ratios of the experiments regarding the requirements
Paired Samples Statistics
air 1 E
oE
air 2 E
oE
air E
oE
air E
oE
air E
oE
understandin errors
understandin errors
easy-to-use errors
easy-to-use errors
timeliness time
timeliness time
usefulness Num erOfFixations
usefulness Num erOfFixations
usefulness en thOfFixations
usefulness en thOfFixations
Mean
Std
Std Error
N De iation
Mean
2 0
2 92
10 08
2 8
9 00
00
29 1
2 2
1 8 92
1 1 92
12
12
12
12
12
12
12
12
12
12
11
18
12
09
2 89
2 09
28
1
09
89
8
0
18 29
18 1
10 2
1 1
Paired Samples Correlations
N Corr Si
air 1 E
oE
air 2 E
use
air
E
air
E
oE
air
E
oE
understandin errors
understandin errors
easy-to-use errors oE
easy-toerrors
timeliness time
usefulness Num erOfFixations
usefulness Num erOfFixations
usefulness en thOfFixations
usefulness en thOfFixations
12 - 11
12
22
0
12
2
2
12 - 0
12
21
8
2
Regarding table 4 please note that compared to the assessments of the interviewees
small values are better than big ones (e.g. less errors are better than more errors or
less time consuming is better than more time consuming). Obviously, regarding the
186
8
means the oEPC language performs better for all requirements than the eEPC variant
except for the requirement understanding. Focusing on understanding one can see
that users answered with fewer errors considering the eEPC. In this context, we reason that eye tracking is an appropriate method to measure and assess requirements for
user satisfaction in business process modeling in our experiment. Therefore, Hypothesis 3 can be considered to be confirmed. T-tests regarding the differences of the means
of the measurements reveal that the oEPC variant dominates significantly the eEPC
variant (Table 5). Only the requirement understanding cannot be considered significant. The requirements easy to use and timeliness are significant on a 95% level
(see column Sig. (2-tailed)). The two measurements for usefulness are significant
only on a 90% level. Table 8 does not contain the requirement completeness since
all users where able to model the predefined process in the appropriate experiment
with the eEPC and the oEPC variant. Thus, the requirement completeness remains
undecided.
Table 5. t-tests of the experiments / measurements for top 5 requirements
Paired Samples Test
aired Differences
90 Confidence Inter al
Mean
air 1
air 2
air
air
air
E
oE
E
use
E
oE
E
oE
E
oE
understandin errors understandin errors
easy-to-use errors - oE
easy-toerrors
timeliness time timeliness time
usefulness Num erOfFixations usefulness Num erOfFixations
usefulness en thOfFixations usefulness en thOfFixations
Std De iation Std Error Mean
- 2
0
00
92
o er
22
pper
2 98
-
1 09
1 91 2 10 11
8
1
2
2 00
1
2
df
-1 0
12
91
t
-2 1
Si
(2-tailed)
11
Si (1tailed)
2 9
0
0 0
8 11
01
91 8 1
11
12
00
0 2
1 1
11
1
0
inner
E
oE
oE
oE
oE
After having analyzed the results of the survey and the experiments separately, we
now want to focus on a comparison of both as an overall result. Table 6 shows the
results of this comparison.
Table 6. Comparison of survey and results, * 90% significance, ** 95% significance
Requirement
"Winner"
survey
"Winner
experiment
Understanding
Completeness
Easy to use
Usefulness
Timeliness
oEPC
oEPC
oEPC*
oEPC*
oEPC
eEPC
N/A
oEPC**
oEPC*
oEPC**
Flexibility
oEPC**
No measurement
Accuracy
N/A
No measurement
Comment
no significance
no significance
significant for oEPC
significant for oEPC
Experiment significant for
oEPC, Survey not significant
for oEPC
significant for oEPC, no
experiment
no significance
Obviously, a significant result can be detected regarding the requirements easy to
use and usefulness. Only significant in one area, but on an average the oEPC dominates regarding the requirement timeliness. Considering the requirement flexibility, the oEPC has been significantly assessed better than the eEPC, but we did not
conduct an experiment for that requirement. For the requirement accuracy we cannot deduct a winner. Furthermore, for the requirements understanding and com-
187
pleteness there is no significant result. To sum up the oEPC seems to dominate most
requirements (often significant). The eEPC variant succumbs or is equal to the oEPC
variant. As a result, the literature based common user related requirements are appropriate to check subjective assessments of users in a survey with objective experiments. Therefore, hypothesis 2 can be considered to be confirmed regarding our experiment.
4 Summary of the results
The eye tracking method was used for first time in the measurement and assessment
of the user satisfaction in business process modeling. In the application case, the eye
tracking method was used to visualise objective measurement and progression data, in
addition to the survey techniques. The graphical depictions from the eye tracking
(heatmaps) were able to visually support the investigation results. The research
framework was used on the basis of extended and object-oriented event-driven process chains (eEPC and oEPC). We used always the same modeling tool to minimize
external effects when testing our framework (bflow Toolbox [BT10]). To what extent
investigations on the basis of other semi-formal modelling languages lead to other
results remains reserved for other investigations. The hypotheses (H1 to H3) could be
confirmed on the basis of two variants of the event-driven process chain (EPC). To
what extent investigations on the basis of other semi-formal modelling languages in
the field of business process modeling lead to other results remains to be validated in
follow-up investigations. The initial results of this investigation are interesting, in
particular with regard to the options to visualise measurement results in the representation of heatmaps. Here, however, further investigations should follow in order to be
able to validate the relevance of the eye tracking method for investigations in the
research area of information systems moreover.
References
[A03]
Aladwani, A.M.: A Deeper Look at the Attitude-Behavior Consistency Assumption in
Information Systems Satisfaction Research. In: The Journal of Computer Information
Systems (44:1), (2003), pp. 57--63
[ANC08] Au, N., Ngai, E.W.T., Cheng, T.C.E.: Extending the Understanding of End User
Information Systems Satisfaction Formation: An Equitable Needs Fulfillment Model
Approach. In: MIS Quarterly Vol. 32 (2008), 1, pp.4366
[BT10] Bflow Toolbox. www.bflow.org. Accessed 2011-07-31
[RP07] Rosen, D.E., Purinton, E.: Website Design: Viewing the Web as a Cognitive Landscape. In: Journal of Business Research, 57 (2007) 7, pp.787--794
[SNZ97] Scheer, A.W., Nüttgens, M., Zimmermann, V.: Objektorientierte Ereignisgesteuerte
Prozeßketten (oEPK) Methode und Anwendung (in german). In: working paper vol.
141, institute of information systems, (1997), Saarbrücken (germany)
[Y03]
Yom, M.: Web usability von Online-Shops. Better-Solutions-Verlag Gierspeck,
Göttingen (2003)
188
Creative Personality and Business Process Redesign
†
Kathrin Figl† , Barbara Weber‡
Vienna University of Economics and Business (WU Vienna)
kathrin.figl@wu.ac.at
‡
University of Innsbruck
barbara.weber@uibk.ac.at
Abstract: The purpose of this article is to discuss the influence of creative personality
on process redesign. Building on creativity theories stemming from the field of cognitive psychology, we identify important individual factors during process redesign, and
hypothesize their contributions to creative process design using a modelling tool. We
present an integrated research model and illustrate how we seek to test the model using
the Cheetah Experimental Platform.
1 Introduction
It has long been recognized that business process redesign and innovation can lead to improved processes and better business performance and customer satisfaction. Process improvement was the number one item for top management according to the Gartner Group
in 2010 [Gro10]. In practice, process models are used for analyzing, documenting and
improving business processes. They help to identify process weaknesses as well as possible improvement opportunities. An important factor for process improvement are the
employees working with the process models. For creating novel and valuable ideas for
process redesign human creative ability is needed. Cook describes creativity as “source of
competitive strength within organizations” [Coo98, 179]. Companies have to identify their
creative potential to enable business innovation. Identifying individual creative potential is
relevant for assembling teams and deciding who should be part of a process redesign team.
Additionally, creativity training can be of relevance for teams working on processes.
Despite increasing consciousness about the need to consider creativity in researching business innovation, most research up to now was undertaken in the field of product innovation. The business process management community has barely recognized the importance
of creativity and little research has been undertaken to improve and understand the relationship between creativity and business process redesign. There are few exceptions as
for instance Seidel et al. [SR08] who discuss the concept and management of creativity in
business process management.
Given the high practical relevance of creative process redesign, the aim of this paper is
filling that gap by investigating the relationship between individual creativity and business process redesign. More specifically, we address the research question how creative
189
personality style and creative capacity influence creativity in process redesign tasks.
2 Theoretical Background
Business Process Redesign. Business process reengineering and continuous process improvement constitute two different ends on a continuum of business process redesign.
Business process reengineering refers to “fundamental rethinking and radical redesign of
business processes to achieve dramatic improvements in critical, contemporary measure of
performance, such as cost, quality, service and speed” [HC93]. Thereby, business process
reengineering largely ignores the existing process and in the most extreme case follows a
clean sheet approach creating the redesigned process from scratch. Continuous process
improvement, in turn, typically takes existing processes as a starting point and gradually
improves them in an incremental manner [Dav93]. While both approaches differ in their
focus, both aim at the redesign of existing business processes and lead to a transformation
of the original process model (i.e., AS-IS process) into a redesigned version (i.e., TO-BE
process) improving one or more performance measures. To achieve this transformation
different redesign patterns can be used as for instance cutting non-value adding tasks,
changing the order of tasks or automating tasks [RLM05].
Creativity and Creative Personality. Creativity and creative problem-solving form the
basis for business innovation and process redesign [MM02]. Similarly to other design activities, there is no single solution and no clear path to a solution in a process redesign
task. A creative solution is defined as being both original and novel as well as relevant and
valuable in the specific context [Ama83, 358]. When researching creativity three main
aspects are relevant: the creative person, the creative process and the creative product
[Ama83]. In our study we intend to include all three aspects and examine how creative
personality influences the solutions to a process redesign-task (creative product) using a
modelling tool (creative process). One of the most influential personality models in the
context of creativity is the Adaption-Innovation theory by Kirton [Kir76]. While adaptors tend to “doing things better”, innovators are likely to be “doing things differently”
[Kir76, 622]. Adaptors work out few, but realizable solutions, they pay attention to details and work out the solution incrementally with well-known techniques [FH92, 967]. In
contrast, innovators propose more, but less realistic solutions and might ignore rules and
existing paradigms. This distinction directly relates to the continuum between continuous,
incremental improvement [Dav93] and radical redesign of processes [HC93].
3 Research Model and Hypotheses
Having laid out the relevant theoretical foundation related to creativity and business process redesign, we will now draw several propositions to suggest which factors will influence business process redesign. Concerning person-related characteristics creative style
and creative capacity are two independent dimensions [Kir78, 697] which are both rel-
190
evant for solutions to process improvement tasks. Besides person-related characteristics
there are two more factors contributing to the creative process [Ama83]: domain-relevant
skills and task motivation. We summarize our expectations about relevant influence factors
in light of the theoretical considerations in the research model shown in Figure 1.
Figure 1: Research Model
The model proposes that the quality of a redesigned model in a process redesign task
will be a function of person-related characteristics (style and capacity), process modelling
competence (domain-relevant skills) and task motivation. Concerning the influence of creative style we anticipate the following effects: In contrast to innovators, adaptors perceive
“boundaries less elastic and permeable” [Kir78, 697]. Therefore, we expect adaptors to
make only small, continuous improvements to the model, innovators to radically change
the process and do a complete reengineering. Innovators are more likely to “break existing
patterns” [Mud95, 167]. As innovators do not pay as much attention to details they might
make more syntax errors, but produce more innovative ideas. We state:
H1: Higher innovative style is positively associated with the amount of model changes.
H2: Higher adaptive style is positively associated with the correctness of model solutions.
According to [GT02, 1] there are two basic indicators of creativity: fluency is defined
as “the ability to produce quantities of ideas which are relevant to the task instruction”
and originality as “the ability to produce uncommon ideas or ideas that are totally new or
unique”. We expect creative capacity to positively influence both criteria. Therefore:
H3: Creative ability is positively associated with originality and fluency in the redesign.
Next, we consider process modelling competence and task motivation. We speculate both
factors to contribute to the creative quality of process redesigns in terms of fluency and
originality. While lower modelling experience might hinder creative persons from fully
unfolding their capacity to provide creative ideas, high task motivation might foster creativity.
H4: Process modelling competence is positively associated with originality and fluency in
the redesign.
H5: Task motivation is positively associated with originality and fluency in the redesign.
191
4 Planned Empirical Study
For answering these research questions we are currently conducting an empirical study
using the Cheetah Experimental Platform [PZW10]. Cheetah guides participants through
a variety of questionnaire parts, offers a tutorial on process modelling and a process modelling tool for the redesign tasks, which logs every modelling action (e.g., adding and
deleting of activities) to enable later analysis. The next paragraphs introduce main parts of
the experimental design.
4.1
Measurement of Independent Variables
Creative Style: Innovative vs. Adaptive Problem-Solving Style (KAI (Kirton AdaptionInnovation Inventory): The Kirton Adaption-Innovation Inventory [Kir76] measures individual problem-solving style relating to the quality of problem solutions. Respondents
have to rate themselves on 32 items. It measures three different scales: sufficiency of
originality, efficiency and rule governance.
Creative Capacity: Abbreviated Torrance Test of Creative Thinking (ATTA): The
Torrance Tests of Creative Thinking (TTCT) measures divergent thinking and assesses the
quantity and quality of creative ideas. It is a widely used (by over 2000 studies) instrument
[GT02]. For the purpose of our study we use the verbal subscale of the abbreviated test
version as a screening indicator of creative thinking abilities. The scores in the test are
based on fluency (number of ideas) and originality (unusualness of ideas).
Process Modelling Competence: Participants are asked about the extent to which they
have previously been involved with modelling in education and work. Additionally, we
use a test on theoretical knowledge of process modelling developed by [MS08].
Task Motivation: For measuring intrinsic motivation we use three items derived from a
scale by [DBW92]. An example item was: ‘I found the tasks of providing improvement
ideas for the processes to be enjoyable.’
4.2
Measurement of Dependent Variables: Process Redesign
Experimental Redesign Tasks In our experiment participants are asked to work out improved TO-BE models for two given AS-IS process models. We use 2 measure-invoked
redesign tasks [SK10] targeting the measure- (customer) quality for the construction of
creative redesign tasks. The two AS-IS models were selected from different domains such
that we could expect that they are understandable with no special domain knowledge. In
the first task we ask participants to optimize a new check in and guidance service of an
airline to best support costumers as well as business interests. The second task asks for a
process redesign of a coffee shop service to foster customer satisfaction.
Measurement of Process-Redesigns For analyzing the process redesigns we intend to use
192
quantitative as well as qualitative measures.
• Change Distance: The amount of process changes can be measured as the number
of change operations needed to obtain the redesigned TO-BE process model from
the original AS-IS process model. Thereby, this measure considers operations related to the insertion or deletion of activities/edges.
• Correctness is assessed in terms of syntax as well as execution semantics. In particular, we measure whether syntactic requirements imposed by BPMN are met, and
whether the model is free of behavioral anomalies such as deadlocks. To this end,
we apply the soundness criterion for syntactically correct models [van98].
• Creativity of Redesign Ideas is assessed by a team of independent experts according to originality and fluency. We deploy an iterative consensus-building process to
ensure validity and reliability of our assessment.
5 Example of Preliminary Analysis
Figure 2: Detail of Airport Model and Exemplary Redesigns
Data collection is still ongoing and participants are being recruited from modelling courses.
We would like to discuss two redesign examples of a process snippet, as can be seen in
Figure 2. The study participant with the ‘adaptive solution’ scored low on the AdaptiveInnovative continuum (2.60 on a scale from 1 to 5), the participant with the ‘innovative’
solution scored high (3.49). In the redesign solution we can see, that the ‘adaptor’ simply added one activities to notify the customer if the flight is not delayed, demonstrating
incremental improvement. The ‘innovator’ added two activities and he reengineered the
complete order of the process. This example would support hypothesis 1, but further data
collection and detailed analysis will be needed to test hypotheses.
References
[Ama83] T. M. Amabile. The social psychology of creativity: A componential conceptualization.
Journal of Personality and Social Psychology (1983), 45(2):357–376, 1983.
193
[Coo98] Peter Cook. The creativity advantage - is your organization the leader of the pack? Industrial and Commercial Training, 30:179–184, 1998.
[Dav93] T.H. Davenport. Process innovation: reengineering work through information technology.
Harvard Business School Press, Boston, MA., 1993.
[DBW92] Fred D. Davis, Richard P. Bagozzi und Paul R. Warshaw. Extrinsic and Intrinsic Motivation to Use Computers in the Workplace. Journal of Applied Social Psychology,
22(14):1111–1132, 1992.
[FH92] Gordon R. Foxall und Paul M. W. Hackett. The factor structure and construct validity of the
Kirton adaption-innovation inventory. Personality and Individual Differences, 13(9):967–
975, 1992.
[Gro10] Gartner Group. Leading in Times of Transition: The 2010 CIO Agenda (EXP Premier
Report No. January 2010). Bericht, Gartner, Inc., 2010.
[GT02] Kathy Goff und E. Paul Torrance. Abbreviated Torrance Test for Adults (ATTA). Manual.
Scholastic Testing Service, Bensenville, Illinois, 2002.
[HC93] M. Hammer und J. A. Champy. Reengineering the Corporation: A Manifesto for Business
Revolution. Harper Business Books, 1993.
[Kir76] M. J. Kirton. Adaptors and innovators: A description and measure. Journal of Applied
Psychology, 61(5):622–629, 1976.
[Kir78] M. J. Kirton. Have adaptors and innovators equal levels of creativity?
Reports, 42:695–698, 1978.
Psychological
[MM02] R. McAdam und J. McClelland. Individual and team-based idea generation within innovation management: organisational and research agendas. European Journal of Innovation
Management, 5:86–97, 2002.
[MS08] Jan Mendling und Mark Strembeck. Influence Factors of Understanding Business Process
Models. In Proc. BIS 2008, Seiten 142–153. Springer, 2008.
[Mud95] Samuel Mudd. Kirton adaption-innovation theory: organizational implications. Technovation, 15(3):165–175, 1995.
[PZW10] J. Pinggera, S. Zugal und B. Weber. Investigating the Process of Process Modeling with
Cheetah Experimental Platform. In Proc. ER-POIS ’10, Seiten 13–18, 2010.
[RLM05] H. A. Reijers und S. Liman Mansar. Best practices in business process redesign: an
overview and qualitative evaluation of successful redesign heuristics. Omega, 33(4):283–
306, 2005.
[SK10] Avraham Shtub und Reuven Karni. Business Process Improvement. In ERP, Seiten 217–
254. Springer US, 2010.
[SR08] Stefan Seidel und Michael Rosemann. Creativity Management The New Challenge for
BPM. BPTrends, 2008.
[van98] W.M.P van der Aalst. The Application of Petri Nets to Workflow Management. JCSC,
8(1):21–66, 1998.
194
Abstractions in Actor and Activity Modeling
Matthias Wester-Ebbinghaus, Daniel Moldt
University of Hamburg, Department of Informatics
Vogt-Kölln-Straße 30, D-22527 Hamburg
{wester,moldt}@informatik.uni-hamburg.de
Abstract: In this paper we argue that actor-centered models are well suited for sociotechnical systems of systems (like enterprise and especially cross-enterprise scenarios). Results can especially be drawn from the field of multi-agent system modeling.
However, existing approaches reveal a lack of possibilities for switching between different levels of abstraction. This does not only concern more or less abstract models
for a given situation (model abstraction), but also to have models with actors of varying granularity, including individual as well as collective actors (actor abstraction).
We present a modeling approach that addresses both these aspects. It is based on the
core concepts of actors and activities and especially the concept of a collective actor
is emphasized. For supporting different levels of model abstraction, we present multiple modeling techniques. The semantic core of all models is based on high-level
Petri nets, although this is hidden for the more abstract models.
1 Introduction
Modern perspectives on software systems (e.g. ultra large scale systems [Nor06] or application landscapes [Mat08]) heavily rely on the concept of systems of systems. Core
distinguishing features of systems of systems are the operational as well as managerial independence of component systems. In addition, they are software-intense sociotechnical systems where a suitable joint optimization of social and (software-)technical
system parts has to be pursued.
Agent-orientation has emerged as a powerful and general conceptual framework for
socio-technical systems of systems. Both the idea of independent, loosely-coupled component systems (the agents) and the establishment of system-level/organizational structures are addressed equally [Jen00, Dig09].1 As in a system of systems setting, a component system might be a system of systems itself, approaches for collective agency have
been brought forth in recent years (cf. [HMMF08, AG09]). These are motivated by the
social and especially organization-theoretic concept of collective actors: Collectivities
(networks of social relations between social actors) are Janus-faced, social contexts as
well as collective actors that are embedded in more global contexts.
Nevertheless, on the modeling side, there still exists what we call an actor abstraction
problem as there is a lack of modeling techniques that allow for a seamless integration of
1 Many of the concepts and ideas of agent-orientation nowadays appear in the “guise” of service-oriented
systems (cf. concepts like autonomous, intelligent or collaborating services where the service metaphor is quite
stressed and the agent metaphor is better suited). However, we will not deepen on this topic in this paper.
195
actors of varying granularity/abstraction. In this paper we present a modeling approach
with the specific aim of addressing this problem. Our main goal is to support planning and design of actor systems, possibly including social as well as artificial actors.
We address abstraction in two dimensions: (1) more or less abstract models for specific
scenarios (model abstraction), (2) possibilities to integrate actors of different abstraction/granularity in scenarios (actor abstraction). For more or less model abstraction, we
present three modeling techniques and each of the techniques incorporates the concept
of collective actors.
2 Different Levels of Abstraction for Actor/Activity Models
So called Actor/Activity-Flow Models provide the semantic core for the more abstract
models. They are high-level Petri net models and thus have a precise operational semantic. Figure 1 shows an Actor/Activity-Flow Model of two Collectivity Units. With
a Collectivity Unit, we denote an entity that is a context for embedded actors and that
may actually be an actor itself. A Collectivity Unit is modeled in terms of actor positions
and activity patterns. Actors initiate activities, participate in them and may be added or
removed from their positions in the course of activities. In the example, we show two activity patterns transaction and replace-seller for Collectivity Unit A and the associated
actor positions seller, buyer, governor and administrator. An activity pattern consists
of connected transitions that represent actions. Places connect the transitions in a way
to make up the life lines of activity roles (vertically connected) and exchanges between
activity roles (horizontally connected). For each activity pattern, we have one role that
is the initiating role. In the example we have the two roles sender and receiver for the
transaction activity pattern (where the sender is the initiating role) and two exchanges
between them (for an item sent and the following acknowledge).
The example also includes a second Collectivity Unit B. This Collectivity Unit is actually an actor itself and is embedded in the context of Collectivity Unit A as a buyer.2
The figure exemplifies that collective agency in our model results from what we call role
implementation via role implementation. On the level of Collectivity Unit A, there is a
buyer actor embedded that implements the role receiver in the context of a transaction activity and thus carries out the actions for receiving an item and for sending an
acknowledge. Here however, this actor is itself a Collectivity Unit and in order to implement the role receiver on the next higher system level, there are the two internal roles
distributor and inspector. For these internal roles, there are specified several peripheral
actions that correspond to (/synchronize with) the actions for the role receiver on the
next higher system level. To summarize, Collectivity Unit B implements the role receiver in the context of Collectivity Unit A by means of embedding some actors itself
that in turn implement the roles distributor and inspector where peripheral actions synchronize with the actions of the receiver role. In general, such action recursion across
system levels can be arbitrarily deep and individual actors (not modeled here) represent
the recursion termination.
2 Such hierarchies of Petri nets being embedded in other Petri nets is covered by the high-level Petri net
concept of nets-within-nets [Val04] and is for example supported by the Renew tool (www.renew.de).
196
Figure 1: Excerpt from a more complex Collectivity Unit
Actor/Activity-Flow Models are quickly becoming large and unhandy. Thus we will
present means for abstracting away from details. In the upcoming models, the underlying Petri net semantic is no longer directly visible but it still provides the basis for
understanding them. The first abstraction step is in omitting details of activity flows
and instead just referring to the roles of activities. Figure 2 shows an example of such
an Actor/Activity-Role Model. It is the abstraction of the Actor/Activity-Flow Model
from Figure 1. Actor positions are now just associated with activities (initiation, participation, addition, removal associations) and internal details of the activities are hidden.
The information revealed is similar to common Use Case Diagrams. However, for our
purpose to regard Collectivity Units as collective actors, we need an extension in the
form of associations across system levels. We use a notation of combining activity roles
via logical operators. Above, we have characterized collective agency basically as role
197
Figure 2: Abstracting away from the activity flows of Figure 1
implementation via role implementation. For Actor/Activity-Role Model models, we introduce AND/XOR configurations. For each AND/XOR configuration, we combine a set
of roles of an activity pattern via a logical AND. For each instance (concrete activity) of
the activity pattern, these roles together implement exactly one role of a set of roles that
are combined with a logical XOR. In the example, it is simply modeled (just as before)
that the implementation of the roles distributor and (logical and) inspector of a reception activity of Collectivity Unit B lead to the implementation of the role receiver of a
transaction activity of Collectivity Unit A.
One might also want to consider additional logical operators for combining roles. For
now, we have intentionally chosen the restricted means of AND/XOR configurations
here as in our opinion, they lead to simple yet expressive possibilities for cross-level role
associations. Two canonical applications are shown in Figure 3.
We introduce one further step of abstraction. Instead of regarding singular actor positions
and activity patterns, we regard arbitrarily comprehensive actor and activity sets and
consequently Actor-/Activity-Set Models. The transition from the former Actor/ActivityRole Models to Actor-/Activity-Set Models is shown in Figure 4. On the top of the figure
is (one level of) the previous Actor/Activity-Role Model. A preliminary abstraction step
is shown in the middle of the figure where each actor position and each activity pattern is
regarded as a singleton actor or activity set respectively. Addition/removal associations
are now subsumed under participation associations. Based on this basic Actor-/ActivitySet Model, different more abstract models can be derived. We call these different perspectives on the initial model and we show some of them at the bottom of the figure.
For each perspective it is noted which set unions were applied. Here, we distinguish
between absolute associations between an actor set A and an activity set B (each sub-set
Asub ⊆ A has this association to one of the sub-sets Bsub ⊆ B) and contingent associa-
198
Figure 3: Examples of embedding situations based on AND/XOR configurations
Figure 4: Transition from Actor/Activity-Role Models to Actor-/Activity-Set Models
tions (for at least one but not for all sub-sets Bsub ⊆ B, there is an association to one of
the sub-sets Asub ⊆ A).
199
For Actor-/Activity-Set Models it also has to be taken into account that Collectivity Units
can be collective actors. For this purpose, it is no longer regarded which activity roles
are cross-level associated but just which activity sets on different levels have which ”interleaving” associations (for example, whether an interleaving is specified for each or
just for some activities of an activity set). Due to space limitations, we do not cover the
corresponding modeling primitives here.
3 Conclusion
In this paper, we present a modeling approach for actor/activity models that explicitly
addresses abstraction in two dimensions: (1) more or less abstract models for a given
scenario (model abstraction) and (2) more or less actor granularity (actor abstraction).
The results we present are part of our ongoing research. We yet have to provide tool
support for our modeling approach. Furthermore, up to now the models were mainly
used as means to conceptualize quite generic and abstract archetypes of organizational
units in multi-level organizational scenarios (cf. [WE10], in german). For the future, we
need to validate the practical use of the modeling approach by applying it to real-world
scenarios.
References
[AG09]
AOS-Group.
Jack Intelligent Agents Team Manual.
http://www.aosgrp.com/
documentation/jack/JACK Teams Manual WEB/index.html, 2009.
Available at:
[Dig09]
Virginia Dignum, editor. Handbook of Research on Multi-Agent Systems: Semantics
and Dynamics of Organizational Models. Information Science Reference, 2009.
[HMMF08] Christian Hahn, Cristian Madrigal-Mora, and Klaus Fischer. A Platform-Independent
Metamodel for Multiagent Systems. Autonomous Agents and Multi-Agent Systems,
18(2):239–266, 2008.
[Jen00]
Nicholas Jennings. On agent-based software engineering. Artificial Intelligence,
177(2):277–296, 2000.
[Mat08]
Florian Matthes. Softwarekartographie. Informatik-Spektrum, 31(6):527–536, 2008.
[Nor06]
Linda Northrop. Ultra-Large-Scale Systems: The Software Challenge of the Future.
Software Engineering Institute, Carnegie Mellon, 2006.
[Val04]
Rüdiger Valk. Object Petri Nets: Using the Nets-within-Nets Paradigm. In Jörg
Desel, Wolfgang Reisig, and Grzegorz Rozenberg, editors, Lectures on Concurrency
and Petri Nets: Advances in Petri Nets, volume 3098 of Lecture Notes in Computer
Science, pages 819–848. Springer, 2004.
[WE10]
Matthias Wester-Ebbinghaus. Von Multiagentensystemen zu Multiorganisationssystemen – Modellierung auf Basis von Petrinetzen. Dissertation, Universität Hamburg, Fachbereich Informatik. Elektronische Veröffentlichung im Bibliothekssystem der Universität Hamburg: http://www.sub.uni-hamburg.de/opus/
volltexte/2011/4974/, 2010.
200
A Generic Multi-purpose Conceptual Model Analysis
Approach Conceptual Specification and Application
Sebastian Herwig, Łukasz Lis, Matthias Steinhorst, Jörg Becker, Patrick Delfmann
University of Münster
European Research Center for Information Systems (ERCIS)
Leonardo-Campus 3
48149, Münster
Germany
{herwig | lis | steinhorst | becker | delfmann}@ercis.uni-muenster.de
Abstract: In this paper, we introduce a model analysis approach suitable for multiple analysis purposes. The approach is based on generic graph pattern matching
and semantic standardization of model elements. Due to its generic nature, it is applicable to conceptual models of any graph-based modelling technique. Through
semantic standardization, the approach relies on unambiguous model contents.
Therefore, its focus is on analyses with regards to the structure and contents of
conceptual models.
1 Introduction
The analysis of conceptual models addresses different goals, e. g., searching for corresponding model sections for model integration in distributed modelling projects or evaluating process compliance, amongst others. Manual model analysis can be costly, as the
number of models to be analysed may rank in the thousands [YDG10]. We therefore
argue that a semi-automated model analysis approach being suitable for different modelling languages and for different application scenarios is highly beneficial. Such an approach has to consider two basic aspects: First, it should be able to recognize structures
occurring within a model. Second, the approach should recognize the labels of the model
elements. It is crucial that the model elements contents are unambiguous, thus the approach should incorporate a mechanism that relies on semantic standardization.
In this contribution, we present a semi-automated model analysis approach, which makes
use of semantic standardization to assure comparability as a precondition for model
analysis and allows for flexible pattern specification and matching. Our research follows
the design science paradigm outlined by [He04]. This integrated approach is based on
our previous research on naming conventions [DHL09] and generic pattern matching
[De10]. The remainder of this contribution proceeds as follows: In Section 2, we present
a literature overview of existing model analysis approaches. Section 3 introduces the
conceptual specification of the approach. Section 4 provides an application example. We
close with an outlook to future research in Section 5.
201
2 Related Work
Different analysis approaches proposed in the literature can be divided according to their
primary goal into quality evaluation, exploration, and comparison approaches. Notable
work concerned with evaluating the syntactical as well as semantic quality of conceptual
models has, for example, been proposed by [Fr09], [LSM09], or [Go08]. In terms of
model exploration, various analysis approaches have been proposed by [Az09] [Es09],
or [TIR07]. To compare two or more models to one another, algorithms have been developed by [Di11], [YDG10], or [DDM08], amongst others. These approaches deal with
particular analysis task and are therefore only suitable in their respective domains. We
argue that it would be beneficial to have a generic analysis approach being suitable for
any modelling language and in any application domain. The following section introduces
such an approach. It is based on our previous work on enforcing naming conventions
[DHL09] and identifying arbitrary structures in conceptual models [De10].
3 Conceptual Specification
The overall procedure of our analysis approach is subdivided into two main steps as
depicted in Figure 1 (black-shaded elements are adapted from the pattern matching approach, grey-shaded elements are adapted from the semantic standardization approach,
and non-shaded elements are new). First, the analysis is defined (cf. section Analysis
Definition in Figure 1). Second, the analysis is applied to a set of models in order to
generate a report as a result (cf. section Report Generation in Figure 1). An Analysis is
composed of one or more Sub-analyses, each of them having a particular Search Criterion. The search criterion describes the properties of the expected analysis results (e.g.,
find all receipt structures containing the term invoice). Furthermore, it has to be defined whether the scope of an analysis is to explore a single set of models or to perform a
comparison of two model sets (cf. attribute scope). To assure that all sub-analyses within
an analysis can be applied to the same model base, an analysis can be composed only of
those sub-analyses which have the same scope.
Each sub-analysis is based on a Search Criterion specifying the fact to be searched for.
Any search criterion consists of either a single Atomic Search Criterion or a Composed
Search Criterion. Atomic search criteria are further specialized as:
! Pattern Equivalence Class (PEC) to search for different structural patterns regarded
as equivalent (e.g., two different data structures being recognized as similar)
! Structural Pattern to search for a specific model structure (e.g., a pattern representing
activities in process models that are related to an application system)
! Element Types (e.g., all application systems)
! Phrase Syntax to search for model elements whose labels follow a certain syntax
(e.g., model elements whose labels follow the syntax <verb, imperative> <noun, singular> in order to name process activities)
! Word Class to search for all words of a certain type within model elements labels
(e.g., all nouns to identify business process objects)
202
! Word to search for a particular word within the label of model elements (e.g., the
particular word invoice in order to identify all data models containing invoice data)
! Comparison Type to define a criterion for the comparison of two model sets. A
comparison type can take the values structural pattern, pattern equivalence class,
element type, word class, word, or phrase syntax. If an analysis has the scope
comparative, a result found in the one model set is compared to all results found in
the other model set. If both results are equivalent concerning the comparison type,
they are returned as a result pair marked as equivalent. For example, finding similar
structures in data models requires defining a pattern equivalence class containing
these structures as structural patterns. Then, a comparative analysis can use this class
as a search criterion. In addition, it uses the comparison type pattern equivalence
class to relate every pattern match in the one model to every match in the other
model, and possibly a further comparison type word to only return structure pairs
containing the same term (e.g., invoice).
Analysis Definition
Criterion
Restriction
(0,n)
Analysis
(1,n)
Analysis
Composition
(1,n)
(0,n)
Sub-analysis
(1,1)
SubanalysisCriterion
(0,n)
(0,n)
Scope
(0,1)
Search
Criterion
Atomic
Search
Criterion
D,T
D,T
Pattern
Equivalence
Class
(0,n)
Scope
Criterion
Structure
(1,n)
Composed
Search
Criterion
Structural
Pattern
Output
Type
Operator
Analysis
Result
Subanalysis
Result
(1,1)
(1,1)
Report
(1,n)
Report
Composition
(1,1)
Report
Element
Comparison
Match
(0,n)
(0,n)
Report
Element
Fact
(1,1)
(1,2)
Fact
(1,1)
Report
Target
(0,n)
Model Set
(0,n)
(1,n)
Model SetAssignment
(0,n)
Model
(0,n)
Model-FactAssignment
D,T
Element
Type
Structural
Pattern
Occurrence
Phrase
Syntax
Element
Occurrence
Word Class
Phrase
Syntax
Occurrence
Word
Word
Occurrence
Comparison
Type
Report Generation
Figure 1. Conceptual Specification of the Analysis Approach
As a comparison type defines only a criterion for the comparison of two model sets, this
criteria is only available for comparative sub-analysis (i.e., the attribute scope is set to
comparative). All the other atomic criteria can be used regardless of the specified
scope. A search criterion that contains more than one atomic search criterion is defined
as Composed Search Criterion. It is composed of further complex or atomic search criteria, which are connected through a logical operator. The search criteria to be combined
are specified by the Criterion Structure. To define the logical operator to be used for the
combination of these search criteria, we introduce the attribute Operator, which can hold
203
one of the values AND, OR, XOR or NOT. A Criterion Restriction is used to
refine an atomic search criterion. Here, a search criterion either an atomic or a complex
one serves as a constraint for an atomic search criterion restricting the resulting set of
model fragments. Therefore, the restricting search criterion is directly assigned to the
atomic search criterion to be restricted.
Finally, an analyst has to specify the granularity of the analysis results to be displayed.
Therefore, we introduce the attribute Output Type. For example, an analysis is defined to
search for structural pattern occurrences. By defining patterns as output type, the pattern
occurrences contained in the analysis result are included in the report. By setting the
value to element, the report is straightened to visualize occurrences of single model
elements, which are contained in the returned pattern occurrences. By defining phrase
syntax as output type only phrase syntax occurrences contained in the returned pattern
occurrences are visualized, and so on. To avoid empty reports, the output type has to be
defined in respect to specified search criteria.
The results of an analysis are visualized as a Report. Depending on the scope of an analysis, a report is targeted at either one Model Set for an explorative analysis or two model
sets for a comparative analysis. Each model set consists of one or more Models to be
analysed. A report is composed of one or more Report Elements, each of them resulting
from one sub-analysis. A report element represents a single row in a report and shows
the particular Facts returned as search results from the corresponding sub-analysis. A
fact defines a particular match of the sub-analysis search criterion, which is shown in
the report. In respect to the possible output types of a sub-analysis, a fact can be a Pattern Occurrence, an Element Occurrence, a Phrase Syntax Occurrence, or a Word Occurrence. For one particular sub-analysis of the aforementioned example, all process
activities that are related with a specific application system are included in the report
element as particular element occurrences. Facts resulting from a comparative analysis
are linked to each other as Comparison Matches. This way, equivalent result pairs are
defined by relating a fact occurrence in the one model set to the matching fact occurrence in the other one.
4 Application
To demonstrate the feasibility of our model analysis approach, we developed a prototypical implementation combining both structural pattern search and linguistic standardization features. As an implementation basis, we used a meta-modelling tool which was
available from a previous research project. Since our model analysis approach is generic
in nature, it is applicable to a variety of business scenarios. To evaluate its general feasibility, however, we decided to provide an application example in an area where we believe the approach is most beneficial, namely in case of mergers and acquisitions. A
major challenge in merging two or more companies is integrating the respective IT landscapes [MB07]. A first step toward IT integration in merger scenarios is identifying the
different application systems that support a particular business activity in all involved
companies.
204
To identify such structures, we define a simple analysis with the output type pattern. The
report will show us all model structures that match the predefined search criteria. For the
example of application systems supporting specific business activities we therefore defined a structural pattern called Function->Application System for EPCs. This pattern
returns all functions that are directly related to an application system. As far as the linguistic features are concerned, we define the words customer and CRM to comprise
the necessary vocabulary for this domain. Given this pattern and these terms we can
construct the search criterion that allows for searching occurrences of the Function>Application System pattern. The search is further restricted to include only those pattern occurrences containing either the word customer or CRM at least once.
Figure 2. Exemplary Analysis
As a model basis, assume that we have two fictional companies. For each of these companies we modelled the three business processes campaign execution, order processing, and request processing as EPCs. Running the analysis as described above
results in the report depicted in Figure 2. In this case, the business process campaign
execution contains two occurrences of the specified search criterion. The processes
order processing and request processing each contain three occurrences. Clicking on
one model allows for navigating to it. Each fact is then highlighted, so the analyst can
easily locate the different fact occurrences. In the example, all functions that are directly
connected to an application system and contain the terms customer or CRM are highlighted.
5 Outlook
The contribution of this paper is two-fold: on the one hand, it provides a convenient tool
to practitioners and professionals that can be used in a multitude of application scenarios,
regardless of what modelling languages they prefer. On the other hand, we provide an
innovative approach to the model analysis body of knowledge. It takes into account both
the intensively discussed requirements of semantic interoperability and structural pattern
matching. It is applicable, reusable and thus evaluable in multiple scientific, educational
and professional scenarios. In future research, we aim at developing an automated or at
least semi-automated approach to integrate a number of conceptual models to one consolidated model. Furthermore, we intend to develop meaningful metrics for the analysed
models.
205
References
[Az09]
Azevedo, L. G.; Santoro, F.; Baião, F.; Souza, J.; Revoredo, K.; Pereira, V.; Herlain, I.:
A Method for Service Identification from Business Process Models in a SOA Approach. In (van der Aalst, W.M.P.; Mylopoulos, J.; Sadeh, N. M.; Shaw, M. J.; Szyperski, C.; Halpin, T. et al. Eds.): Enterprise, Business-Process and Information Systems
Modeling, LNBIP, vol. 29. Springer, Heidelberg, 2009; pp. 99-112.
[DDM08] van Dongen, B. F.; Dijkman, R.; Mendling, J.: Measuring similarity between business
process models. In (Bellahsene, Z.; Léonard, M. Eds.): Proceedings of the 20th International Conference on Advanced Information Systems Engineering, Montpellier,
2008, pp. 450-464.
[De10]
Delfmann, P.; Herwig, S.; Lis, L.; Stein, A.; Tent, K.; Becker, J.: Pattern Specification
and Matching in Conceptual Models. A Generic Approach Based on Set Operations.
In: Enterprise Modelling and Information Systems Architecture 5 (2010) 3, pp. 24-43.
[Di11]
Dijkman, R.; Dumas, M.; van Dongen, B.; Käärik, R.; Mendling, J.: Similarity of
business process models: Metrics and evaluation. In: Information Systems 36 (2011) 2;
pp. 498-516.
[DHL09] Delfmann, P.; Herwig, S.; Lis, L.: Unified Enterprise Knowledge Representation with
Conceptual Models Capturing Corporate Language in Naming Conventions. In: Proceeding of the 30th International Conference on Information Systems (ICIS 2009),
Phoenix, 2009, pp. 13-26.
[Es09]
Esswein, W.; Weller, J.; Stark, J.; Juhrisch, M.: Identifikation von Services aus
Geschäftsprozessmodellen durch automatisierte Modellanalyse. In (Hansen, H. R.; Karagiannis, D.; Fill, H.-G. Eds.): Proceedings der 9. Internationalen Tagung Wirtschaftsinformatik, Vienna, 2009; pp. 513-522.
[Fr09]
Friedrich, F.: Measuring semantic label quality using WordNet. In: Proceedings des 8.
GI-Workshops Geschäftsprozessmanagement mit Ereignisgesteuerten Prozessketten
(EPK). CEUR-WS.org, 2009; pp. 7-21.
[Go08]
Governatori, G.; Hoffmann, J.; Sadiq, S.; Weber, I.: Detecting Regulatory Compliance
for Business Process Models through Semantic Annotations. In (van der Aalst,
W.M.P.; Mylopoulos, J.; Sadeh, N. M. et al. Eds.): Business Process Management
Workshops. Springer, Heidelberg, 2008; pp. 5-17.
[He04]
Hevner, A. R. et. al.: Design Science in Information Systems Research. MIS Quarterly
28, 2004; pp. 75-105.
[LSM09] Leopold, H.; Smirnov, S.; Mendling, J.: On labeling quality in business process models. In: Proceedings des 8. GI-Workshop Geschäftsprozessmanagement mit Ereignisgesteuerten Prozessketten (EPK). CEUR-WS.org, 2009; pp. 42-57.
[MB07]
Mirklitz, T.; Buxmann, P.: IT Standardization and Integration in Mergers and Acquisitions. A Decision Model for the Selection of Application Systems. In: Proceedings of
the 15th European Conference on Information Systems (ECIS 2007), St. Gallen, 2007,
pp. 1041-1051.
[TIR07] Thom, L. H.; Iochpe, C.; Reichert, M.: Workflow patterns for business process modeling. In: Proceedings of Workshops and Doctoral Consortium of the 19th International
Conference on Advanced Information Systems Engineering (CAiSE 2007), Trondheim, 2007; pp. 349-358.
[YDG10] Yan, Z.; Dijkman, R.; Grefen, P.: Fast business process similarity search with featurebased similarity estimation. In: Proc. of the 2010 international conference on the move
to meaningful internet systems, Hersonissos, 2010; pp. 60-77.
206
System Architecture Validation with UML
André Pflüger, Wolfgang Golubski
Westsächsische Hochschule Zwickau
Fachgruppe Informatik
Dr.-Friedrichs-Ring 2A
08056 Zwickau
{Andre.Pflueger|Golubski}@fh-zwickau.de
Stefan Queins
SOPHIST GmbH
Vordere Cramergasse 13
90478 Nürnberg
Stefan.Queins@sophist.de
Abstract: The foundation of every system is the architecture. It has to be designed
according to the corresponding system requirements by the system architect who is
also responsible for ensuring that it fits to the system requirements even if these
change due to new conditions during development process. Our approach defines a
model driven process for the architect to validate system architecture against system requirements based on UML. It supports the architect in designing the architecture and in analysing the impacts of requirements changes.
1. Motivation
Designing and validating system architectures are time- and work-consuming tasks. The
responsible person for those tasks, the system architect, must have special expertise in
this area to create the basic foundation of the system, experience and social skills are
also appreciated. The number of requirements is growing with the systems complexity
which also increases the number of dependencies between them. The system architect
does not only have to look at all these requirements, he also has to handle requirements
changes and their impacts on the system architecture. Due to customer’s desire for high
flexibility, requirement changes are no longer limited to analysis phase and early design
phase. They occur up to the end of implementation phase making architect’s work a big
challenge. A common proceeding can be described as follows: The architect creates an
architecture draft and improves it iteratively. The validation is done manually by stepping through the architecture-relevant requirements. There is no defined validation process which might cause unintentional omitting of important aspects. The validation is
based on estimations due to the architect’s experiences and only in special cases, mostly
for critical requirements, it is based on verifiable results. According to the required work
effort documentation and validation are only done completely for important, e.g. review
appointments. We think that the process described is not sufficient for developing complex systems because there is no defined validation process to achieve repeatable results.
Figure 1 shows the four levels of our approach to support the architect in his work. The
top-level defines our model-based validation process which is independent from any
specific modelling language. According to apply this process we choose the Unified
modelling Language (UML) and develop UML-specific support tool approaches in levels two and three. The implementation of these tool approaches are part of the fourth
level. The validation process is used in conjunction with the iterative design of the system architecture which can be partly automated. So-called validation targets sum up
architecture-relevant requirements on an abstract level suitable for architecture design.
207
Figure 1: different levels of the approach
This level gives the architect a better view on the system architecture according to the
architecture-relevant requirements and makes complex correlations visible. For every
validation target the architect creates a process to check if the associated requirements
are fulfilled by the system architecture. The architecture is valid if all validation targets
are valid. Requirements can be associated to more than one validation target supporting
the architect during impact analysis after requirement changes. This approach depends
on a model-based system documentation used as data source for the validation target
processes and for a seamless documentation of system architecture design. For every
validation target the architect may create a specific notation allowing him to model all
necessary information for the validation target process. Based on the choice of UML
information not available is added by using the UML profile mechanism [UML10].
In section 2 we will discuss related work to our approach. Section 3 describes the validation process in detail and section 4 deals with the validation target specific notations
using the UML profile mechanism and the separating of development and validation
model. In section 5 we apply the introduced approach to an example from the area of
embedded systems. Section 6 lists up benefits and our experiences before section 7 concludes the paper and overviews future work.
2. Related Work
According to our research there are no approaches defining a model driven process for
the architect to validate system architectures against requirements which includes the
support for the architect for designing the architecture and analysing the impacts of requirement changes. The project System Architecture Virtual Integration (SAVI) [FEI10]
initiated by the Aerospace Vehicle Systems Institutes is based on three key concepts: a
reference model, a model repository with model bus and a model based development
process. Models using i.e. UML, AADL or SysML and mapped to the reference model
can be compared with the content of a requirements repository. In theory a validation of
system architectures against requirements would be possible but the overhead would be
high because this is not the projects intention. Furthermore, the project only exists in the
description of a feasibility study. There are modelling-language-specifications available
for special topics like the System Modelling Language (SysML) [OMG10a], the Architecture Analysis and Design Language (AADL) [CMU11] or the Modelling and Analysis of Real-Time and Embedded Systems (MARTE) profile [OMG10b] which could be
used but they are limited to a specific domain and they cannot be extended to add validation specific data.
208
The tool PREEvision of aquintos [Aqu11] supports a model based development and
evaluation of electric and electronic architectures. The modelling language is based on a
graphical notation using an own domain specific meta-model, architecture relevant information are provided by a data model. Our approach is focused on the whole system
architecture including software components and uses an extension of the domain specific
notation by using the profile mechanism of the UML.
3. Validation process
The first step towards a model driven validation of system architectures is defining a
process for the architect. This process should be independent from any special system or
modeling language. It describes the necessary steps initially building up the validation
and how to react on requirement changes. First the validation targets are detected from
the requirements, especially non-functional requirements. The architect creates a validation process for every validation target and a validation target specific notation based
upon the domain notation to add validation specific information to the validation model.
By running all validation processes, manual or automatic, the architecture is validated. In
failure case the architecture has to be adjusted and revalidated. These modifications are
transformed into development model in case of a successful validation. The validation
process restarts after requirements change. [Pfl10] describes this process in detail.
4. Modeling language UML
According to the four levels presented in figure 1 our approach does not depend on any
particular modeling language but for practical usage a choice is necessary. We decided
for the UML because it is a widespread and accepted standard in research and industry.
It provides two possibilities to extend its meta-model: meta-model extension and the
profile mechanism [UML10]. By creating and using a profile composed of stereotypes,
tagged values and constraints for every validation target the architect can add the relevant information to UML elements without much effort. An UML element can own
several stereotypes from different profiles, i.e. from different validation targets, thus
there is no need for modeling dependencies between different validation targets just
because they add information to the same UML element easing profile reuse. The UML
like other meta-model based languages supports the exchange of data between different
models. For this purpose the specification defines the XML Metadata Interchange (XMI)
format. We use the Eclipse Modeling Framework (EMF) [EMF10] for implementing
tools supporting the system architect in creating the validation model by transferring
selected UML elements from development model to validation model and in synchronizing the two models after system architecture modifications. More details about this topic
can be found in [Pfl10].
5. Radar system example
A radar system receives echoes of electro-magnetic signals from an antenna, processes
these data closed to real-time and provides tracks which are the desired objects the radar
engineer wants to see on his radar display. The data processing in a radar system is almost sequential. It is just separated by data processing variants due to different types of
signals which are in our example: sea targets, air targets and near range targets.
209
These signals are characterized by different electro-magnetic wave forms and processing
algorithms allowing the radar engineer to detect objects in a short or far away distance
with different resolutions. The system processes a continuous stream of data so that the
computing performance for data of a specific point in time is influenced by processing of
past and future signals. For this example we consider the following requirement:
- The system shall provide tracks for fed in test data after 320ms. (A1)
- The system shall provide plots for fed in test data after 280ms. (A2)
- The system shall be able to compute 3500 MBit/s data received from antenna. (A3)
- The system shall consume less than 220Watt/h. (A4)
- The temperature of each FPGA shall be less than 50°C. (A5)
- The temperature of each microprocessor shall be less than 70°C. (A6)
The system architecture describes software and hardware components, their connections
and behaviour. For this example we focus on the components and their connections.
Each software component has to have a dependency connection with stereotype
executedOn to a hardware component indicating that it is executed on it. There are three
hardware boards with two processing units, one microprocessor and one field programmable gate array (FPGA), each.
According to the validation process the architect detects validation targets from system
requirements in the first step. A1 and A2 are assigned to validation-specific aspect processing time (VT1), A3 to communication infrastructure (VT2), A4 to energy consumption (VT3) and A5 and A6 to system temperature (VT4). For the validation target specific notations the stereotype software has the tagged value mflops specifying needed
floating point operations for the software component and the stereotype hardware has
the tagged value performance providing the performance of processing units in MFlop/s.
If we assume linear energy consumption and linear temperature rising for processing
units in combination with a basic value, these data can be calculated by using processing
unit load deduced by VT1’s examination process. We add the tagged values energyConsumption and basicEnergyConsumption to the stereotype hardware as validation target
specific notation for VT3 and temperature and basicTemperature for VT4 (see figure 3).
We add the stereotype communication with tagged value bandWidth to the association
representing physical link between the board and the hardware component executing the
first software component in the processing chain. After determine the validation targets,
their corresponding examination processes and notations the architect creates the validation model illustrated in figure 2.
Figure 2: Validation model for radar system validation
210
Figure 3: Validation target specific notations for VT1, VT2, VT3 and VT4
Table 1: result examination process VT1
Data
Processing time [ms]
tracks
313
plots
277
Table 2: result examination process VT2
Communication device
Antenna radar system
Band width
[MBit/s]
5200
Table 3: result examination process VT3 and VT4
Processing unit
B1_FPGA
Processor
load [%]
Temperature
[°C]
26
48
35
B1_PowerPC
13
63
42
B2_FPGA
24
47
34
B2_PowerPC
8
60
37
B3_FPGA
14
43
28
B3_PowerPC
12
63
Total energy consumption
Table 4: result examination process VT1
Data
Processing time [ms]
tracks
306
plots
270
Table 5: result examination process VT2
Communication device
Antenna radar system
Band width
[MBit/s]
5200
Energy
[Watt/h]
41
218
Table 6: result examination process VT3 and VT4
Processing unit
B1_FPGA
Processor
load [%]
Temperature
[°C]
Energy
[Watt/h]
25
47
34
B1_PowerPC
12
63
40
B2_FPGA
15
44
29
B2_PowerPC
28
74
55
B3_FPGA
13
43
27
B3_PowerPC
11
62
Total energy consumption
39
221
It contains validation specific information only and it is used as data source for automated validation. Some information of the development model has been transferred,
some is not required and therefore has not been transferred and other has been added by
the introduced profiles. The results of the system architecture validation, i.e. of the four
validation targets, are shown in table 1 to 3.The system architecture passes the validation
and therefore fulfils the regarded system requirements. There has been no architecture
modification in the validation model so that synchronization of the models is not necessary. During development requirements are changing. In our example the customer
changes requirement A1: Tracks shall be provided after 310ms instead of 320ms.
211
The revalidation after requirement change fails because VT1 cannot be validated: 313ms
is greater than 310ms. The architect has to change the system architecture. He assigns
software component sweep processing to the PowerPC of Board_2. This modification is
memorized because it has to be transferred to the development model in case of a successful validation. Although VT1 is valid after modification, the whole validation fails
due to energy consumption (see table 4 to 6). By analyzing validation results including
former ones the architect is able to identify the problem. The processing unit load of the
PowerPC on Board_2 has been raised causing raising energy consumption which cannot
be compensated by falling energy consumption of FPGA on Board_2. The architect
restores the former architecture, exchanges the software components burst processing
and near range target processing and reruns architecture validation. These modifications
are transferred to development model because validations for all validation targets are
successful.
6. Conclusion and future work
System development has to deal with evolving conditions challenging the system architect who has to design and document the architecture just as validate it against system
requirements considering influencing parameter. The introduced approach defines a
model-based validation process integrated into iterative development of system architectures. The process can be partly automated reducing time and effort for validation. The
approach supports the architect in keeping track of architecture specific aspects and their
dependencies such as analysing impacts after requirement changes. The approach has
been developed with the help of projects of the SOPHIST GmbH and our experiences
from system developments. Our experience shows that extra work caused by this validation process amortizes within several iterations of architecture design and requirement
changes. Currently we are developing tools to support the architect in model-to-model
transformation and using non-UML data sources during transfer process. In combination
with the validation target management tool we try to provide a tool chain for the whole
validation process increasing the usability. We also develop a validation target process
simulation to gain experience with heavily dependent validation targets.
References
[Aqu11] PREEvision by aquintos, http://www.aquintos.com, last access on 2011-02-04
[CMU11] Architecture Analysis & Design Language (AADL), Carnegie Mellon University,
http://www.aadl.info, last access on 2011-01-03
[EMF10] Eclipse
Modeling
Framework
(EMF),
Eclipse
Foundation,
http://www.eclipse.org/emf/, last access on 2010-11-28
[Fei10]
P. Feiler, L. Wrage, J. Hansson, System Architecture Virtual Integration: A Case
Study, Embedded Real-time Software and Systems Conference, May 2010
[OMG10a] System
Modeling
Language
(SysML),
Object
Management
Group,
http://www.omgsysml.org/, last access on 2010-11-28
[OMG10b] UML-Profil MARTE, http://www.omgwiki.org/marte/, Object Management Group,
last access on 2010-11-28
[Pfl10]
Pflüger A., Golubski W., Queins S.: Validation Of System Architectures Against
Requirements, International Joint Conferences on Computer, Information and System
Sciences, Engineering (CISSE’2010), published by Springer in Summer 2011
[UML10] Unified Modeling Language (UML) Version 2.3, http://www.uml.org, 2010
212
Tool Support for the Comprehensive Modeling of
Quality Information within
Business Process Models
Robert Heinrich, Alexander Kappe, Barbara Paech
Institute of Computer Science, University of Heidelberg
Im Neuenheimer Feld 326, 69120 Heidelberg, Germany
{heinrich, paech}@informatik.uni-heidelberg.de
kappe@stud.uni-heidelberg.de
Abstract: Business process modeling is commonly used to document information
about structure and behavior of a business process. However, current business
process modeling notations do not support well the expression of quality
information relating to business processes. Organizations are interested in the
capturing of quality information for quality improvement of business processes and
supporting IT systems. We are developing an approach to capture quality
information comprehensively within business process models. In contrast to
existing approaches, our notation allows the capturing of a broad range of quality
characteristics as well as detailed attributes and measures. The approach is
implemented as an extension of a CASE tool. Moreover, we discuss lessons
learned from the application of the approach to a business process from practice.
1 Introduction
Business process modeling is widely used within organizations as a method to increase
awareness and knowledge of business processes, and to deconstruct organizational
complexity [BGR05]. Current business process modeling notations do not aim at
expressing quality information (QI) such as information about maturity or security of a
business process (cf. [Ko08], [PZ08], [SZS10]). Hence, quality requirements are often
not considered at the process modeling stage, which results in increased costs and delays
in the further development of business processes and involved IT systems. Annotating
the process model with QI contributes to a model that provides a more complete
representation of the overall business process [PZ08]. A (graphical) expression of QI
together with information on structure and behavior within a single model would
increase the modeler’s focus on quality at the process modeling stage. Therefore, as
stated in related work (cf. [SZS10], [PZ08]), it facilitates the capturing of quality
requirements and results in a more complete set of requirements. Although the benefit of
QI captured in a process model for early requirements elicitation has already been
identified by other authors, current approaches only focus on single QI [HKP11]. Our
research aims at capturing a comprehensive set of business and IT quality requirements
213
and the coordination of these requirements. A first step in this direction is the
comprehensive modeling of QI within business process models, as business process
models are a starting point of requirements elicitation [Ko08].
In contrast to software product quality, which for example is standardized in the
ISO/IEC 9126 quality model [ISO01], there is no common quality standard for business
processes. Therefore, we developed a comprehensive quality model for business
processes that is based on software product quality standards [HP10a], [HP10b].
In this paper, we describe an approach to present the QI of our quality model within a
business process model and provide prototypical tool support of our approach. The paper
is structured as follows: In Section 2, as a background, we sketch our quality model.
Section 3 describes our approach to model QI and the prototypical tool support. In
Section 4 we present lessons learned from an exemplary application of the approach and
tool. Section 5 concludes the paper and presents future work.
2 Background
In the quality initiative domain process quality is in the focus of research and practice for
some decades and there are many high-level and expert-based techniques like TQM,
Kaizen or Six Sigma. However, a comprehensive view on the – in particular nonfinancial – quality aspects of a business processes is still missing.
Therefore, we developed the comprehensive Business Process Quality Reference-Model
(BPQRM) [HP10a], [HP10b] using characteristics we transferred from software product
quality standards. To the characteristics we allocated a broad range of detailed quality
aspects from business process management literature. We use a hierarchical structure of
QI defined as follows. A business process quality characteristic is a category of business
process quality attributes, for example the maturity of an activity. A business process
quality attribute is an inherent property of a business process that can be distinguished
quantitatively or qualitatively, for example the error density of an activity. A business
process quality measure is a variable to which a value is assigned as the result of
measurement. Measures can be defined as base measures or derived measures. A base
measure is a measurement for which the value is directly applicable to the process, e.g.
the number of errors or the number of (sub) activities. A derived measure is a measure
that is defined as a function of two or more values of base measures, e.g. the number of
errors per activity size. In the following, we use the term QI as a superset of
characteristics, attributes and measures.
Business process quality refers to the components of a business process. Components are
the activities of the process, the actors performing these activities, the objects handled
and created by the process as well as the resources necessary for its execution. As an
activity can be subdivided into sub activities, we consider a process itself as an activity.
In the BPQRM we associated a set of quality characteristics adapted from software
product quality standards to each component of a business process. Figure 1 shows the
BPQRM. The nodes correspond to the components and the characteristics are listed
214
either within a node or on an edge between nodes. See [HP10b] for further information.
Fig. 1: Business Process Quality Reference-Model
3 Approach and Tool Support
We conducted an extensive literature and tool survey [HKP11] on research approaches
and tools from practice for modeling QI within business process models. As shown in
the survey, neither of them was able to express QI covered by more than 9 characteristics
of the BPQRM. Moreover, in most cases the surveyed tools were not able to express QI
in the process modeling view. The deficiencies identified in the survey motivated us to
develop an approach to model QI comprehensively within business process models and
to provide a new tool support. The research question thereby is how to enable the
modeling of a large set of different QI without a major increase of complexity of the
modeling notation. We decided to add small graphical symbols to already existing model
elements. To each characteristic in the BPQRM we associate a graphical symbol. A
detailed allocation of the symbols to the characteristics can be found in [Ka11]. To each
symbol we associate attributes, derived measures, base measures and the related values
in tabular form. The approach is implemented prototypically as an extension of the
Eclipse-based CASE tool UNICASE (http://unicase.org/).
215
Figure 2 shows a screenshot of our tool. The process model is presented in a split screen
view together with the corresponding tables of attributes and measures. Therefore, – in
contrast to the surveyed tools – information on structure and behavior as well as
information on quality can be presented in a single view. In the process modeling view
(Business Process Model) the modeler can capture information on structure and behavior
of the business process as a BPMN [OMG11] model and can additionally capture the
quality characteristics of the BPQRM. The modeler can add quality characteristics to a
process model element by dragging and dropping the characteristic icons from the
toolbar beside the process model to the corresponding process model element. By
clicking a characteristic icon in the process model, the table of attributes and measures
related to the selected characteristic appears in the QI Details view below. Here the
modeler can capture the attributes, measures and their values. Thus, it is possible to
model a large set of different QI (in the form of characteristics) as well as to capture
detailed QI (in the form of attributes and measures) without a major increase of
complexity of the process model.
Fig. 2: Screenshot of the Process Modeling Editor
Note that in our view it is important to model QI simultaneously with information on
structure and behavior because QI is often elicited together with information on structure
and behavior. We think it is not sufficient to enter QI ex post into the model as QI may
influence the structure and behavior of a business process. Our approach is applicable to
an arbitrary graphical modeling notation. We utilize the BPMN as an example as it is an
216
up-to-date and wide-spread modeling notation for business processes. Moreover, the
BPMN is well suited to be extended.
4 Exemplary Application
In the previous section, we described our approach and the prototypical tool support. In
the following, we discuss lessons learned by the application of the tool in a first
evaluation. We wanted to get first feedback whether our approach is an adequate means
to model the QI of a business process from practice. As an example we choose the
process of writing discharge letters in a hospital, because all the components of the
BPQRM are contained in the process and there is a large number of QI to be captured. A
discharge letter is a summary of the performed patient treatment and is used for the
communication between physicians for follow-up treatments. The modeled process
consists of 15 activities and handles 5 information objects. 4 actors and 1 IT system are
involved in the process. Altogether, we captured 12 different characteristics of the
BPQRM and related attributes and measures. For activities we captured the
characteristics maturity, time behavior, interoperability, attractiveness and resource
utilization. For resources we captured the characteristics maturity, attractiveness and
learnability and for information objects we captured currentness, compliance,
availability and operability. We did not capture QI related to actors as this QI was
excluded on request of the hospital. In Figure 2, as an example, a part of the process is
modeled in the process modeling editor. The diagram shows 4 activities and 2 data
objects annotated with characteristic icons which are relevant in the process. Further
information on the exemplary application can be found in [Ka11]. Next, we discuss
lessons learned by applying our approach in practice.
The example showed that our approach is a valid and practically applicable means to
model the 12 different characteristics that are relevant in the process. In general,
annotating a large set of characteristics to a single model element (the worst case is 26
characteristics per model element for the process components activity and resource) very
likely will reduce the clarity of a model developed using our approach. It turned out,
however, that in the example the number of modeled characteristics per model element
was much lower (typically 1 to 5 characteristics per model element) because not all of
the characteristics were relevant to every model element in the specific process.
In the process model we only visualize the quality characteristics. One characteristic
icon can represent several attributes and measures. The limitation to characteristics is a
useful means to allow a compact overview of the QI. However, the modeler cannot
access a specific attribute or measure directly. Instead s/he has to click on the
corresponding characteristic icon. For example, if the modeler wants to view the value of
the measure number of errors, s/he first has to click on the maturity icon (see Figure 2).
In the future the tool support may provide functionality for the direct retrieval of a
specific attribute or measure, e.g. by searching for its name.
The split screen view is a useful means to show details of a single characteristic together
with the process model element the characteristic is annotated to. Moreover, the tool
217
allows the switching between characteristics quickly. However, selecting single
characteristic icons may be cumbersome, if the modeler wants to view QI aggregated
over different characteristics or model elements. Future work on the tool is required.
5 Conclusion and Future Work
In this paper, we presented an approach to model QI comprehensively within a business
process model and provide prototypical tool support as an extension of the Eclipse-based
CASE tool UNICASE. The approach puts quality in the focus of business process
modelers and therefore helps to capture relevant QI early at the modeling stage. It is built
on the results of an extensive literature and tool survey. To the best of our knowledge
this approach is the first one which allows the modeling of a comprehensive set of QI
within a business process model and provides guidance on the QI to be modeled.
As a next step we plan to conduct further evaluations of the approach. We want to
compare our approach to goal-oriented approaches that model quality requirements
separately in a goal hierarchy and link them to processes. Moreover, the usability of the
prototypical tool support has to be revised. Extending the tool support by additional
functionality for automatic QI calculation, analysis or simulation is also desirable.
References
[BGR05]
Bandara, W., Gable, G.G., Rosemann, M.: Factors and measures of business process
modelling: Model building through a multiple case study, European Journal of
Information Systems Vol. 14 No. 4, pp. 347-360 (2005)
[HP10a] Heinrich, R., Paech, B.: Defining the Quality of Business Processes, In: Engels, G.,
Karagiannis, D., Mayr, H.C., eds.: Modellierung 2010, Lecture Notes in Informatics
Vol. P-161, GI, pp. 133-148 (2010)
[HP10b] Heinrich, R., Paech, B.: Business Process Quality - A Technical Report, Technical
Report, Software Engineering Heidelberg (2010)
[HKP11] Heinrich, R., Kappe, A., Paech, B.: Modeling Quality Information within Business
Process Models, In: Proceedings of the 4th SQMB Workshop, TUM-I1104, pp. 4-13
(2011)
[ISO01] ISO/IEC 9126-1: Software engineering — Product quality — Part 1: Quality model,
First edition (2001)
[Ka11]
Kappe, A.: Entwicklung und Umsetzung eines Konzepts zur Modellierung von
Qualitätsinformationen in einem Geschäftsprozessmodell, Master Thesis, Software
Engineering Heidelberg (2011) (in German)
[Ko08]
Korherr, B.: Business Process Modelling: Languages, Goals, and Variabilities, VDM
Verlag, Saarbrücken, Germany (2008)
[OMG11] OMG: Business Process Model and Notation (BPMN), Version 2.0, Object
Management Group (2011)
[PZ08]
Pavlovski, C.J., Zou, J.: Non-functional requirements in business process modeling, In:
Proceedings of the 5th Asia-Pacific conference on Approachual Modelling, (APCCM
2008), CRPIT, 79. Hinze, A., Kirchberg, M. (eds.) ACS, pp. 103-112 (2008)
[SZS10] Saeedi, K., Zhao, L., Sampaio, P.R.F.: Extending BPMN for Supporting CustomerFacing Service Quality Requirements, In: Web Services, IEEE International
Conference on, IEEE Computer Society, pp. 616-623 (2010)
218
Understanding IT-Management and IT-Consulting
Teaching as Product-Service System:
Application of an Engineering Model
Matthias Boehm, Carl Stolze, Oliver Thomas
University of Osnabrück
Information Management and Information Systems
Katharinenstraße 3
49074 Osnabrück
{ matthias.boehm | carl.stolze | oliver.thomas }@uni-osnabrueck.de
Abstract: In this research-in-progress paper we conceptualize the teaching of
IT-Management and IT-Consulting as a hybrid package of products (resources)
and services (teaching). This understanding offers a new perspective on teaching
approaches and creates new opportunities for all stakeholders. Following a wellestablished procedure model for product-service systems (PSS) engineering, we
derive the customer requirements to the hybrid package. Than a product model is
developed. Based on these findings, an Education Integration Platform Solution
(EIPS) is prototypically designed. Finally, a conclusion and outlook are given.
1
Introduction The Challenge of Teaching ITMC
The current business and information technology (IT) environment can be characterized
by dynamically evolving concepts and technologies as well as unprecedented volatility
[OB11]. This fact poses challenges on the future working force, which can only be faced
by permanent training of skills and knowledge [La10]. A combination of several crossdisciplinary qualifications is necessary for managing the complexity of the environment
[Wi02]. However, not only the professionals have to cope with new technology and
structures, also educational systems face challenges, like for example technological and
organizational change, globalization and international commerce as well as environmental challenges [OB11]. It has been shown for example that sustainability concerns have
an impact on teaching approaches [SBZ11]. Furthermore, what people learn, how they
learn, and where they learn will radically change in future [Wa07].
IT management and IT consulting (ITMC) are two important fields, which both are very
personal-intensive [CM09]. IT management covers different areas like managing IT
resources (and personnel) as a competitive advantage or the development and operation
of information systems [Lu04, CM09]. In this field, a variety of IT professionals, like
programmers, analysts, IS managers and others, are employed. In the following, we
focus on analysts and IS managers. IT consultants focus on building, managing and
219
operating information systems [VSB12]. They work as intermediaries between IT and
business function [BD95].
Teaching ITMC involves several substitutable products and services [BSB11]. One cannot solely focus on the actual act of teaching (service) [MT03]. The learning resources,
in terms of presentations, working sheets, exercises and other documentation [TR05] are
the physical part of the teaching product-service system (PSS). The special thing about
understanding teaching ITMC as a PSS is the fact that the physical resources play a more
important role than in other teaching cases. Participants often cannot attend every presence lecture due to frequent travelling in their jobs (this is especially true for IT consultants) and therefore a good documentation is required. Mobile and individual e-learning
opportunities can support their learning. Furthermore, the ITMC field is wide and the
content, which has to be taught, is complex and difficult to understand. Therefore, personal teaching and coaching components are highly relevant. Hence, the provision of this
hybrid PSS has to be made transparent and measureable to all involved stakeholders.
That is the reason why our central research question can be formulated as follows: How
can PSS engineering methods be used to develop an integrated solution that allows and
effective and efficient teaching of ITMC?
2
Towards an Integrated Solution
2.1 Method
As we develop a holistic product-service system that incorporates the necessary aspects
of teaching ITMC as well as satisfying the requirements of all stakeholders, we focus
especially on the integrated design of products and services [TWL08]. Therefore, we use
a method for PSS engineering, developed by [WSB04] that has been successfully applied in research [BFS11].
Following [BFS11], the procedure consists of three phases: (1) definition, (2) synthesis,
and (3) analysis. First, the requirements of the PSS are derived and based on the to-be
properties have to be defined. By separating the inquiry of the requirements from the
definition of properties of the product-service system, the transformation of them into
the properties of the PSS can be made transparent [TWL08]. In the second phase, new or
changed characteristics and components of the hybrid service bundle with its inner relationships and dependencies are derived. Finally, the as-is properties of the service bundle
are evaluated by comparison with the to-be properties. Here one has to clarify which
requirements already have been fulfilled and which still have to be covered.
2.2 Requirements
In order to determine the requirements to the PSS, we conducted qualitative expert interviews with six practitioners from international consulting companies and representatives
of five universities during a workshop in February 2010 [BSB11]. As the findings of
these interviews provide only a rough overview on the requirements, a deeper analysis is
220
necessary. Therefore, we conducted a literature review [WW02]. Search terms like
teaching, training, information technology, IT management, IT consulting, Learning
Management System or education system were used to search for relevant papers in
leading journals (top 20 journals according to AIS) and conferences (ICIS, ECIS, WI,
AMCIS, HICCS) within the last 20 years. These findings have been integrated into those
from the expert interviews. The result is depicted in Table 1.
Social and personcentric requirements
! Diffusion and transfer of knowledge between the stakeholders
! Giving opportunities for developing interpersonal skills
! Acquisition of needed technical skills and competencies
Economic
requirements
! Providing certificates after completion of program (value for money)
! Monetary or time savings and/or noticeable quality improvements
! Ensuring efficiency of the methods by measuring economic and ecological
values
Ecological
requirements
! Minimizing the ecological impact of the program through state-of-the-art
learning methods (e-learning, blended learning, simulation)
! Creating awareness among the stakeholders for ecological issues
Execution-related
requirements
! Integrated platform for all phases from enrollment over execution till evaluation and evolution (4E of Teaching)
! Rapid adoption of teaching new technologies and concepts
! Ensuring mobile as well as offline e-learning access
! Safeguarding and motivation for long-term impact and top management
awareness and acknowledgement
Table 1: Customer Requirements on the Teaching PSS.
3
Conceptualization and Implementation
3.1 Definition of To-Be Properties and Synthesis of Properties
The to-be properties now can be defined based on the requirements. The collected data
of the previous step (cf. Table 1) is grouped into functional aspects and subsequently
transferred to the level of properties of the product model of the PSS. In Figure 1 an
excerpt of the developed model is shown. The characteristics and properties as well as
their relations are depicted.
Within the synthesis phase, the to-be properties of the PSS are realized by the definition
of system characteristics. Existing logical relations and interdependencies between product and service components within the PSS are illustrated by the inner relationships
between characteristics (cf. Figure 1). Based on the customer requirements new characteristics of the PSS are added in order to analyze their impact on the previously determined requirements in the form of relations. As a result teaching ITMC is especially
related to content, cooperation and mutual exchange together with the enabling technological platform. Hence, we focus on these aspects in a way that the characteristics of the
software component of the PSS are constituted and their relations to the level of properties are determined (cf. Figure 1).
221
Inner Relationships
Level of Characteristics
Teaching ITMC
Relations
Stakeholder
Participant
Researcher
Practitioner
Level of Properties
4Es of Teaching
Enrollment
Execution
Evaluation
Evolution
Content and Knowledge
Adoption new Technologies
Overcoming academia-practice gap
Social Aspects
Changed Behaviour
Application of Skills
Motivation
Top Management Acknowledgement
Certification
Monitoring
Costs
Economic goals
Ecological goals
Resources
Presentation Slides
Worksheets and Exercises
Documentation
Teaching
Module
Training Unit
Content
Method
Skill
Competence
Integration
Software interfaces
Data storage
Document management
Processes
EIPS
Data storage
Integrated DWH Solution
Reporting
Maintenance
Usability
Intuitive operation
Data quality
Timeliness
Granularity
Scalability
Adaption to growth of company
Infrastructure Mgmt.
Figure 1: Excerpt of the Level of Characteristics and Level of Properties of the Product Model.
3.2 Analysis of As-Is Properties
By analyzing the as-is properties one can monitor the fulfillment of the design goals
from various perspectives. For doing so, we use a descriptive and argumentative evaluation [Fr06]. We developed the PSS using expert interviews and a systematic literature
review. The resulting product model fulfills the requirements, which have been gained
by the definition and synthesis phase (cf. section 3.1). During the analysis we recognized
that a data warehouse (DWH) needs to be integrated into the IT architecture [BK10] in
order to cope with requirements like quality of data, the degree of integration as well as
the generation of reports (cf. Figure 1). This DWH is incorporated into an Education
Integration Platform Solution (EIPS) and additionally offers central services like a role
concept and an authorization mechanism. Unlike existing Learning Management Systems, which have been mainly developed by looking at technical possibilities [ZZZ04],
our system uses a holistic approach in looking at the user requirements on the complete
ITMC teaching process as a whole.
A high-level data model of the prototype is depicted in Figure 2 as Entity-Relationship
Diagram. Of course, the stakeholders and their affiliation, e.g. university or company
name, are represented. Stakeholders in EIPS can be participants, researchers and practitioners while feedback can be given to from any stakeholder to any other one. Researchers and practitioners work together in a partnership and therefore act as a tutor within a
training unit. This training unit has several participants and consists of a hierarchy of
training units which together form the module list. The necessary modules for teaching
ITMC have been outlined by [BSB11]. Each training unit has its at least one associated
content, but each content should only be covered by one training unit in order to avoid
redundancies. The content is linked to a set of skills. Skills are ability to apply the
learned content and competencies are defined as the proven ability to apply content
222
[EE11]. These two concepts are modeled separately in EIPS in order to document the
learning progress. For each skill there are several teaching methods [HMT05] represented in the system. New teaching methods like simulation [LPP11] and practical case
study exercises [TR05] are also supported by EIPS. Additionally, specific resources can
be stored and assigned to one or more of the methods.
M-R
Feedback
(0,m)
Stakeholder
Module
(0,1)
(0,m)
Participant
Participation in
relevant units
(0,n)
(1,n)
(0,1)
Training Unit
Method
Lecture
(0,n)
C-T
(1,n)
Researcher
Resource
(0,n)
(0,m)
(1,n)
(0,n)
Literature Study
M-S
(1,n)
(1,n)
(0,n)
(0,n)
Partnership
Affiliation
Practitioner
(1,1)
Tutoring
Service
(1,n)
Content
(0,n)
(0,n)
...
Skill
(0,n)
(0,n)
S-Cont
S-Comp
Competence
Figure 2: Data Model of the EIPS.
4
Conclusion and Outlook
We conceptualized teaching ITMC as a hybrid package of products (resources) and services (teaching) in this paper. Synthesis-analysis cycles of a systematic procedure model
have been than run through in order to derive and implement of our integration concept.
As our research is currently in progress, only the Education Integration Platform Solution (EIPS) has been briefly presented.
The proposed approach offers some limitations and therefore raises further questions:
First, we have not been able to evaluate the PSS entirely. This is the task of future research. Also a detailed comparison with other Learning Management Systems is required
and needs to be done in a real world environment. Second, our approach cannot cover all
requirements on an integrated approach. The success of teaching strongly depends on the
persons who teach and those who are willing to learn (social aspects like mutual acceptance and understanding, motivation etc.). Future research might cover these issues in
investigating mechanisms for facilitating them. The task of future research is to find out,
how the proposed approach can be implemented and offered in an efficient way. However, we are convinced that the utilization of the PSS engineering method in the context of
teaching is a novel approach which results in an effective solution.
5
Acknowledgements
This paper was written in the context of the research project IMUCON which is funded
by the European Regional Development Fund (ERDF). The authors acknowledge the
support by ERDF and all involved project partners as well as the anonymous reviewers.
223
6
References
BD95
BFS11
BK10
BSB11
CM09
EE11
Fr06
HMT05
La10
LPP11
Lu04
MT03
OB11
SBZ11
TR05
TWL08
VSB12
Wa07
Wi02
WSB04
WW02
ZZZ04
Bloomfield, B. P.; Danieli, A.: The Role of Management Consultants in the Development of Information Technology. The Indissoluble Nature of Socio-Political and Technical Skills. In Journal of Management Studies, 1995, 32; pp. 2346.
Boehm, M. et al.: Towards an Integrated Approach for Resource-Efficiency in Server
Rooms and Data Centers: ECIS 2011 Proceedings, 2011; Paper 100.
Baars, H.; Kemper, H.-G.: Business Intelligence in the Cloud? In PACIS 2010 Proceedings, 2010; Paper 145.
Boehm, M. et al.: An Integrated Approach for Teaching Professionals IT Management
and IT Consulting: AMCIS 2011 Proceedings, 2011; Paper 72.
Cragg, P.; Mills, A.: Understanding IT Management in SMEs. In Proceedings of the
European Conference on Information Management & Evaluation, 2009; pp. 116123.
European Parliament; European Council: The European Qualifications Framework
(EQF) for Livelong Learning. http://ec.europa.eu/education/lifelong-learningpolicy/doc44_en.htm, accessed 20 May 2011.
Frank, U.: Towards a Pluralistic Conception of Research Methods in Information Systems Research. ICB Research Report No. 7, University Duisburg-Essen, 2006.
Heim, G. R. et al.: Experiential Learning in a Management Information Systems Course:
Simulating IT Consulting and CRM System Procurement. In Communications of the
ACM, 2005, 15; Article 25.
LaFrance, G.: Bridging the IT Skills Gap Through Industry and Academic Collaboration.
In Employment Relations Today, 2010, 36; pp. 2530.
Léger et al.: Business Simulation Training in Information Technology Education: Guidelines for New Approaches in IT Training. In Journal of Information Technology Education, 2011, 10; pp. 3953.
Luftman, J. N.: Managing the Information Technology Resource. Leadership in the
information age. Prentice Hall, Upper Saddle River, N. J, 2004.
Minch, R. P.; Tabor, S. W.: Networking Education for the New Economy. In Journal of
Information Technology Education, 2003, 2; pp. 5159.
Osorio, J.; Bulchand, J.: Moving towards 2020: A Tentative Approach to ITEM. In
(Tatnall, A.; Kereteletswe, O.; Visscher, A. Eds.): Information Technology and Managing Quality Education. Springer Boston, 2011; pp. 104112.
Stolze, C. et al.: Towards Sustainable IT by Teaching Governance Practices for InterOrganizational Dependencies: Proceedings of IFIP 8.6 Hamburg Conference, 2011.
Tatnall, A.; Reyes, G.: Teaching IT Project Management to Postgraduate Business Students. A Practical Approach. In Journal of Information Technology Education, 2005, 4;
pp. 153166.
Thomas, O.; Walter, P.; Loos, P.: Design and usage of an engineering methodology for
product-service systems. In Journal of Design Research, 2008, 7; pp. 177195.
Valacich, J. S.; Schneider, C.; Behl, R.: Information Systems Today. Managing in the
digital world. Pearson, New Jersey, 2012.
Warschauer, M.: The paradoxical future of digital learning. In Learning Inquiry, 2007, 1;
pp. 4149.
Winter, R.: An Executive MBA Program in Business Engineering. A Curriculum Focusing on Change. In Journal of Information Technology Education, 2002, 1; pp. 279288.
Weber, C. et al.: Modelling of Product-Service Systems (PSS) Based on the PDD Approach: International Design Conference (Design 2004), 2004; pp. 547554.
Webster, J.; Watson, R. T.: Analyzing the Past to Prepare for the Future: Writing a Literature Review. In MIS Quarterly, 2002, 26; pp. xiiixxiii.
Zhang, D. et al.: Can e-learning replace classroom learning? In Communications of the
ACM, 2004, 47; pp. 75!79.
224
Modeling Choreographies:
BPMN 2.0 versus BPEL-based Approaches
Oliver Kopp, Frank Leymann, Sebastian Wagner
Institute of Architecture of Application Systems
University of Stuttgart, Universitätsstraße 38, 70569 Stuttgart, Germany
firstname.lastname@iaas.uni-stuttgart.de
Abstract: Choreographies capture the collaboration aspects between two or more processes. Explicit choreography notations have been included in the upcoming version 2.0
of the Business Process Model and Notation language (BPMN 2.0). This paper presents
a first evaluation of the choreography modeling capabilities of BPMN 2.0 and presents
a summary of the evaluation of BPEL-based approaches. The result is that BPMN 2.0
does not support reference passing and is tightly tied to technical configurations.
1 Introduction
The collaboration between two or more processes is captured by process choreographies
[Pel03]. There are two paradigms of choreography languages available: interconnection
models and interaction models [DKB08]. In both modeling styles, the participants interact
with each other and activities are connected. On the one hand, an interconnection model
connects communication activities belonging to two participants. Thus, each send/receive
message exchange is expressed using a connection between participants. On the other
hand, interaction models express each send/receive message exchange as atomic interaction.
Generally said, the terms “interaction model” and “interconnection model” are derived
from the way each modeling paradigm expresses a send/receive message exchange and
not from the fact that activities are connected or general interactions between processes
are presented. The current version of BPMN 2.0 [Obj10] supports both, interconnection
models and interaction models.
We use the requirements by Decker et al. [DKLW09] to evaluate the choreography capabilities of BPMN 2.0 and provide a comparison with BPEL-based choreography approaches. In
this work, we dropped “R10: Integration with service orchestration languages”. BPMN 2.0
also offers modeling orchestration of Web services and thus a discussion of mapping concepts from BPMN 2.0 to BPEL is obsolete. In fact, it is possible that a workflow engine
accepts both formats [Ley10].
A variant of the requirements by Decker et al. [DKLW09] is presented by Schönberger
[Sch11]. In principle, “Decomposability” and “Standardization” have been added, whereas
“Correlation” and “Message Formats” have been dropped. We did not adopt that requirements framework to keep the existing evaluations of BPELlight , WSFL, WS-CDL, Let’s
Dance, BPMN 1.x, iBPMN, BPSS/UMM, and SCA comparable with our work. Our future
225
work is to consolidate these requirements with the work of Schönberger and re-evaluate all
choreography languages using the consolidated requirements. Neither the requirements by
Decker et al. nor by Schönberger take comprehensibility of the language itself into account.
Such a quality assessment might be done by using the quality framework presented by
Krogstie et al. [KSJ06].
2 Chorographies in BPMN 2.0
In BPMN 2.0 interconnection models are implemented by collaboration diagrams. These
diagrams consist of two or more pools, where each of them represents a participant in
the overall collaboration. Within a pool, the internal process behavior of a participant
is modeled with BPMN common elements (e. g., activities, events, and sequence flows).
Alternatively, pools can be depicted as black boxes, i. e., no internal process behavior is
modeled. In the discussion below we focus on pools that contain internal processes as they
provide more information about the collaboration.
A choreography language has to support multi-lateral interactions (requirement R1). Multilateral interactions are supported by the pools and message flows in BPMN collaboration
diagrams.
It must be possible to specify the participant topology, i. e. the number and types of
participants involved in a choreography (R2). The participant types can be expressed by
pools and by using attributes, the minimum and maximum number of participants can be
specified. Hence, R2 is supported as well.
The choreography language has to provide also means to model an arbitrary number of
participants of the same type (“Service sets”, R3), where the concrete number is determined
at runtime. Multiplicity can by indicated in BPMN collaborations by three black parallel
lines in the bottom center of the pool, thus R3 is supported.
Another important choreography feature is “reference passing” (R4), where participant A
enables participant C to communicate with participant B by passing the reference of B to C.
This is not explicitly covered by the BPMN specification. The only way to indicate that
messages contain references are documentation elements.
Message formats (R5) cannot be expressed by the graphical notation of BPMN, i. e., there
exist no graphical elements in the specification to define the format of a message. A
message, however, may have an optional itemRef property (which is invisible in the
diagrams). This property represents an item definition, which contains a reference to a
concrete data schema (e. g., XML schema), thus R5 is fully met.
To enable an easy exchange of choreography implementations (as concrete WSDL portTypes and WSDL operations [W3C01]) the technical configurations of a choreography
have to be interchangeable and should be separated from choreography model (R6). In
BPMN collaborations, participants can be associated to interfaces (which are no graphical
elements). These abstract interfaces may contain references that point to a corresponding
technical implementation, such as WSDL portTypes. This, in turn, leads to the fact,
that a reference has to be changed if the name of the portType has been changed, thus
226
requirement R6 is not supported.
The modeling of time constraints (R7) has to be supported by choreographies, too. In
BPMN collaborations they can be modeled by timer intermediate events.
Another choreography requirement is “Exception handling” R8. This is also supported by
BPMN collaborations. An exception can be caught and raised by error events.
As choreographies may be enacted multiple times messages have to be routed to the
correct process instance that belongs to the choreography instance, thus a message correlation mechanism is required (R9). In BPMN correlation information cannot be specified
graphically in a diagram, but BPMN collaborations provide means to define correlation
information for the collaboration (i. e., R9 is met). For each collaboration a property named
correlationKey can be defined. This property consists of one or many correlation
properties that provide expressions to extract certain data from the payload of a message in
order to associate the message to a particular process instance.
Interaction models are realized by choreography diagrams in BPMN 2.0. These diagrams
define a special kind of processes that reflect the interaction sequence, i. e. the message
exchanges between two or more participants (which are represented by pools in the collaboration diagrams). The atomic building block in choreography diagrams is the choreography
task (briefly called “task” in the following). “[An atomic task] represents an Interaction,
which is one or two Message exchanges between two Participants.” [Obj10].
Requirement R1 is satisfied, because the purpose of choreography diagrams is the definition
of interaction behavior between two or more participants. If one participant has to interact
with multiple partners, a task for each interaction can be created.
Choreography diagrams have full support of requirement R2: One can specify the type
of a service involved in a choreography diagram. It is not possible, however, to determine the exact number of participant instances involved in an interaction. Using the
participantMultiplicity attribute, it is possible to specify minimum and maximum number of participants.
Three black parallel lines in the bottom center of the participant bands indicate that multiple
instances of a particular participant interact with another participant. Thus, requirement R3
is satisfied.
There is no explicit support for reference passing mechanisms (R4) in choreography diagrams. As explained in the context of interconnection models, the only way to indicate that
a message carries service references are natural language descriptions in documentation
elements, which lack formality.
Choreography diagrams support message formats (R5). Although messages cannot be
graphically visualized in the choreography diagram, each atomic task contains one or more
messages. As mentioned in the context of interconnection models, a message may have an
optional itemRef property that contains a reference to a concrete data schema.
Participants in collaborations can be associated to interfaces. This is also true for choreographies but as with collaborations an interface references to a portType (or any other
technical implementation) by using its qualified name. Consequently, the reference has to
be changed after the portType name has been changed. Thus, requirement R6 is violated
227
as the technical configuration cannot be interchanged without adapting the choreography
accordingly.
Requirement R7 (time constraints) is satisfied by BPMN choreography diagrams as timer
intermediate events can be added to the choreography.
A requirement that is not satisfied by BPMN choreographies is exception handling (R8):
Exceptions are only visible to the participant where the error occurred. That means that
the other participants are not aware of any exception that was thrown within a particular
participant. As a consequence, they cannot handle this exception. The only way for a
participant to inform the other participants about an exception are ordinary messages, i. e.,
there are no special error message types.
Like in BPMN collaboration diagrams, correlation information (requirement R9) cannot be
depicted in a choreography diagram either. BPMN 2.0 choreographies, however, provide a
property named correlationKey to define correlation information, thus R9 is met.
3 BPEL-based Choreography Approaches
BPEL supports modeling of single business processes. It does not directly support choreographies. Hence, BPEL4Chor [DKLW09] has been introduced to support interconnection
models and BPELgold [KEvL+ 11] has been introduced to support interaction models.
BPEL4Chor directly supports interconnection models. The behavior of each participant is
modeled using BPEL. Each behavior description is called “participant behavior description”.
Communication activities must not have WSDL-specific attributes such as partnerLink
and operation. The interconnection to other participants is established by “message
links”, which connect sending and receiving activities. A message link contains the attribute
participantRefs, which lists the participant references transmitted on the message
flowing on the message link. Message links themselves are listed in a “topology”. Besides
message links, the topology contains a list of participant types (referring to participant
behavior descriptions) and a list of participants and participant sets which denote the
participants taking part in one instance of the choreography. Decker et al. [DKLW09]
already evaluated BPEL4Chor using the requirements. We summarize the findings in the
following.
BPEL4Chor introduces a topology (R2) with sets of services (R3). BPEL4Chor allows
an arbitrary number of participant types and thus supports multi-lateral interactions (R1).
BPEL4Chor supports participant references by the attribute participantRefs on a
message link (R4). Specification of message formats is fully supported by BPEL4Chor
(R5). Technical configurations are stored in a separate grounding document, which enables
interchangeability of technical WSDL information in different settings (R6). Time constraints may be specified using BPEL’s control flow constructs (R7). BPEL’s exception
handling can be reused in BPEL4Chor (R8). Correlation may be specified on a name-basis
or directly using message properties (R9).
BPELgold is a choreography language providing interaction models using the control flow
semantics of BPEL. “gold” stands for global definition. The complete language and an
228
evaluation of the requirements is presented by Kopp et al. [KEvL+ 11]. Here, we provide a
summary of the language and the evaluation according to the requirements.
BPELgold reuses the topology and the grounding of BPEL4Chor. The participant behavior
descriptions are replaced by a single process interaction description. A message link of
the topology relates a sending participant to a receiving participant and excludes concrete
activities. The attribute participantRefs still exists.
The process interaction description is a BPEL process following the Abstract Process
Profile for Basic Interaction Models. “Basic” denotes that only interactions between listed
participants are allowed. BPELgold introduces the Abstract Process Profile for Extended
Interaction Models, where a global observer may also interact with the choreography.
This is used for runtime compliance checking of choreographies as described by Kopp
et al. [KvLN08]. This profile is out of scope of this paper. In the case of the basic
profile, BPEL’s communication activities are disallowed in BPELgold . Interactions are
solely described by an interaction activity. This activity presents an atomic message
exchange between two participants: One participant sends a message and another participant
receives that message.
As BPELgold builds on BPEL4Chor, it supports a topology (R2) with sets of services (R3).
BPELgold also supports arbitrary numbers of participant types and thus multi-lateral interactions (R1). The attribute participantRefs is still present in message links and thus
participant references can be passed over a message link (R4). Message formats can be specified using variables in the process model (R5). The concept of a grounding is still present
(R6): In a grounding, an interaction activity is assigned a partnerLink/operation
combination, which in turn is used in the services implementing the described participants.
Time constraints can be modeled using the constructs provided by BPEL (R7). BPELgold
supports the exception handling mechanisms of BPEL. Thus the exceptional path through
the choreography model can be explicitly described (R8). BPEL’s concept of correlation
is reused in BPELgold : Each interactionActivity contains a correlationSet
element referring to a correlation set used for correlation (R9). BPELgold can be mapped
to BPEL4Chor: each interactionActivity is split into an invoke/receive
pair. The name of each communication activity is generated. Thus, all information of a
BPEL4Chor message link is available.
4 Conclusion and Outlook
We compared the choreography modeling capabilities of BPMN 2.0 using the evaluation
criteria of Decker et al. [DKLW09]. For each requirement, we showed the constructs
offered by BPMN 2.0 to implement each requirement. Section 3 summarized the evaluation
of BPEL-based approaches. Table 1 summarizes the evaluation. On the one hand, it
turns out, that BPEL4Chor and BPELgold provide a complete support of the requirements,
whereas BPMN 2.0 lacks support: Reference passing and interchangeability of technical
configurations are not fully supported. On the other hand, only BPMN offers a full graphical
notation for both interaction models and interconnection models. Thus, a next step is a
detailed investigation of the integration of BPMN 2.0 with BPEL4Chor and BPELgold .
229
Modeling Style
R1. Multi-lateral interactions
R2. Service topology
R3. Service sets
R4. Reference passing
R5. Message formats
R6. Interchangeability of technical
configurations
R7. Time constraints
R8. Exception handling
R9. Correlation
BPMN 2.0
Collaboration
interconnection
+
+
+
–
+
–
BPMN 2.0
Choreography
interaction
+
+
+
–
+
–
BPEL4Chor
BPELgold
interconnection
+
+
+
+
+
+
interaction
+
+
+
+
+
+
+
+
+
+
–
+
+
+
+
+
+
+
Table 1: Assessment of BPMN 2.0 and BPEL-based approaches
This may be an extension of BPMN to support all constructs of BPEL4Chor similar to the
approach taken by Pfitzner et al. [PDKL07] in the case of BPMN 1.0 and BPEL4Chor.
Currently, there exists a mapping of BPMN 1.0 to BPEL4Chor [PDKL07] and no mapping
of BPMN 2.0 to BPEL4Chor. In case such a mapping is available, BPMN 2.0 collaborations
can be transformed to BPEL4Chor choreographies. Such a transformation would lead to a
fulfillment of requirement R6: The technical configuration is not done in the BPMN 2.0
model, but using the mechanisms offered by BPEL4Chor.
References
[DKB08]
G. Decker, O. Kopp, and A. Barros. An Introduction to Service Choreographies. it Information Technology, Special Issue on Service-oriented Architectures, 50(2), 2008.
[DKLW09] G. Decker, O. Kopp, F. Leymann, and M. Weske. Interacting services: from specification
to execution. Data & Knowledge Engineering, 68(10):946–972, April 2009.
[KEvL+ 11] O. Kopp, L. Engler, T. van Lessen, F. Leymann, and J. Nitzsche. Interaction Choreography Models in BPEL: Choreographies on the Enterprise Service Bus. In S-BPM ONE
2010, CCIS. Springer, 2011.
J. Krogstie, G. Sindre, and H. Jørgensen. Process models representing knowledge
[KSJ06]
for action: a revised quality framework. European Journal of Information Systems,
15(1):91–102, 2006.
[KvLN08] O. Kopp, T. van Lessen, and J. Nitzsche. The Need for a Choreography-aware Service
Bus. In YR-SOC, 2008.
[Ley10]
F. Leymann. BPEL vs. BPMN 2.0: Should You Care? In 2nd International Workshop
on BPMN, LNBIP. Springer, 2010.
Object Management Group. Business Process Model and Notation (BPMN) Version 2.0.
[Obj10]
OMG, 2010.
[PDKL07] K. Pfitzner, G. Decker, O. Kopp, and F. Leymann. Web Service Choreography Configurations for BPMN. In WESOA, volume 4907 of LNCS. Springer, 2007.
C. Peltz. Web Services Orchestration and Choreography. IEEE Computer, 36(10):46–52,
[Pel03]
2003.
[Sch11]
A. Schönberger. Do We Need a Refined Choreography Notion? In ZEUS. CEURWS.org, 2011.
[W3C01]
W3C. Web Services Description Language (WSDL) 1.1, March 2001.
230
Process Model Verification with SemQuu
Michael Fellmann, Oliver Thomas
Institut für Informationsmanagement und Unternehmensführung
Universität Osnabrück
Katharinenstr. 3
49074 Osnabrück, Germany
{michael.fellmann, oliver.thomas}@uni-osnabrueck.de
Abstract: In this contribution, we present an overview of our ongoing work in
developing an integrated tool support for semantic process model verification. We
present some of the requirements and insights gained by the implementation of a
platform for semantic process model verification. The platform is based on OWL
and SPARQL and provides for a user-friendly way of semantic model verification.
First empirical evaluations have been very promising and confirm that the tool can
be used to effectively query and verify semi-formal process knowledge.
1 Introduction
1.1 Background
A fundamental idea in the context of the model-driven development of information
systems is to use models to manage the complexity accompanied with the development
task. They provide a means for understanding business processes and serve for
optimization, reengineering, and development of supporting IT systems. Especially in
regard to the latter aspect, the quality of the models is of vital importance since it
directly affects the quality of the information systems developed on the basis of the
models. According to ISO 8402, quality is “the totality of characteristics of an entity that
bear on its ability to satisfy stated and implied needs”. One fundamental aspect of model
quality is the correctness of a model. There has been extensive research regarding
syntactical correctness. In contrast to that, approaches to verify the semantic correctness,
i.e. the correctness of the model contents, have been considered less frequently. With the
term “verification”, we denote ensuring correctness with respect to explicitly stated
requirements. In contrast to that, validation means the eligibility of a model in respect to
its intended use [De02, p.24] – in other words: if the criteria is something outside the
model [Me09, p.2]. Following this distinction, we call the procedures to ensure semantic
correctness “semantic verification”.
A major problem regarding semantic verification is how to automate it. This problem is
rooted in natural language being used for labeling model elements, thus introducing
terminological problems such as ambiguity (homonyms, synonyms) and other linguistic
231
phenomena. Model creators and readers do not necessarily share the same understanding
as the concepts they use are usually not documented and mix both discipline-specific
terminology and informal, ordinary language. Hence, it is hard for humans to judge if a
model is semantically correct and almost impossible for machines (apart from using
heuristics) because the model element labels are not backed with machine processable
semantics. The result is that the machine cannot interpret the contents of model
elements. Therefore, we have developed an approach to encode the model element
semantics in a precise, machine readable form using OWL ontologies. On top of the
ontology-based representation, SPARQL-queries are used to check the correctness of a
model based on its graph structure. We use this query language as (a) it is a language
which has been standardized by the World Wide Web Consortium (W3C) and gained
broad acceptance in the Semantic Web and AI community, (b) it provides a simple
structure for queries consisting of triples which intuitively reflect the structure of process
graphs and (c) the query language is not specialized to any specific process modeling
language such as EPC, BPMN or UML Activity Diagrams. The approach is described in
more detail in [FeTh11]. In contrast to our previous work where we elaborated on our
concept of semantic verification, in this contribution we concentrate on the description
of our prototypical implementation. We describe the requirements, design decisions and
insights gained by extending a modeling tool with support for semantic verification.
1.2 Related Work
Approaches to ontology-based semantic verification can be found e.g. in the context of
Semantic Web Services. Semantically annotated process models are verified with an
emphasis on logical preconditions and effects which are specified relative to an ontology
[Dr07], [We09], [WHM10]. These approaches usually require both the annotation of
preconditions and effects and hence enable to check if the model is consistent. They do
not build upon a formal representation of the (intentional) semantics of individual model
elements (i.e. what is “inside the box” of a model element). In contrast to that, the
underlying paradigm of our approach is capturing the semantics of model elements by
using a single annotation of a model element in order to associate it with its intended
meaning explicitly specified in a formal ontology. Semantic constraints are then
formulated as SPARQL queries which are executed against the model repository. Such
queries focus constraints related to the structure of a model (e.g. to verify the absence of
an anti-pattern) and are in this respect orthogonal to approaches emphasizing the correct
execution of process instances [Ly09]. In respect to the IT-support of SPARQL queries,
query builders such as Konduit VQB have been developed [AMH10]. Whereas these
tools provide complex forms aiming at a general-purpose support for SPARQL, our
approach is tailored to query process models. We provide a simple yet powerful user
interface and we have evaluated the effectiveness in a user experiment.
232
2 Requirements
In this section, we describe basic requirements regarding the functionality of the
integrated tool support for our approach to semantic model verification described in
[FeTh11]. To explore the general topic of queries and search in process models, we
conducted a survey. The participants of the survey (50 from both academia and practice)
clearly favored form-based approaches over all other approaches for querying model
contents. As form-based approaches require less knowledge in comparison to the usage
of a query language, it can be concluded that the users want to have a simple and userfriendly way of specifying queries. However, by relying exclusively on form-based
queries, much of the freedom and flexibility a query language such as SPARQL for
querying graph like structures offers is lost. We therefore suggest using a hybrid
approach where triple patterns of SPARQL queries can be constructed using a form and
adapted as plain-text manually. In order to execute multiple queries e.g. when checking
if a set of constraints is met, the tool has to provide for the batch-execution of queries.
Finally, we do not want to confine ourselves to a specific modeling language such as
BPMN or EPC. Rather, the tool should transform different model representations into
the ontology-based process representation devised in our previous work. Below, we
summarize the requirements.
(R1) Form-based, fail-safe query construction: Since our survey revealed that formbased approaches of query construction are favored over other alternatives, the tool has
to offer support for query construction using forms. The user should construct a query
exclusively by interacting with the form, i.e. no textual input should be required. Only
valid queries both in respect to syntax and semantics should be producible with the formbase query construction feature.
(R2) User-friendly query refinement: Since we propose the usage of SPARQL for
querying process models and for specifying semantic constraints, the user should be able
to use the full expressivity of this query language. However, the user should be
supported in refining a query which in its initial form is generated using the form-based
approach of query construction (cf. R1).
(R3) Execution of batch queries: In order to execute multiple queries at once, the tool has
to provide for the batch-execution of queries. The results should be summarized in a
report whereby it should be possible to assign a type to each query result as either being
information, a warning or an error.
(R4) Support of multiple modeling languages: In order to achieve a broad applicability
of the tool, it should not be dependent on a specific modeling language such as the EPC,
BPMN etc. but instead should be generic.
Moreover, other requirements such as performance and scalability were considered as
well during the development process. Due to space limitations, we have not listed them
as requirements but mention them in the appropriate sections of the platform description.
233
3 Using SemQuu for Process Model Verification
In the following, we describe the tool we have developed for process model verification.
The tool is called SemQuu – Semantic Query Utility, since much of the functionality is
also relevant when exploring ontologies in general. SemQuu is implemented using the
Jena library (jena.sourceforge.net), the Pellet inference engine (pellet.owldl.com), the
Tomcat web server and JSP, XSLT, JavaScript, CSS and HTML for the user interface.
3.1 Import of Process Models
Process models can be imported from arbitrary tools in arbitrary formats, since they can
be transformed on the server into the ontology-based representation which is
accomplished by a plugin-converter for each format (R4 is met). As an example, we have
implemented an extension of Visio which can export EPC process models being
annotated with ontology concepts (see Fig. 1 for an illustration of the following
procedure). For export from Visio, we use the standard EPML format (Event-driven
Process Chain Markup Language) with special attributes capturing the semantic
annotation of the model. After having been exported from Visio to SemQuu, the model
is transformed to an OWL-DL ontology and added to the repository (cf. Fig 1 right).
Fig. 1 SemQuu Converter and Repository
3.2 Querying the Semantic Process Repository using SPARQL
An overview of SemQuu is provided by Fig. 2 (B). In order to query the repository of
ontology-based process representations with SPARQL, the user can use a simple formbased query builder (A) for successively constructing a graph pattern (R1 is met). This is
done by inserting multiple triple patterns with the help of drop-down list boxes (A)
which are aware of rdfs:domain and rdfs:range information so that no semantically
wrong query can be constructed (R1 is met).
234
Fig. 2 Overview of SemQuu
Moreover, drop-down list boxes are dynamically updated upon selection of a value in
one of the boxes. Alternatively to the drop-down list boxes, the user can leverage the full
expressivity of SPARQL by using the text area input field. Moreover, when the user
modifies the query she or he is supported by an “intelligent” auto-completion feature (D)
which is fully aware of the ontology schema and instance data and only proposes
meaningful suggestions (R2 met). When queries are executed in batch mode, the result of
the queries can be displayed as an information, warning or error with respective
graphical symbols appearing for each type (C). The result for each query is initially
collapsed but can be unfolded if the user clicks on the “+” sign symbols (R3 is met).
4 Evaluation
In order to measure the effectiveness of SPARQL queries for the task of semantic
correctness checking, we conducted an experiment with 21 participants. We have
described this experiment in more detail in [FeTh11]. We give a short overview of the
results of this experiment here. We used models in three different sizes with 11, 39 and
217 model elements. The participants had to answer 10 questions per model (a) by
browsing and searching in the model with the tool Microsoft Visio or (b) by using
SemQuu. We conducted the experiment without the form-based query construction
feature since it was not been available at the time we conducted the experiment
(however, we plan to re-evaluate the tool). Fig. 3 shows the time required for answering
the 10 questions for each of the three models in relation to the model size. As can be
seen easily, at first there is a learning curve when using SPARQL. When the model size
exceeds approx. 190 model elements, the query language outperforms manual search.
235
Time
00:23:02
00:20:10
00:17:17
00:14:24
00:11:31
00:08:38
00:05:46
00:02:53
00:00:00
Brow sing
SPARQL
0
75
small
150
225
medium
large
Model size
Fig. 3 Time Needed to Answer the Questions Dependent on Model Size
5 Conclusion and Outlook
In this contribution, we have introduced SemQuu, our prototype for semantic process
verification. Results of an initial evaluation regarding the query formulation have been
very promising. The tool is under active development, currently we are working on
improvements regarding the import and export of model data from Microsoft Visio.
Literature
[AMH10] Ambrus, O.; Möller, K.; Handschuh, S.: Konduit VQB: a Visual Query Builder for
SPARQL on the Social Semantic Desktop. In (Handschuh, S. et al. Eds.): Proc. of
VISSW 2010, Hong Kong, China, 2010; 5 Pages.
[De02]
Desel, J.: Model Validation – A Theoretical Issue? In (Esparza, J., Lakos, C. Eds.)
Proc. of ICATPN 2002, Adelaide, Australia. Springer (LNCS 2360), 2002; pp. 23–43.
[Dr07]
Drumm, C. et al.: Dynamic Composition Reasoning Framework and Prototype. In:
Project IST SUPER, Deliverable 3.2, SAP, 2007.
[FeTh11] Fellmann, M.; Thomas, O.; Busch, B.: A Query-driven Approach for Checking the
Semantic Correctness of Ontology-based Process Representations. In Proc. of BIS
2011, Poznan, Poland, 2011. Springer (LNBIP 87), Berlin; pp. 62–73.
[Ly09]
Ly, L.T. et al.: On enabling integrated process compliance with semantic constraints in
process management systems. In: Information Systems Frontiers, 2009; pp. 1-25.
[Me09]
Mendling, J.: Empirical Studies in Process Model Verification. In (Jensen, K.; van der
Aalst, W.M.P. Eds.): Transactions on Petri Nets and Other Models of Concurrency II.
Springer (LNCS 5460), Berlin, 2009; pp. 208–224.
[W09]
Weber, I.M.: Verification of Annotated Process Models. In: Semantic Methods for
Execution-level Business Process Modeling, Springer (LNBIP 40), Berlin, 2009; pp.
97–148.
[WHM10] Weber, I.; Hoffmann, J.; Mendling, J.: Beyond soundness: on the verification of
semantic business process models. In: Distributed and Parallel Databases (27), 2010;
pp. 271–343.
236
Designing a Risk-based Partner Selection Process
for Collaborative Cloud Computing Environments
Benedikt Martens1, Novica Zarvi!2, Frank Teuteberg1, Oliver Thomas2
1
Accounting and Information Systems
Information Management and Information Systems
University of Osnabrueck
Katharinenstr. 1-3
49069 Osnabrueck
{benedikt.martens; novica.zarvic; frank.teuteberg; oliver.thomas}@uni-osnabrueck.de
2
Abstract: Cloud Computing represents a shift in technology and IT (information
technology) service management and it opens up possibilities to build new
organizational configurations. As in IT outsourcing, the selection of the right
partner/vendor is of crucial importance. Thus, in this article we present a partner
selection process for four Cloud Computing Environments: public cloud,
community cloud, cloud chain and industrial cloud. We included data sensitivity
and the risk attitude of the decision maker as major decision factors for partner
selection. The constructed IT artifact (i.e. partner selection process) is evaluated by
means of a conceptual evaluation (expert interviews) that demonstrates the
applicability of the selection process.
1 Introduction
The advent of the internet, as well as advances in related concepts and technologies (as
e.g. Cloud Computing) have led to an increase in research on Collaborative Networks
and inter-organizational business practices [CA05]. In the course of that, researchers
increasingly focused on the inter-organizational use of IT instead of taking a solely intraorganizational view (cf. [KD96]). Cloud Computing is a new paradigm for the delivery
of scalable IT services via the internet on a pay-per-use basis [Ar10]. In this context, the
term cloud refers to data centers that offer virtualized computing resources and
services. The combined investigation of Cloud Computing Services and Collaborative
Networks results in new collaboration types with a focus on resource sharing. In this
paper we posit that Cloud Computing works on the same basic principles as all
Collaborative Networks: two or more businesses that are largely autonomous and
geographically distributed collaborate with each other and are supported in this by IT
networks [CA05, WRT09]. Thus, the sourcing of Cloud Computing services always
implies some degree of dependency on external partners and a certain level of risk for
each participating unit [KD96, Ar10]. One of the biggest challenges regarding the
success of Collaborative Networks is the selection of partners in inter-organizational
business settings [WCB91]. In view of all this we propose a partner selection method for
conceivable Cloud Computing Environments with a focus on risk and security factors.
237
We aim at developing a new partner selection method for the field of Cloud Computing.
The conceptual elaboration of the IT artifact is based on a systematic literature review
focusing on terms from the field of Cloud Computing and provider/vendor selection that
have been applied to publisher independent journal data bases [MT09]. Overall we
identified a sample of 113 articles published in journals and conference proceedings.
Finally we improved the artifact by means of three expert interviews. The interviewees
were experienced practitioners and academics with 3 to 10 years of work experience in
the field of Collaborative Networks, Partner Selection and IT outsourcing.
3 Reviewing Cloud Computing and Partnership Formation
Researchers have especially focused on gaining more insights into Cloud Computing and
its multiple facets. For instance, Youseff et al. [YBS08] propose an ontology, which
illustrates the relevant factors of Cloud Computing and their relationships. As yet, little
research has been conducted on the success factors and the risks associated with Cloud
Computing Services [MT09]. Koehler et al. [KAD10] found that the average reputation
of the Cloud Computing Service provider and the use of standard data formats are more
important than financial aspects such as cost reduction or pricing tariffs. Armbrust et al.
[Ar10] present a list of ten obstacles for Cloud Computing, of which the following four
are assumed to affect adoption: availability/business continuity, data lock-in, data
confidentiality and auditability. We found that several authors point out the importance
of risk and security methods for Cloud Computing Services [YBS08, MT09]. From a
vendor perspective, there are some obstacles affecting the growth of Cloud Computing
as well as policy and business issues, e. g. data transfer bottlenecks [Ar10]. Nevertheless,
Cloud Computing facilitates the introduction of new products and services without large
investments in IT infrastructure [Pu09]. Pricing strategies and revenue models are
suggested for instance by Pueschel et al. [Pu09] in order to exploit the economic
opportunities offered by this emerging paradigm. The literature review revealed that only
little research has been conducted in the field of Collaborative Networks. The only
article on this topic that we could identify is the one by Wlodarczyk et al. [WRT09], who
deal with the aggregation of several private Clouds to reduce security risks, which can
also be regarded as a form of Collaborative Networks.
The duration of collaborations between different companies or organizations can be
described by the so-called life cycle of enterprise networks [TJ01]. By and large, the
enterprise network life cycle can be divided into four different phases: a preparation
phase (preparing, first sourcing of partners etc.), a formation/setting up phase (final
partner selection, legal issues, contracts etc.), an operation phase (daily management of
the network) and a dissolution phase (decomposition of the network). Therefore,
partnership formation happens both during the preparation phase and in the
formation/setting up phase. Partner identification is concerned with the identification of
possible business partners. For example, Large [La09] distinguishes next to the
different request types also the internet as a source for identifying conceivable
partners. Additionally he lists classical sources such as personal contacts, professional
journals, annual reports, trading catalogues and alike. Apart from these two categories, it
is also conceivable to identify possible partners indirectly, that is, via intermediaries.
238
What all these categories have in common is the fact that they are self-initiated
approaches. Partner selection, as the second and final step in the process of partnership
formation, can be performed by means of several methods and techniques. In the
literature review by Weber et al. [WCB91] and Zhang et al. [Zh04] 123 articles have
been reviewed paying special attention to the criteria and methods used in the partner
selection process. As regards the methods applied, the following three categories were
initially proposed by Weber et al. [WCB91]: linear weighting approaches, mathematical
programming techniques and statistical/probabilistic approaches. We have added
another category to this three-way categorization, which includes all other approaches
found in the literature that did not fit any of these three categories as, for instance,
serious games and simulations [Ha08], agent-based approaches [Pe07] or activity-based
costing [RK97]. However, the results of the literature review give the impression that the
selection of partners in the context of Cloud Computing has not yet received sufficient
attention from the research community. To the best of our knowledge, we are the first
researchers to apply the concept of partner selection to the realm of Cloud Computing
Environments with a specific focus on risk and security.
4 Collaboration Types in Cloud Computing Environments
A selecting process for Cloud Computing partners/vendors needs to consider the
individual characteristics of each Cloud Computing Environment in order to be effective.
The following typology was developed deductively on the basis of scientific literature
and evaluated by means of expert interviews. In Thompsons [Th67] seminal work on
organizational theory, he described pooled, sequential and reciprocal interdependencies
for an intra-organizational perspective, which were later mapped to the interorganizational context by Kumar & van Dissel [KD96]. In the field of Cloud Computing
we have identified collaboration configurations that share the same structural properties
as these inter-organizational system types. This analogy led to the definition of three
distinct Cloud Computing types which are conceptually depicted in Figure 1. The
following description points to security and risk-related aspects that are relevant for the
development of a partner selection method. In Pooled Cloud Computing one business
actor sources a service from the central point. This might be represented either by a
(standard) public cloud environment or a community cloud environment [Ar10]. In
contrast to public clouds, community clouds are only available for members of a specific
community (i. e. shared needs of a business community), where interdependencies
remain the same, which is the reason why both belong to the same collaboration type.
Community clouds provide a slightly higher security level than public clouds [WRT09].
Public availability has an enormous impact on risk management, for it requires high risk
management effort, whereas limited availability to specific communities reduces the
effort. Sequential Cloud Computing includes public clouds and represents a chain of
interdependencies by means of service flows (cloud chain). The access to this type of
cloud is usually open, resulting in a low security level comparable to that of (standard)
public clouds. Finally, in Networked Cloud Computing the output of each point can
become input for the other points, as happens, for example, in so-called industrial clouds
[WRT09]. However, the fact that they constitute a closed network of private clouds
guarantees for a high level of security [WRT09].
239
Figure 1. A comprehensive and risk-based overview of Cloud Computing Environments.
5 Development of a Partner Selection Process
The developed approach combines a standard partner selection process [CNR95, MF97]
with a special focus on risk and security issues in Cloud Computing. The process is
illustrated in Figure 2. The preparation phase starts with the prerequisites of a sourcing
decision: the selection of a particular Cloud Computing Environment and the
identification of potential partners. The formation phase comprises the actual partner
selection and the initialization of the partnership. In addition to the actual process we
added methods, theories and documents to individual process steps. Since we investigate
the outsourcing of standardized Cloud Computing Services, we assume a newly arising
requirement as the first process step (e. g. the demand for a new IT service triggered by a
new business opportunity). This new requirement needs to be documented, analyzed and
aligned to the business strategy and the IT strategy of the focal company [MF97]. This
step is of particular importance for the follow-up activities because it accounts for
security and risk requirements, too. The make or rent decision is made on the basis of the
availability of internal resources and the ratio of production and sourcing costs. Next, the
company needs to determine its requirements regarding the risk and security level of the
IT service. During this step the risk attitude of the decision maker needs to be taken into
account. If the company considers a relatively low security level sufficient and decides
to enter an open access cloud it needs to distinguish between the Cloud Computing
Environments public cloud (pooled Cloud Computing) and cloud chain (sequential
Cloud Computing) (cf. Figure 1 and Figure 2). Within the former environment the focal
company acts as an end user and within the latter environment as an intermediator. For
the following steps of this track, we recommend applying the approaches discussed in
Section 3. On the other hand, the process paths for closed access Cloud Computing types
(community cloud and industrial cloud) with medium and high security levels can
support companies in their search for existing community clouds or industrial clouds.
Following the identification of suitable networks, the decision-making company
retrieves more detailed information on these networks in order to review them and to set
240
up a ranking list. Finally, the company submits a request for membership to the best
fitting network. If access to this network is denied, the next best one is applied for.
Figure 2. Partner selection process for specific Cloud Computing Environments.
6 Conclusions and Outlook
The presented methodic approach supports companies in selecting partners for
collaborations in Cloud Computing environments and provides a strategic framework for
such collaborations from a managerial, technological and organizational perspective. In
this context, a strong focus was put on risk and security issues. In particular, the method
serves to support companies with knowledge in the field of vendor management in
Cloud Computing. We built the conceptual model (cf. Figure 2) on a deductive
theoretical basis (theories and methods) and improved it iteratively by means of semistructured expert interviews [Wal06]. Limitations of our approach lie in the limited
number of decision factors we draw on (security and risk issues). Furthermore, we
concentrated on the organizational forms Cloud Computing offers to companies. We are
planning to adapt our method to the requirements of the different Cloud Computing
Service types, which will mainly result in a modification of the first part of the process.
241
Furthermore, a closer look at interdependencies within clouds or between the
participating collaboration partners could lead to important insights regarding the
discussed Cloud Computing Environments. There are many interdependencies within
industrial clouds. Partners in such a network are strongly dependend on each other,
which offers new business opportunities, but also new risk factors.
References
[Ar10]
Armbrust, M.; Fox, A.; Griffith, R.; Joseph, A.D.; Katz, R.H.; Konwinski, A.; Lee, G.;
Patterson, D.A.; Rabkin, A.; Stoica, I.; Zaharia, M.: A View of Cloud Computing. In
Communications of the ACM, Vol. 53, No. 4, 2010; pp. 50-58.
[CA05] Camarinha-Matos, L.M.; Afsarmanesh, H.: Collaborative Networks: A New Scientific
Discipline. In Journal of Intelligent Manufacturing, Vol. 16, No. 4-5, 2005; pp. 439-452.
[CNR95] Chaudhury, A.; Nam, K.; Rao, H.R.; Management of Information Systems Outsourcing:
A Bidding Perspective. In Journal of MIS, Vol. 12, No. 2, 1995; pp. 131159.
[Ha08] Hans, C.: Supporting partner identification for virtual organisations in manufacturing. In
Journal of Manufacturing Technology Management, Vol. 19, No. 1, 2008, pp. 497-513.
[KAD10]Koehler, P.; Anandasivam, A.; Dan, M.A.: Cloud Services from a Consumer Perspective
Cloud Services from a Consumer Perspective. AMCIS 2010, Peru.
[KD96] Kumar, K.; v. Dissel, H.: Sustainable Collaboration: Managing Conflict and Cooperation
in Interorganizational Systems. In MISQ, Vol. 20; No. 3, 1996; pp. 279-300.
[La09] Large, R.: Strategisches Beschaffungsmanagement Eine praxisorientierte Einführung.
Gabler, Wiesbaden, Germany, 2009.
[MT09] Martens, B.; Teuteberg, F.: Why Risk Management Matters in IT Outsourcing - A
Systematic Literature Review and Elements of a Research Agenda. ECIS 2009, Italy.
[MF97] Michell, V.; Fitzgerald, G.: The IT outsourcing market-place: vendors and their
selection. In Journal of Information Technology, Vol. 12, No. 3, 1997; pp. 223-237.
[Pe07] Petersen, S.: Virtual enterprise formation and partner selection. In International Journal
of Networking and Virtual Organisations, Vol. 4, No. 2, 2007; pp. 201-215.
[Pu09] Pueschel, T.; Anandasivam, A.; Buschek, S.; Neumann, D.: Making Money with Clouds
Revenue Optimization Through Automated Policy Decisions. ECIS 2009, Italy.
[RK97] Roodhooft, F.; Konings, J.: Vendor selection and evaluation an Activity Based Costing
approach. European Journal of Operational Research, Vol. 96, No. 1, 1997; pp. 97-102.
[TJ01] Thoben, K.-D.; Jagdev, H.: Typological issues in enterprise networks. In Production
Planning & Control, Vol. 12, No. 5, 2001; pp. 421-436.
[Th67] Thompson, J.D.: Organizations in Action: Social Science Bases of Administrative
Theory. McGraw-Hill, New York, USA, 1967.
[Wal06] Walsham, G.: Doing Interpretive Research. In European Journal of Information Systems,
Vol. 15, No. 3, 2006; pp. 320-330.
[WCB91]Weber, C.A.; Current, J.R.; Benton, W.C.: Vendor selection criteria and methods. In
European Journal of Operational Research, Vol. 50, No. 1, 1991; pp. 2-18.
[WRT09]Wlodarczyk, T.W.; Rong, C.; Thorsen, K.A.H.: Industrial Cloud: Toward Interenterprise Integration. In LNCS 5931, Springer, Berlin, 2009; pp. 460471.
[YBS08] Youseff, L.; Butrico, M.; da Silva, D.: Toward a Unified Ontology of Cloud Computing.
In 2008 Grid Computing Environments Workshop, Austin, USA, 2008.
[Zh04] Zhang, Z.; Lei, J.; Cao, N.; To, K.; Ng, K.: Evolution of supplier selection criteria and
methods. In Proceedings of the Second Globelics Conference Innovation Systems and
Development, Emerging Opportunities and Challenges, Beijing, China, 2004.
242
Introducing a Co-Creation Perspective to Service Business
Models
Andreas Zolnowski, Martin Semmann, Tilo Böhmann
ISS International Business School of Service Management
Hans-Henny-Jahnn-Weg 9
22085 Hamburg
zolnowski@iss-hamburg.de
Institut für Informatik
Universität Hamburg
Vogt-Kölln-Straße 30
22527 Hamburg
semmann@informatik.uni-hamburg.de
boehmann@informatik.uni-hamburg.de
Abstract: Due to the growing importance of services for many companies and the
resulting transformation of product based business models to service based business models, the paper focuses on the link between business models and services.
For this purpose the business model canvas of Osterwalder is adapted to address
the shortcoming relating to co-creation of extant business models.
1 Introduction
Services are a key driver of growth and profitability for many companies [CS06]. Particularly firms in technology industries, such as IT, aerospace, medical technology, and
automotive capture an increasing share of their income and profits with services [Ne08].
Thus, for these enterprises, services have become an essential part of their business
models, leading more and more of these companies to transform their product based
business models to service based business models. According to this transformation
approaches for modeling are necessary. As a recent study shows, extant business model
approaches have a lack in focusing specific aspects regarding to services [ZB11]. This
paper therefore proposes the use of an adaptation of Osterwalder´s business model canvas to support the modeling of service business models.
The remainder of the paper is structured as follows. Conceptual foundations about the
understanding of business models in general and the business model canvas of Osterwalder as well as a general understanding of co-creation are given in chapter two. Based on
this theoretical background in chapter three we derive the problem of considering service
specific aspects in recent business model approaches. Moreover, we adapt the business
model canvas of Osterwalder to match the requirements of service based business models. Finally we sum up the findings and give an outlook.
243
2 Conceptual foundations
2.1 Business models
The academic literature offers a variety of possible conceptualizations of the business
model construct. Recently, however, the different approaches seem to converge. AlDebai [Al10] summarizes a business model as “[...] an abstract representation of an organization, be it conceptual, textual, and/or graphical, of all core interrelated architectural, co-operational, and financial arrangements designed and developed by an organization, as well as all core products and/or services the organization offers based on these
arrangements that are needed to achieve its strategic goals and objectives.” [Al10]. Similarly Osterwalder [Os04] defines a business model as a “[...] conceptual tool that contains a set of elements and their relationships and allows expressing a company's logic of
earning money. It is a description of the value a company offers to one or several segments of customers and the architecture of the firm and its network of partners for creating, marketing and delivering this value and relationship capital, in order to generate
profitable and sustainable revenue streams.” [Os04]. Osterwalder also developed the
business model canvas, which represents a visualization of the business model dimensions [OP10]. We decided to use this approach by Osterwalder, because of the systematic and easy use of the structure as well as the handy visualization. Figure 1 shows the
Osterwalder business model canvas which is based on nine different dimensions.
Key
Ressources
Partner
Key
Activities
Value
Proposition
Customer
Relationship
Customer
Channel
Cost Structure
Revenue Streams
Figure 1: Business model framework (based on [OP10])
2.2 Co-creation
During the last decades the perspective of value creation turned from a value-inexchange view where value for customers is embedded in products to a value-in-use
view where value for customers is generated during the value-generating processes
[Gr08]. This reflects the shift from a traditional goods-dominant logic with the focus on
the exchange of goods to a service-dominant logic focusing on the creation of value
[VL06].
According to this value is not created by buying products but by using them in a specific
context [GKW11]. This reflects renunciation from distinct roles of customers and producers towards a broad engagement of the customer in value creation [PR04].
244
This new perspective emphasizes on the understanding of the customer as part of valuecreation [ETG10][SP08]. From this point of view the customers can tailor the product or
service pursuant to their needs, which results in an enhanced value created [KMJ08].
This also implies that customers can be part of the value creation along the complete
value creating activities e.g. maybe from the development to the delivery of a product or
service by providing customer-specific knowledge [GKW11].
After a brief introduction into the basics of the business model and the co-creation, the
main problem of the existing business model construct will be outlined and a possible
solution will be introduced.
3 Service Business Models
3.1 Problem
The business model construct will be used in order to analyse, describe and evolve the
business model of products as well as services [ZB10]. Nevertheless, the consideration
of business models has a lack of service-specific aspects. One important gap can be
found in the structure of a business model. Co-creation, as a key characteristic of services, is not considered in most business model approaches [ZB11]. If it is considered,
like in the Osterwalder Business model canvas, then only with little impact on the whole
business model. For example Osterwalder limits co-creation to the customer relationship
dimension [OP10].
This aspect is emphasized by the graphical illustration of the business model (figure 1).
The business model canvas and the arrangement of the nine dimensions follow a goodsdominant logic view, where the dimensions are organized along a value chain. This
chain starts with partners and key resources as well as key activities leading to the value
proposition and at the end to the handling of the customer. This structure suggests that
neither customers can participate in the value creation nor partners are involved in customer-specific aspects of the business model.
3.2 Proposal for a solution
A possible way to address the described problem is to change the perspective on business models from a product to a service based point of view. To achieve this it is necessary that the framework permits a structure where the customer can be part of all aspects
of the business model. A possible way for this is to change the composition of the nine
blocks in Osterwalder´s business model canvas. The new structure is given in figure 2.
245
Customer
Cost
Structure
Key
Ressources
Key
Activities
Value
Proposition
Customer
Relationship
Channel
Revenue
Streams
Partner
Figure 2: Adapted business model canvas (based on [OP10]).
As figure 2 shows by encompassing the rest of the blocks with customers as well as
partners the deep integration into the value co-creation is represented. With this adaptation it is possible to reflect business models where a customer may have influence on
each dimension of the business model. In the following examples are given, which represent potential resources of co-creation.
In many cases of service provision there is a need for customer integration. This is due to
the fact that most services need to be provided individually for the customer. This leads
to a dynamic role of the customer in the service provision where the provider´s mission
is to help the customer to create value according to the specifications [GR11]. The value
creation is focusing on shared monetary benefits and revenue, for example, sharing of
financial gain and gaining joint contracts. Moreover, the relationship with the customer
can be designed based on the desired customer experience. Analogous to this the channel
has to be chosen. Further, customers have to be integrated into communication, sales and
distribution as well as into open innovation processes, for example, through social media. On the one hand, customers can provide resources like infrastructure or technology
and on the other hand they can influence the selection and use of resources. The service
provision is directly influenced by the customers and delivered by joint activities. Depending on the objectives of the service provision costs are borne by customers. Table 1
sums up the outlined impacts of co-creation in service business models.
Element
Cost Structure
Key Resources
Key Activities
Value Proposition
Customer Relationship
Channel
Revenue Streams
Customer perspective – Evidence of co-creation in business
model elements
Cost incurred / borne by customers
Customer providing resources
Customers influence on resource selection and use (e.g. selecting a
specific consultant or varying volume of demand)
Joint activities
Customers influence on activities
Specification of how service provider helps customers to create
value
Desired customer experience
Derived customer experience
Open innovation
Integration of customers in communication, sales and distribution
(e.g. through social media)
Shared monetary benefits / revenue (e.g. sharing of financial gain
and gaining joint contracts)
Table 1: Customer perspective –Evidence of co-creation in business model elements
246
In addition to the customer, partners can also affect the service business model and influence all dimensions. On the one hand, the provider can request help from partners, on
the other hand the provider can offer a platform for his partners to distribute own services and products. In this case the service provider holds an infrastructure that can be
used by partners to access customers. An example of this kind of business model is Apple Inc., which is selling devices with access to an app store where partners are able to
offer products to their customers. Therefore the main differences between partners and
customers can be found in the elements Value Proposition, Customer Relationship and
Channel, where partner support the provision of the service. The influence of the partners of the business model is shown in a table 2.
Element
Cost Structure
Key Resources
Key Activities
Value Proposition
Customer Relationship
Channel
Revenue Streams
Partner perspective –Evidence of co-creation in business model
elements
Cost incurred / borne by partners
Partner providing resources
Mutual influence on resource selection and use
Joint activities
Partner enhances co-creation of provider and vice versa
Provider opens customer relationship to partners and vice versa
Partner opens channels to customers for partners and vice versa
Shared monetary benefits / revenue
Table 2: Partner perspective –Evidence of co-creation in business model elements
4 Conclusion and Outlook
The paper shows how an adaptation of Osterwalder´s business model canvas can be
utilized to take co-creation into account of business modeling, which is a main shortcoming of recent business model approaches. Hence, the adapted model gives a framework
that is eligible to model service businesses, because one of the main characteristics of
them is supported considering customers as well as partners. This allows representing a
deep integration of the customers and partners in the value creation. According to the
understanding of business implied by service dominant logic this is a critical aspect and
is not applied in recent business model approaches. Furthermore with the adapted model
it is possible to analyze, describe and evolve business models with emphasis on services.
As this paper represents research-in-progress the described solution cannot be seen as a
complete solution for all paucities of recent business model approaches, but it gives an
implication how to deal with service specifics in the context of business modeling. The
next step on the research agenda addresses a refinement of the business model canvas
according to service-dominant logic. Further, a meta model for the service business
model will be abstracted. Beyond this it will be necessary to prove the adaptation by
specific examples in order to evaluate the adapted business model canvas.
247
References
[Al10]
Al-Debei, M.: The design and engineering of innovative mobile data services: An ontological Framework founded on business model thinking. Brunel University, 2010.
[CS06]
Chesbrough, H.; Spohrer, J.: A research manifesto for services science. In: Communications of the ACM, 49, 7, 2006; pp. 35-40.
[ETG10] Edvardsson, B.; Tronvoll, B.; Gruber, T.: Expanding understanding of service exchange
and value co-creation: a social construction approach. In: Journal of the Academy of
Marketing Science, 37, 2, 2010; pp. 327–339.
[GKW11]Gustafsson, A.; Kristensson, P.; Witell, L.: The Dimensions of Co-Creation and its Contribution to Market Success. In (Rhee, B. van der; Victorino, L. Eds.): Proc. 12th Int. Reseach Symposium on Service Excellence in Management, Ithaca, New York, 2011, Cayuga Press, Ithaca, New York, 2011; pp. 494-503.
[Gr08]
Grönroos, C.: Service logic revisited: who creates value? And who co-creates? In: European Business Review 20, 4, 2008; pp. 298-314.
[GR11] Grönroos, C.; Ravald, A.: Service as business logic: implications for value creation and
marketing. In: Journal of Service Management, 22, 1, 2011; pp. 5-22.
[KMJ08] Kristensson, P.; Matthing, J.; Johansson, N.: Key strategies for the successful involvement of customers in the co-creation of new technology-based services. In: International
Journal of Service Industry Management, 19, 4, 2008; pp. 474-491.
[Ne08]
Neely, A.: Exploring the financial consequences of the servitization of manufacturing.
In: Operations Management Research, 1, 2, 2008; pp.103-118.
[OP10]
Osterwalder, A.; Pigneur, Y.: Business Model Generation. John Wiley & Sons, Hoboken, 2010.
[OPT05] Osterwalder, A.; Pigneur, Y; Tucci, C.: Clarifying Business Models: Origins, Present
and Future of the Concept. In: Communications of the Association for Information Systems, 16, 1, 2005; pp. 1-25.
[PR04]
Prahalad, C K; Ramaswamy, Venkat: Co-creating unique value with customers. In:
Strategy & Leadership, 32, 3, 2004; pp. 4-9.
[Sp08]
Spohrer, J. et al.: The Service System is the Basic Abstraction of Service Science. In:
Proc. 41st Hawaii Int. Conf. on System Science, Big Island, 2008; pp. 104-114.
[VL06] Vargo, S.; Lusch, R.: Service-Dominant Logic: What It Is, What It Is Not, What It Might
Be. In (Vargo, S.; Lusch, R. eds.): The Service-Dominant Logic of Marketing: Dialog,
Debate, and Directions, M.E. Sharpe, Armont, New York, 2006; pp. 43-56.
[ZB10]
Zolnowski, A.; Böhmann, T.: Stand und Perspektiven der Modellierung von Geschäftsmodellen aus Sicht des Dienstleistungsmanagements. Proceedings des Workshops
Dienstleistungsmodellierung, In (Thomas, O.; Nüttgens, M. eds.): Dienstleistungsmodellierung. Heidelberg Physica, 2010; pp. 25-39.
[ZB11]
Zolnowski, A.; Böhmann, T.: Business modelling for services – Current state and research perspectives. In: AMCIS 2011 Proceedings – All Submissions. Paper 394, 2011.
248
GI-Edition Lecture Notes in Informatics
P-1
P-2
P-3
P-4
P-5
P-6
P-7
P-8
P-9
P-10
P-11
P-12
P-13
P-14
P-15
P-16
P-17
Gregor Engels, Andreas Oberweis, Albert
Zündorf (Hrsg.): Modellierung 2001.
Mikhail Godlevsky, Heinrich C. Mayr
(Hrsg.): Information Systems Technology
and its Applications, ISTA’2001.
Ana M. Moreno, Reind P. van de
Riet (Hrsg.): Applications of Natural
Lan-guage to Information Systems,
NLDB’2001.
H. Wörn, J. Mühling, C. Vahl, H.-P.
Meinzer (Hrsg.): Rechner- und sensorgestützte Chirurgie; Workshop des SFB
414.
Andy Schürr (Hg.): OMER – ObjectOriented Modeling of Embedded RealTime Systems.
Hans-Jürgen Appelrath, Rolf Beyer, Uwe
Marquardt, Heinrich C. Mayr, Claudia
Steinberger (Hrsg.): Unternehmen Hochschule, UH’2001.
Andy Evans, Robert France, Ana Moreira,
Bernhard Rumpe (Hrsg.): Practical UMLBased Rigorous Development Methods –
Countering or Integrating the extremists,
pUML’2001.
Reinhard Keil-Slawik, Johannes Magenheim (Hrsg.): Informatikunterricht und
Medienbildung, INFOS’2001.
Jan von Knop, Wilhelm Haverkamp
(Hrsg.): Innovative Anwendungen in
Kommunikationsnetzen, 15. DFN Arbeitstagung.
Mirjam Minor, Steffen Staab (Hrsg.): 1st
German Workshop on Experience Management: Sharing Experiences about the
Sharing Experience.
Michael Weber, Frank Kargl (Hrsg.):
Mobile Ad-Hoc Netzwerke, WMAN
2002.
Martin Glinz, Günther Müller-Luschnat
(Hrsg.): Modellierung 2002.
Jan von Knop, Peter Schirmbacher and
Viljan Mahni_ (Hrsg.): The Changing
Universities – The Role of Technology.
Robert Tolksdorf, Rainer Eckstein
(Hrsg.): XML-Technologien für das Semantic Web – XSW 2002.
Hans-Bernd Bludau, Andreas Koop
(Hrsg.): Mobile Computing in Medicine.
J. Felix Hampe, Gerhard Schwabe
(Hrsg.): Mobile and Collaborative Business 2002.
Jan von Knop, Wilhelm Haverkamp
(Hrsg.): Zukunft der Netze –Die Verletzbarkeit meistern, 16. DFN Arbeitstagung.
P-18
P-19
P-20
P-21
P-22
P-23
P-24
P-25
P-26
P-27
P-28
P-29
P-30
P-31
Elmar J. Sinz, Markus Plaha (Hrsg.):
Modellierung betrieblicher Informationssysteme – MobIS 2002.
Sigrid Schubert, Bernd Reusch, Norbert
Jesse (Hrsg.): Informatik bewegt – Informatik 2002 – 32. Jahrestagung der Gesellschaft für Informatik e.V. (GI) 30.Sept.-3.
Okt. 2002 in Dortmund.
Sigrid Schubert, Bernd Reusch, Norbert
Jesse (Hrsg.): Informatik bewegt – Informatik 2002 – 32. Jahrestagung der Gesellschaft für Informatik e.V. (GI) 30.Sept.-3.
Okt. 2002 in Dortmund (Ergänzungsband).
Jörg Desel, Mathias Weske (Hrsg.):
Promise 2002: Prozessorientierte Methoden und Werkzeuge für die Entwicklung
von Informationssystemen.
Sigrid Schubert, Johannes Magenheim,
Peter Hubwieser, Torsten Brinda (Hrsg.):
Forschungsbeiträge zur “Didaktik der
Informatik” – Theorie, Praxis, Evaluation.
Thorsten Spitta, Jens Borchers, Harry M.
Sneed (Hrsg.): Software Management
2002 – Fortschritt durch Beständigkeit
Rainer Eckstein, Robert Tolksdorf
(Hrsg.): XMIDX 2003 – XMLTechnologien für Middleware – Middleware für XML-Anwendungen
Key Pousttchi, Klaus Turowski (Hrsg.):
Mobile Commerce – Anwendungen und
Perspektiven – 3. Workshop Mobile
Commerce, Universität Augsburg,
04.02.2003
Gerhard Weikum, Harald Schöning,
Erhard Rahm (Hrsg.): BTW 2003: Datenbanksysteme für Business, Technologie
und Web
Michael Kroll, Hans-Gerd Lipinski, Kay
Melzer (Hrsg.): Mobiles Computing in
der Medizin
Ulrich Reimer, Andreas Abecker, Steffen
Staab, Gerd Stumme (Hrsg.): WM 2003:
Professionelles Wissensmanagement –
Er-fahrungen und Visionen
Antje Düsterhöft, Bernhard Thalheim
(Eds.): NLDB’2003: Natural Language
Processing and Information Systems
Mikhail Godlevsky, Stephen Liddle,
Heinrich C. Mayr (Eds.): Information
Systems Technology and its Applications
Arslan Brömme, Christoph Busch (Eds.):
BIOSIG 2003: Biometrics and Electronic
Signatures
P-32
P-33
P-34
P-35
P-36
P-37
P-38
P-39
P-40
P-41
P-42
P-43
P-44
P-45
P-46
P-47
Peter Hubwieser (Hrsg.): Informatische
Fachkonzepte im Unterricht – INFOS
2003
Andreas Geyer-Schulz, Alfred Taudes
(Hrsg.): Informationswirtschaft: Ein
Sektor mit Zukunft
Klaus Dittrich, Wolfgang König, Andreas
Oberweis, Kai Rannenberg, Wolfgang
Wahlster (Hrsg.): Informatik 2003 –
Innovative Informatikanwendungen
(Band 1)
Klaus Dittrich, Wolfgang König, Andreas
Oberweis, Kai Rannenberg, Wolfgang
Wahlster (Hrsg.): Informatik 2003 –
Innovative Informatikanwendungen
(Band 2)
Rüdiger Grimm, Hubert B. Keller, Kai
Rannenberg (Hrsg.): Informatik 2003 –
Mit Sicherheit Informatik
Arndt Bode, Jörg Desel, Sabine Rathmayer, Martin Wessner (Hrsg.): DeLFI
2003: e-Learning Fachtagung Informatik
E.J. Sinz, M. Plaha, P. Neckel (Hrsg.):
Modellierung betrieblicher Informationssysteme – MobIS 2003
Jens Nedon, Sandra Frings, Oliver Göbel
(Hrsg.): IT-Incident Management & ITForensics – IMF 2003
Michael Rebstock (Hrsg.): Modellierung
betrieblicher Informationssysteme – MobIS 2004
Uwe Brinkschulte, Jürgen Becker, Dietmar Fey, Karl-Erwin Großpietsch, Christian Hochberger, Erik Maehle, Thomas
Runkler (Edts.): ARCS 2004 – Organic
and Pervasive Computing
Key Pousttchi, Klaus Turowski (Hrsg.):
Mobile Economy – Transaktionen und
Prozesse, Anwendungen und Dienste
Birgitta König-Ries, Michael Klein,
Philipp Obreiter (Hrsg.): Persistance,
Scalability, Transactions – Database Mechanisms for Mobile Applications
Jan von Knop, Wilhelm Haverkamp, Eike
Jessen (Hrsg.): Security, E-Learning.
E-Services
Bernhard Rumpe, Wofgang Hesse
(Hrsg.): Modellierung 2004
Ulrich Flegel, Michael Meier (Hrsg.):
Detection of Intrusions of Malware &
Vulnerability Assessment
Alexander Prosser, Robert Krimmer
(Hrsg.): Electronic Voting in Europe –
Technology, Law, Politics and Society
P-48
P-49
P-50
P-51
P-52
P-53
P-54
P-55
P-56
P-57
P-58
P-59
P-60
P-61
P-62
P-63
Anatoly Doroshenko, Terry Halpin,
Stephen W. Liddle, Heinrich C. Mayr
(Hrsg.): Information Systems Technology
and its Applications
G. Schiefer, P. Wagner, M. Morgenstern,
U. Rickert (Hrsg.): Integration und Datensicherheit – Anforderungen, Konflikte und
Perspektiven
Peter Dadam, Manfred Reichert (Hrsg.):
INFORMATIK 2004 – Informatik verbindet (Band 1) Beiträge der 34. Jahrestagung der Gesellschaft für Informatik e.V.
(GI), 20.-24. September 2004 in Ulm
Peter Dadam, Manfred Reichert (Hrsg.):
INFORMATIK 2004 – Informatik verbindet (Band 2) Beiträge der 34. Jahrestagung der Gesellschaft für Informatik e.V.
(GI), 20.-24. September 2004 in Ulm
Gregor Engels, Silke Seehusen (Hrsg.):
DELFI 2004 – Tagungsband der 2.
e-Learning Fachtagung Informatik
Robert Giegerich, Jens Stoye (Hrsg.):
German Conference on Bioinformatics –
GCB 2004
Jens Borchers, Ralf Kneuper (Hrsg.):
Softwaremanagement 2004 – Outsourcing
und Integration
Jan von Knop, Wilhelm Haverkamp, Eike
Jessen (Hrsg.): E-Science und Grid Adhoc-Netze Medienintegration
Fernand Feltz, Andreas Oberweis, Benoit
Otjacques (Hrsg.): EMISA 2004 – Informationssysteme im E-Business und
E-Government
Klaus Turowski (Hrsg.): Architekturen,
Komponenten, Anwendungen
Sami Beydeda, Volker Gruhn, Johannes
Mayer, Ralf Reussner, Franz Schweiggert
(Hrsg.): Testing of Component-Based
Systems and Software Quality
J. Felix Hampe, Franz Lehner, Key
Pousttchi, Kai Ranneberg, Klaus
Turowski (Hrsg.): Mobile Business –
Processes, Platforms, Payments
Steffen Friedrich (Hrsg.): Unterrichtskonzepte für inforrmatische Bildung
Paul Müller, Reinhard Gotzhein, Jens B.
Schmitt (Hrsg.): Kommunikation in verteilten Systemen
Federrath, Hannes (Hrsg.): „Sicherheit
2005“ – Sicherheit – Schutz und Zuverlässigkeit
Roland Kaschek, Heinrich C. Mayr,
Stephen Liddle (Hrsg.): Information Systems – Technology and ist Applications
P-64
P-65
P-66
P-67
P-68
P-69
P-70
P-71
P-72
P-73
P-74
P-75
P-76
P-77
P-78
P-79
Peter Liggesmeyer, Klaus Pohl, Michael
Goedicke (Hrsg.): Software Engineering
2005
Gottfried Vossen, Frank Leymann, Peter
Lockemann, Wolffried Stucky (Hrsg.):
Datenbanksysteme in Business, Technologie und Web
Jörg M. Haake, Ulrike Lucke, Djamshid
Tavangarian (Hrsg.): DeLFI 2005: 3.
deutsche e-Learning Fachtagung Informatik
Armin B. Cremers, Rainer Manthey,
Peter Martini, Volker Steinhage (Hrsg.):
INFORMATIK 2005 – Informatik LIVE
(Band 1)
Armin B. Cremers, Rainer Manthey,
Peter Martini, Volker Steinhage (Hrsg.):
INFORMATIK 2005 – Informatik LIVE
(Band 2)
Robert Hirschfeld, Ryszard Kowalcyk,
Andreas Polze, Matthias Weske (Hrsg.):
NODe 2005, GSEM 2005
Klaus Turowski, Johannes-Maria Zaha
(Hrsg.): Component-oriented Enterprise
Application (COAE 2005)
Andrew Torda, Stefan Kurz, Matthias
Rarey (Hrsg.): German Conference on
Bioinformatics 2005
Klaus P. Jantke, Klaus-Peter Fähnrich,
Wolfgang S. Wittig (Hrsg.): Marktplatz
Internet: Von e-Learning bis e-Payment
Jan von Knop, Wilhelm Haverkamp, Eike
Jessen (Hrsg.): “Heute schon das Morgen
sehen“
Christopher Wolf, Stefan Lucks, Po-Wah
Yau (Hrsg.): WEWoRC 2005 – Western
European Workshop on Research in
Cryptology
Jörg Desel, Ulrich Frank (Hrsg.): Enterprise Modelling and Information Systems
Architecture
Thomas Kirste, Birgitta König-Riess, Key
Pousttchi, Klaus Turowski (Hrsg.): Mobile Informationssysteme – Potentiale,
Hindernisse, Einsatz
Jana Dittmann (Hrsg.): SICHERHEIT
2006
K.-O. Wenkel, P. Wagner, M. Morgenstern, K. Luzi, P. Eisermann (Hrsg.): Landund Ernährungswirtschaft im Wandel
Bettina Biel, Matthias Book, Volker
Gruhn (Hrsg.): Softwareengineering 2006
P-80
P-81
P-82
P-83
P-84
P-85
P-86
P-87
P-88
P-90
P-91
P-92
P-93
P-94
P-95
P-96
P-97
Mareike Schoop, Christian Huemer,
Michael Rebstock, Martin Bichler
(Hrsg.): Service-Oriented Electronic
Commerce
Wolfgang Karl, Jürgen Becker, KarlErwin Großpietsch, Christian Hochberger,
Erik Maehle (Hrsg.): ARCS´06
Heinrich C. Mayr, Ruth Breu (Hrsg.):
Modellierung 2006
Daniel Huson, Oliver Kohlbacher, Andrei
Lupas, Kay Nieselt and Andreas Zell
(eds.): German Conference on Bioinformatics
Dimitris Karagiannis, Heinrich C. Mayr,
(Hrsg.): Information Systems Technology
and its Applications
Witold Abramowicz, Heinrich C. Mayr,
(Hrsg.): Business Information Systems
Robert Krimmer (Ed.): Electronic Voting
2006
Max Mühlhäuser, Guido Rößling, Ralf
Steinmetz (Hrsg.): DELFI 2006: 4.
e-Learning Fachtagung Informatik
Robert Hirschfeld, Andreas Polze,
Ryszard Kowalczyk (Hrsg.): NODe 2006,
GSEM 2006
Joachim Schelp, Robert Winter, Ulrich
Frank, Bodo Rieger, Klaus Turowski
(Hrsg.): Integration, Informationslogistik
und Architektur
Henrik Stormer, Andreas Meier, Michael
Schumacher (Eds.): European Conference
on eHealth 2006
Fernand Feltz, Benoît Otjacques, Andreas
Oberweis, Nicolas Poussing (Eds.): AIM
2006
Christian Hochberger, Rüdiger Liskowsky
(Eds.): INFORMATIK 2006 – Informatik
für Menschen, Band 1
Christian Hochberger, Rüdiger Liskowsky
(Eds.): INFORMATIK 2006 – Informatik
für Menschen, Band 2
Matthias Weske, Markus Nüttgens (Eds.):
EMISA 2005: Methoden, Konzepte und
Technologien für die Entwicklung von
dienstbasierten Informationssystemen
Saartje Brockmans, Jürgen Jung, York
Sure (Eds.): Meta-Modelling and Ontologies
Oliver Göbel, Dirk Schadt, Sandra Frings,
Hardo Hase, Detlef Günther, Jens Nedon
(Eds.): IT-Incident Mangament & ITForensics – IMF 2006
P-98
P-99
P-100
P-101
P-102
P-103
P-104
P-105
P-106
P-107
P-108
P-109
P-110
P-111
Hans Brandt-Pook, Werner Simonsmeier
und Thorsten Spitta (Hrsg.): Beratung
in der Softwareentwicklung – Modelle,
Methoden, Best Practices
Andreas Schwill, Carsten Schulte, Marco
Thomas (Hrsg.): Didaktik der Informatik
Peter Forbrig, Günter Siegel, Markus
Schneider (Hrsg.): HDI 2006: Hochschuldidaktik der Informatik
Stefan Böttinger, Ludwig Theuvsen,
Susanne Rank, Marlies Morgenstern (Hrsg.):
Agrarinformatik im Spannungsfeld
zwischen Regionalisierung und globalen
Wertschöpfungsketten
Otto Spaniol (Eds.): Mobile Services and
Personalized Environments
Alfons Kemper, Harald Schöning, Thomas
Rose, Matthias Jarke, Thomas Seidl,
Christoph Quix, Christoph Brochhaus
(Hrsg.): Datenbanksysteme in Business,
Technologie und Web (BTW 2007)
Birgitta König-Ries, Franz Lehner,
Rainer Malaka, Can Türker (Hrsg.)
MMS 2007: Mobilität und mobile
Informationssysteme
Wolf-Gideon Bleek, Jörg Raasch,
Heinz Züllighoven (Hrsg.)
Software Engineering 2007
Wolf-Gideon Bleek, Henning Schwentner,
Heinz Züllighoven (Hrsg.)
Software Engineering 2007 –
Beiträge zu den Workshops
Heinrich C. Mayr,
Dimitris Karagiannis (eds.)
Information Systems
Technology and its Applications
Arslan Brömme, Christoph Busch,
Detlef Hühnlein (eds.)
BIOSIG 2007:
Biometrics and
Electronic Signatures
Rainer Koschke, Otthein Herzog, KarlHeinz Rödiger, Marc Ronthaler (Hrsg.)
INFORMATIK 2007
Informatik trifft Logistik
Band 1
Rainer Koschke, Otthein Herzog, KarlHeinz Rödiger, Marc Ronthaler (Hrsg.)
INFORMATIK 2007
Informatik trifft Logistik
Band 2
Christian Eibl, Johannes Magenheim,
Sigrid Schubert, Martin Wessner (Hrsg.)
DeLFI 2007:
5. e-Learning Fachtagung
Informatik
P-112
P-113
P-114
P-115
P-116
P-117
P-118
P-119
P-120
P-121
P-122
Sigrid Schubert (Hrsg.)
Didaktik der Informatik in
Theorie und Praxis
Sören Auer, Christian Bizer, Claudia
Müller, Anna V. Zhdanova (Eds.)
The Social Semantic Web 2007
Proceedings of the 1st Conference on
Social Semantic Web (CSSW)
Sandra Frings, Oliver Göbel, Detlef Günther,
Hardo G. Hase, Jens Nedon, Dirk Schadt,
Arslan Brömme (Eds.)
IMF2007 IT-incident
management & IT-forensics
Proceedings of the 3rd International
Conference on IT-Incident Management
& IT-Forensics
Claudia Falter, Alexander Schliep,
Joachim Selbig, Martin Vingron and
Dirk Walther (Eds.)
German conference on bioinformatics
GCB 2007
Witold Abramowicz, Leszek Maciszek
(Eds.)
Business Process and Services Computing
1st International Working Conference on
Business Process and Services Computing
BPSC 2007
Ryszard Kowalczyk (Ed.)
Grid service engineering and manegement
The 4th International Conference on Grid
Service Engineering and Management
GSEM 2007
Andreas Hein, Wilfried Thoben, HansJürgen Appelrath, Peter Jensch (Eds.)
European Conference on ehealth 2007
Manfred Reichert, Stefan Strecker, Klaus
Turowski (Eds.)
Enterprise Modelling and Information
Systems Architectures
Concepts and Applications
Adam Pawlak, Kurt Sandkuhl,
Wojciech Cholewa,
Leandro Soares Indrusiak (Eds.)
Coordination of Collaborative
Engineering - State of the Art and Future
Challenges
Korbinian Herrmann, Bernd Bruegge (Hrsg.)
Software Engineering 2008
Fachtagung des GI-Fachbereichs
Softwaretechnik
Walid Maalej, Bernd Bruegge (Hrsg.)
Software Engineering 2008 Workshopband
Fachtagung des GI-Fachbereichs
Softwaretechnik
P-123
P-124
P-125
P-126
P-127
P-128
P-129
P-130
P-131
P-132
Michael H. Breitner, Martin Breunig, Elgar
Fleisch, Ley Pousttchi, Klaus Turowski
(Hrsg.)
Mobile und Ubiquitäre
Informationssysteme – Technologien,
Prozesse, Marktfähigkeit
Proceedings zur 3. Konferenz Mobile und
Ubiquitäre Informationssysteme
(MMS 2008)
Wolfgang E. Nagel, Rolf Hoffmann,
Andreas Koch (Eds.)
9th Workshop on Parallel Systems and
Algorithms (PASA)
Workshop of the GI/ITG Speciel Interest
Groups PARS and PARVA
Rolf A.E. Müller, Hans-H. Sundermeier,
Ludwig Theuvsen, Stephanie Schütze,
Marlies Morgenstern (Hrsg.)
Unternehmens-IT:
Führungsinstrument oder
Verwaltungsbürde
Referate der 28. GIL Jahrestagung
Rainer Gimnich, Uwe Kaiser, Jochen
Quante, Andreas Winter (Hrsg.)
10th Workshop Software Reengineering
(WSR 2008)
Thomas Kühne, Wolfgang Reisig,
Friedrich Steimann (Hrsg.)
Modellierung 2008
Ammar Alkassar, Jörg Siekmann (Hrsg.)
Sicherheit 2008
Sicherheit, Schutz und Zuverlässigkeit
Beiträge der 4. Jahrestagung des
Fachbereichs Sicherheit der Gesellschaft
für Informatik e.V. (GI)
2.-4. April 2008
Saarbrücken, Germany
Wolfgang Hesse, Andreas Oberweis (Eds.)
Sigsand-Europe 2008
Proceedings of the Third AIS SIGSAND
European Symposium on Analysis,
Design, Use and Societal Impact of
Information Systems
Paul Müller, Bernhard Neumair,
Gabi Dreo Rodosek (Hrsg.)
1. DFN-Forum Kommunikationstechnologien Beiträge der Fachtagung
Robert Krimmer, Rüdiger Grimm (Eds.)
3rd International Conference on Electronic
Voting 2008
Co-organized by Council of Europe,
Gesellschaft für Informatik and E-Voting.
CC
Silke Seehusen, Ulrike Lucke,
Stefan Fischer (Hrsg.)
DeLFI 2008:
Die 6. e-Learning Fachtagung Informatik
P-133
P-134
P-135
P-136
P-137
P-138
P-139
P-140
P-141
P-142
P-143
Heinz-Gerd Hegering, Axel Lehmann,
Hans Jürgen Ohlbach, Christian
Scheideler (Hrsg.)
INFORMATIK 2008
Beherrschbare Systeme – dank Informatik
Band 1
Heinz-Gerd Hegering, Axel Lehmann,
Hans Jürgen Ohlbach, Christian
Scheideler (Hrsg.)
INFORMATIK 2008
Beherrschbare Systeme – dank Informatik
Band 2
Torsten Brinda, Michael Fothe,
Peter Hubwieser, Kirsten Schlüter (Hrsg.)
Didaktik der Informatik –
Aktuelle Forschungsergebnisse
Andreas Beyer, Michael Schroeder (Eds.)
German Conference on Bioinformatics
GCB 2008
Arslan Brömme, Christoph Busch, Detlef
Hühnlein (Eds.)
BIOSIG 2008: Biometrics and Electronic
Signatures
Barbara Dinter, Robert Winter, Peter
Chamoni, Norbert Gronau, Klaus
Turowski (Hrsg.)
Synergien durch Integration und
Informationslogistik
Proceedings zur DW2008
Georg Herzwurm, Martin Mikusz (Hrsg.)
Industrialisierung des SoftwareManagements
Fachtagung des GI-Fachausschusses
Management der Anwendungsentwicklung und -wartung im Fachbereich
Wirtschaftsinformatik
Oliver Göbel, Sandra Frings, Detlef
Günther, Jens Nedon, Dirk Schadt (Eds.)
IMF 2008 - IT Incident Management &
IT Forensics
Peter Loos, Markus Nüttgens,
Klaus Turowski, Dirk Werth (Hrsg.)
Modellierung betrieblicher Informationssysteme (MobIS 2008)
Modellierung zwischen SOA und
Compliance Management
R. Bill, P. Korduan, L. Theuvsen,
M. Morgenstern (Hrsg.)
Anforderungen an die Agrarinformatik
durch Globalisierung und
Klimaveränderung
Peter Liggesmeyer, Gregor Engels,
Jürgen Münch, Jörg Dörr,
Norman Riegel (Hrsg.)
Software Engineering 2009
Fachtagung des GI-Fachbereichs
Softwaretechnik
P-144
P-145
P-146
P-147
P-148
P-149
P-150
P-151
P-152
P-153
P-154
Johann-Christoph Freytag, Thomas Ruf,
Wolfgang Lehner, Gottfried Vossen
(Hrsg.)
Datenbanksysteme in Business,
Technologie und Web (BTW)
Knut Hinkelmann, Holger Wache (Eds.)
WM2009: 5th Conference on Professional
Knowledge Management
Markus Bick, Martin Breunig,
Hagen Höpfner (Hrsg.)
Mobile und Ubiquitäre
Informationssysteme – Entwicklung,
Implementierung und Anwendung
4. Konferenz Mobile und Ubiquitäre
Informationssysteme (MMS 2009)
Witold Abramowicz, Leszek Maciaszek,
Ryszard Kowalczyk, Andreas Speck (Eds.)
Business Process, Services Computing
and Intelligent Service Management
BPSC 2009 · ISM 2009 · YRW-MBP
2009
Christian Erfurth, Gerald Eichler,
Volkmar Schau (Eds.)
9th International Conference on Innovative
Internet Community Systems
I2CS 2009
Paul Müller, Bernhard Neumair,
Gabi Dreo Rodosek (Hrsg.)
2. DFN-Forum
Kommunikationstechnologien
Beiträge der Fachtagung
Jürgen Münch, Peter Liggesmeyer (Hrsg.)
Software Engineering
2009 - Workshopband
Armin Heinzl, Peter Dadam, Stefan Kirn,
Peter Lockemann (Eds.)
PRIMIUM
Process Innovation for
Enterprise Software
Jan Mendling, Stefanie Rinderle-Ma,
Werner Esswein (Eds.)
Enterprise Modelling and Information
Systems Architectures
Proceedings of the 3rd Int‘l Workshop
EMISA 2009
Andreas Schwill,
Nicolas Apostolopoulos (Hrsg.)
Lernen im Digitalen Zeitalter
DeLFI 2009 – Die 7. E-Learning
Fachtagung Informatik
Stefan Fischer, Erik Maehle
Rüdiger Reischuk (Hrsg.)
INFORMATIK 2009
Im Focus das Leben
P-155
P-156
P-157
P-158
P-159
P-160
P-161
P-162
P-163
P-164
Arslan Brömme, Christoph Busch,
Detlef Hühnlein (Eds.)
BIOSIG 2009:
Biometrics and Electronic Signatures
Proceedings of the Special Interest Group
on Biometrics and Electronic Signatures
Bernhard Koerber (Hrsg.)
Zukunft braucht Herkunft
25 Jahre »INFOS – Informatik und
Schule«
Ivo Grosse, Steffen Neumann,
Stefan Posch, Falk Schreiber,
Peter Stadler (Eds.)
German Conference on Bioinformatics
2009
W. Claupein, L. Theuvsen, A. Kämpf,
M. Morgenstern (Hrsg.)
Precision Agriculture
Reloaded – Informationsgestützte
Landwirtschaft
Gregor Engels, Markus Luckey,
Wilhelm Schäfer (Hrsg.)
Software Engineering 2010
Gregor Engels, Markus Luckey,
Alexander Pretschner, Ralf Reussner
(Hrsg.)
Software Engineering 2010 –
Workshopband
(inkl. Doktorandensymposium)
Gregor Engels, Dimitris Karagiannis
Heinrich C. Mayr (Hrsg.)
Modellierung 2010
Maria A. Wimmer, Uwe Brinkhoff,
Siegfried Kaiser, Dagmar LückSchneider, Erich Schweighofer,
Andreas Wiebe (Hrsg.)
Vernetzte IT für einen effektiven Staat
Gemeinsame Fachtagung
Verwaltungsinformatik (FTVI) und
Fachtagung Rechtsinformatik (FTRI) 2010
Markus Bick, Stefan Eulgem,
Elgar Fleisch, J. Felix Hampe,
Birgitta König-Ries, Franz Lehner,
Key Pousttchi, Kai Rannenberg (Hrsg.)
Mobile und Ubiquitäre
Informationssysteme
Technologien, Anwendungen und
Dienste zur Unterstützung von mobiler
Kollaboration
Arslan Brömme, Christoph Busch (Eds.)
BIOSIG 2010: Biometrics and Electronic
Signatures Proceedings of the Special
Interest Group on Biometrics and
Electronic Signatures
P-165
P-166
P-167
P-168
P-169
P-170
P-171
P-172
P-173
P-174
Gerald Eichler, Peter Kropf,
Ulrike Lechner, Phayung Meesad,
Herwig Unger (Eds.)
10th International Conference on
Innovative Internet Community Systems
(I2CS) – Jubilee Edition 2010 –
Paul Müller, Bernhard Neumair,
Gabi Dreo Rodosek (Hrsg.)
3. DFN-Forum Kommunikationstechnologien
Beiträge der Fachtagung
Robert Krimmer, Rüdiger Grimm (Eds.)
4th International Conference on
Electronic Voting 2010
co-organized by the Council of Europe,
Gesellschaft für Informatik and
E-Voting.CC
Ira Diethelm, Christina Dörge,
Claudia Hildebrandt,
Carsten Schulte (Hrsg.)
Didaktik der Informatik
Möglichkeiten empirischer
Forschungsmethoden und Perspektiven
der Fachdidaktik
Michael Kerres, Nadine Ojstersek
Ulrik Schroeder, Ulrich Hoppe (Hrsg.)
DeLFI 2010 - 8. Tagung
der Fachgruppe E-Learning
der Gesellschaft für Informatik e.V.
Felix C. Freiling (Hrsg.)
Sicherheit 2010
Sicherheit, Schutz und Zuverlässigkeit
Werner Esswein, Klaus Turowski,
Martin Juhrisch (Hrsg.)
Modellierung betrieblicher
Informationssysteme (MobIS 2010)
Modellgestütztes Management
Stefan Klink, Agnes Koschmider
Marco Mevius, Andreas Oberweis (Hrsg.)
EMISA 2010
Einflussfaktoren auf die Entwicklung
flexibler, integrierter Informationssysteme
Beiträge des Workshops
der GI-Fachgruppe EMISA
(Entwicklungsmethoden für Informationssysteme und deren Anwendung)
Dietmar Schomburg,
Andreas Grote (Eds.)
German Conference on Bioinformatics
2010
Arslan Brömme, Torsten Eymann,
Detlef Hühnlein, Heiko Roßnagel,
Paul Schmücker (Hrsg.)
perspeGKtive 2010
Workshop „Innovative und sichere
Informationstechnologie für das
Gesundheitswesen von morgen“
P-175
P-176
P-177
P-178
P-179
P-180
P-181
P-182
P-183
P-184
Klaus-Peter Fähnrich,
Bogdan Franczyk (Hrsg.)
INFORMATIK 2010
Service Science – Neue Perspektiven für
die Informatik
Band 1
Klaus-Peter Fähnrich,
Bogdan Franczyk (Hrsg.)
INFORMATIK 2010
Service Science – Neue Perspektiven für
die Informatik
Band 2
Witold Abramowicz, Rainer Alt,
Klaus-Peter Fähnrich, Bogdan Franczyk,
Leszek A. Maciaszek (Eds.)
INFORMATIK 2010
Business Process and Service Science –
Proceedings of ISSS and BPSC
Wolfram Pietsch, Benedikt Krams (Hrsg.)
Vom Projekt zum Produkt
Fachtagung des GIFachausschusses Management der
Anwendungsentwicklung und -wartung
im Fachbereich Wirtschafts-informatik
(WI-MAW), Aachen, 2010
Stefan Gruner, Bernhard Rumpe (Eds.)
FM+AM`2010
Second International Workshop on
Formal Methods and Agile Methods
Theo Härder, Wolfgang Lehner,
Bernhard Mitschang, Harald Schöning,
Holger Schwarz (Hrsg.)
Datenbanksysteme für Business,
Technologie und Web (BTW)
14. Fachtagung des GI-Fachbereichs
„Datenbanken und Informationssysteme“
(DBIS)
Michael Clasen, Otto Schätzel,
Brigitte Theuvsen (Hrsg.)
Qualität und Effizienz durch
informationsgestützte Landwirtschaft,
Fokus: Moderne Weinwirtschaft
Ronald Maier (Hrsg.)
6th Conference on Professional
Knowledge Management
From Knowledge to Action
Ralf Reussner, Matthias Grund, Andreas
Oberweis, Walter Tichy (Hrsg.)
Software Engineering 2011
Fachtagung des GI-Fachbereichs
Softwaretechnik
Ralf Reussner, Alexander Pretschner,
Stefan Jähnichen (Hrsg.)
Software Engineering 2011
Workshopband
(inkl. Doktorandensymposium)
P-185
P-186
P-187
P-188
P-189
P-190
Hagen Höpfner, Günther Specht,
Thomas Ritz, Christian Bunse (Hrsg.)
MMS 2011: Mobile und ubiquitäre
Informationssysteme Proceedings zur
6. Konferenz Mobile und Ubiquitäre
Informationssysteme (MMS 2011)
Gerald Eichler, Axel Küpper,
Volkmar Schau, Hacène Fouchal,
Herwig Unger (Eds.)
11th International Conference on
Innovative Internet Community Systems
(I2CS)
Paul Müller, Bernhard Neumair,
Gabi Dreo Rodosek (Hrsg.)
4. DFN-Forum Kommunikationstechnologien, Beiträge der Fachtagung
20. Juni bis 21. Juni 2011 Bonn
Holger Rohland, Andrea Kienle,
Steffen Friedrich (Hrsg.)
DeLFI 2011 – Die 9. e-Learning
Fachtagung Informatik
der Gesellschaft für Informatik e.V.
5.–8. September 2011, Dresden
Thomas, Marco (Hrsg.)
Informatik in Bildung und Beruf
INFOS 2011
14. GI-Fachtagung Informatik und Schule
Markus Nüttgens, Oliver Thomas,
Barbara Weber (Eds.)
Enterprise Modelling and Information
Systems Architectures (EMISA 2011)
The titles can be purchased at:
Köllen Druck + Verlag GmbH
Ernst-Robert-Curtius-Str. 14 · D-53117 Bonn
Fax: +49 (0)228/9898222
E-Mail: druckverlag@koellen.de