You are on page 1of 215

Page 2

CHAPTER 1: INTRODUCTION TO INFORMATION SYSTEMS

1.1 What is an Information System?

Most of us experience the effect of different systems every day - getting to work
or school using a transport system, for example, or making a call using the telephone
system. Computerised systems have transformed the world in countless
ways, not least in increased connectivity between people, as provided by
the Internet and the World Wide Web, for example.

To many people, information systems means computers -- and computers are


indeed part of the world of most information systems professionals. But there is more to
the story than simple technology.

Information

Information is the universal commodity, available to all -- but information adds


value only when it meets a real need and when it is acquired, organized and
disseminated in a systematic fashion".

Consider the following diagramin figure 1.1:

Data
Context
Information

Context

Knowledge

Figure 1.1

This suggests that there is a process by which Information and then Knowledge is
produced, somewhat analogous to a manufacturing process whereby a raw material, in
this case ‘Data’, is refined and worked on to produce an end-product. But what are the
components, which convert data into information?

Copyright © Genetic Computer School, 2002


Page 3

Data and information are an important part of a computer system. A computer exists to
convert data into information.

Data are unprocessed, raw facts, figures, statistics, concepts or instructions. To be


useful, they must be converted into information.

Information is processed data. Data have been analyzed, organized and


summarized to create useful information. Information is data that has been processed so
that an action or decision can be made from it. A computer exists to create this useful
information by converting data into information through a process called data
processing.

Data processing is the process of manipulation, organization and analysis that is


used to convert data into information.

Data and information in a computer system are organized using a directory


structure and stored in files on the storage medium in use.

Data becomes information when it is used for a purpose, which is of value to the
individual or organization. Take, as an example, the following list into which animals
are divided:

a. Those that belong to the Emperor

b. Embalmed ones

c. Those that are trained

d. Suckling pigs

e. Mermaids

f. Fabulous ones

g. Those that tremble as if they were mad

Copyright © Genetic Computer School, 2002


Page 4

As outsiders, the categorization of animals under these headings has no


meaning or value whatsoever but in the context of the ancient Chinese Court where
they were used, it presumably provided valuable information by which the
Emperor managed his livestock portfolio!

The above example (figure 1.1) is rather extreme but to be an effective


information manager, an understanding of the purpose or 'context' in which
information is to be used is essential. That purpose may be to assist the effective
management of the organisation or it may have a wider purpose, for example, to
produce information which can be 'sold' either as a product or a service. This is an
important way of generating profits and, indeed, information related services
account for a higher, percentage of Gross Domestic Product in the U.K. than the
manufacturing sector.

In the same way that one manufacturer's product is more attractive to a


customer because it has features which suit their specific requirements, so too,
information will have characteristics which add value, depending upon the purpose
for which it is required. Some of these are listed below:

Quantity The amount of information provided should be adequate for the


purpose - not so much that the key information is lost, or so
little that it does not present a complete picture.
Suitability It should be appropriate to the skills and competencies of the
manager who will use it and in a form which makes it 'user
friendly'.
Scope The breadth of information supplied will be in accord with the
purpose for which it is to be used, for example, a population
forecast will use census statistics over several decades.
Relevance The subject matter which the information covers is the same as
that which the manager is addressing.
Accuracy As accurate as possible but, in some circumstances, not at the
expense of timeliness - sometimes it's better to be 90% accurate

Copyright © Genetic Computer School, 2002


Page 5

than 100% out of date.


Timeliness It should be available when required.
Compatibility The information is based upon standards which also apply to
other information systems, for example, the accounting year as
opposed to the calendar year.
Presentation The information is presented in an appropriate style, for
example, high quality printing and graphics in the case of an
Annual Report.

Managing information is primarily a question of assessing the context and


purpose for which it is to be used; and then deciding the relative importance of
each of the above. For example, annual financial reports need to be both accurate
and presented in a form that makes them intelligible to non-financial personnel. In
the case of internal budgetary statements, presentation may be less important but
timeliness increasingly critical if potential overspends are to be dealt with quickly
and effectively.

Information Systems

Information systems collect, process, store, and distribute information so that


it can be used by people. The information may be about people, places, things, or
events inside an organization or in the environment that surrounds it. People use
the information to make decisions, to keep track of resources, and to plan for the
future. Information has attributes of accuracy, credibility, and timeliness - old
news is not news!

An information system can be defined technically as a set of interrelated


components that collect (or retrieve), process, store, and distribute information to
support decision making and control in an organization. In addition to supporting
decision making, coordination, and control, information systems may also help
managers and workers analyze problems, visualize complex subjects, and create
new products.

Copyright © Genetic Computer School, 2002


Page 6

Information systems contain information about significant people, places, and


things with the organization or in the environment surrounding it. By information
we mean data that have been shaped into a form that is meaningful and useful to
human beings. Data, in contrast, are streams of raw facts representing events
occurring in organizations or the physical environment before they have been
organized and arranged into a form that people can understand and use.

Three activities in an information system produce the information


organizations need for making decisions, controlling operations, analyzing
problems, and creating new products or services. These activities are input,
processing, and output. Input captures or collects raw data from within the
organization or from its external environment. Processing converts this raw input
into a more meaningful form. Output transfers the processed information to the
people or an activities where it will be used. Information systems also require
feedback, which is output that is returned to appropriate members of the
organization to help them evaluate or correct the input stage.

Formal information systems can be either computer-based or manual. Manual


systems use paper and pencil technology. These manual systems serve important
needs. Computer-based information systems (CBIS), in contrast, rely on computer
hardware and software technology to process and disseminate information. From
this point on, when we use the term information systems we will be referring to
computer-based information systems -- formal organizational systems that rely on
computer technology.

Although computer-based information systems use computer technology to


process raw data into meaningful information, there is a sharp distinction between
a computer and computer program on the one hand, and an information system on
the other. Electronic computers and related software programs are the technical
foundation, the tools and materials, of modern information systems. Computers
provide the equipment for storing and processing information. Computer programs,
or software, are sets of operating instructions that direct and control computer

Copyright © Genetic Computer School, 2002


Page 7

processing. Knowing how computers and computer programs work is important in


designing solutions to organizational problems, but computers are only part of an
information systems.

Housing provides an appropriate analogy. Houses are built with hammers,


nails, and wood, but these do not make a house. The architecture, design, setting,
landscaping, and all of the decisions that lead to the creation of these features are
part of the house and are crucial for finding a solution to the problem of putting a
roof over one's head. Computers and programs are the hammer, nails, and lumber
of CBIS, but alone they can not produce the information a particular organization
needs. To understand information systems, one must understand the problems they
are designed to solve, their architectural and design elements, and the
organizational processes that lead to these solutions. Today's managers must
combine computer literacy with information systems literacy.

1.2 Computer Information System

The Computer Information System is an ensemble of hardware and software


applications. The Information system is an analysis of a problem domain. It is
"information oriented". For example, a banking compensation system or a telecommunication
protocol are "information systems" in our sense. Their definition does not imply
the use of specific software or hardware packages.

Today the computer information system is called by a growing number


"Information system" - maybe they find that it sounds smarter! So that company
hires Oracle specialists for their "information system"!

Whatever the name, the flavor is the same. Please note that in our definition
information systems are not platform specific (hardware, software, language etc...).
They might be called also domain systems, abstract systems, general systems etc. The
interesting fact is that these systems may be seen also as "knowledge objects", one of
the two views related to knowledge management. This view enables the way to
"conceptual objects reuse".

Copyright © Genetic Computer School, 2002


Page 8

The Computer Information System is derived from the Information System by


taking into account the platform specifics.

1.3 Why does information have to be managed?

♦ There is so much more of it these days!

The sheer volume of information available through advances in computing and


communications means that managers have to be clearly focused on what information
is important to them and what is not. Recent articles in newspapers have commented
on the problems associated with too much information and new medical conditions
such as 'information fatigue' and 'information overload’ have started to enter our
vocabulary.

The latest lap-top computers can hold and process vast amounts of data but the capacity
of our 'neck-top' computer has not altered over the last few hundred years!

♦ It is a valuable asset in most organizations

Information-based organizations now generate more wealth than manufacturing


industries in the U.K. In some organizations the only product they sell is information -
and that is often purchased by other organizations which add further value to the
information before selling it on. It is essential, therefore, that such a valuable product
is managed efficiently.

♦ It can provide a competitive advantage in the market place

Information facilitates chance

Exposure to new ideas and concepts from outside the business can stimulate managers to
adopt and adapt these concepts in their own business environment. There is a danger for
managers, who stay in one job with one company for a long time, that they become insular in
their approach and resistant to change. The working environment, as we move into the 21st
Century, requires a workforce that is multi-skilled and able to adapt and keep up to date with
change. There is much that organizations themselves can do to ensure that their employees have
access to new information and learning opportunities. It is also the case that employees will

Copyright © Genetic Computer School, 2002


Page 9

have to take greater charge of their own personal development in the new organization
structures being created in the Information Age.

♦ Information enables managers to manage effectively

If managers do not have the right information about how the business is
performing then they are unlikely to be able to make the right decisions (or are more
likely to, make the wrong ones).

Managers need:

• Essential decision making information in a usable form

• Information when they want it (at a time that suits their workflow)

• Real-time information (so decisions are timely and relate directly to what is
happening)

• To see trends developing (such as cost or time blowouts, teams not


performing)

• Information to manage cross functional teams doing many projects and


tasks

• To be notified of any changes that are likely to affect them

1.4 Roles people play in an information environment

Applications support organizational goals and such goals are held by those for
whom the success of the firm is critical; for this discussion, they are called stakeholders.
This group includes boards of directors, shareholders, unions, communities, and
suppliers. The goals represent desirable conclusions for clients. Clients may be inside or
outside the organization. The classes of individuals that have direct interaction with the
application are called direct users. Assisting these role players are developers, who may
work for an organization, a client or a user, and they may be direct employees or
contracted specialists.

Copyright © Genetic Computer School, 2002


Page 10

CHAPTER 2: FUNDAMENTAL TYPES OF INFORMATION SYSTEM

2.1 Transaction Processing Systems (TPS)

Introduction

A transaction Processing System is an exceptionally powerful and flexible


transaction processing environment for both mobile and fixed-base data collection
equipment. It provides the sophisticated connectivity needed in today's complex
industrial, warehousing, distribution, and other environments.

Transaction Processing System has numerous features important in creating


systems and solutions for demanding transaction processing requirements. With
Transaction Processing System, you can:

• Develop high-performance data collection and bar code printing systems


across client/server networks.

• Revitalize legacy software and hardware in a state-of-the-art transaction


processing and bar code environment Interface with different types of host and
network computers at the same time.

• Access multiple local and remote distributed databases, concurrently and


simultaneously.

• Select and use terminals and printers from multiple leading manufacturers, all
in one system.

• Combine 5250 and 3270 intelligent terminal emulation with data collection
functionality using a GUI development environment that involves no host code
rewrites.

• Run the same transactions on your radio-frequency devices that you run on
fixed devices and PCs, including intelligent emulation.

Copyright © Genetic Computer School, 2002


Page 11

• Select from a variety of automatic, dynamic servers and data backup schemes,
which are available to ensure minimum downtime and maximum productivity.

Transaction Processing System's modular architecture (refer to figure 2.1) allows


you to choose the functionality that you need - you can always add on additional
functionality as your requirements expand.

Transaction processing systems [TPS] support the organization's daily activities


and maintain the majority of the organization's internal data. These systems usually
employ simple but highly repetitive processes but require the capability to deal with
high volumes of transactions with great accuracy and high security. In the past,
transactions were completed using batch processing, but today most transactions are
completed immediately using on-line transaction processing [OLTP] systems.

Figure 2.1 Transaction Processing Systems Architecture

To implement on-line transaction processing, master files containing key


information about important business entities are placed on hard disk drives where they
are directly accessible. The transaction files containing information about activity
concerning these business entities, such as orders placed by customers, are also held in
on-line files until they are no longer needed for everyday transaction processing
activity.. This assures that the transaction data is available to all applications, and that
all data is kept up-to-the-minute.

Copyright © Genetic Computer School, 2002


Page 12

Users with the proper database permissions may download data from the database
to another location and upload changes to these on-line files as soon as changes are
required. When the transaction files are no longer being "actively" used, they are not
deleted. Instead they are archived in a special data storage facility called a data
warehouse.

A Transaction Processing System (TPS) is designed for Operating Management. It


is also called Electronic Data Processing (EDP). It has a limited area of functionality,
and is designed for one area of the business.

It uses only data that is internal to the company. There are some possible
exceptions, however. Inter -branch banking uses TPS as well as direct deposit of
employee's pay from various companies and government institutions.

Reports and other outputs from Transaction Processing Systems

The outputs of TPS provide operational details, summary reports, and exception
reports, which help supervise and control routine operations. Exception reporting
provides a feedback loop to the operational manager about any unusual or unexpected
activity that may require attention. Transaction processing systems maybe divided into
major functional areas. Accounting provides systems to handle general ledger, payroll,
and accounts. Marketing has TP systems to handle sales, promotion and advertising
activity. Manufacturing requires transaction processing systems and operational
feedback from production processes. Manufacturing reporting systems must often be
designed as real-time systems to allow operational managers and supervisors to closely
monitor on-going operations. Human resources needs systems to support its daily
operations in recruiting, maintaining employee records, providing benefits, and
monitoring occupational conditions. Finance requires maintaining information on cash
reserves and investment holdings, and the monitoring of changes to tax and fiscal
regulations.

Many systems, such as order processing, may involve processes that cross
funtional boundaries in the enterprise. For example, an order may concern sales,

Copyright © Genetic Computer School, 2002


Page 13

manufacturing, distribution, and accounting. Transaction processing systems may


depend upon each other in intricate ways that are not always apparent to the users of
these systems.

When transaction data grows old, it can still be of value to the organization.
However, it is not used frequently and takes up valuable disk space, so such data is
often moved to other lower cost storage areas. In the past, archived data was difficult to
access because the data archives were not well organized. Today, such archives are
placed in data warehouses, which are data storage facilities that are managed by special
database programs designed for that purpose. Analysts can then make use of on-line
analytical processing [OLAP] tools to investigate either the current transaction data or
the data that has been retired to the data warehouse facility. Using the company's data
archives to determine trends and to extract other useful information is often referred to
as data mining.

2.2 Management Information Systems (MIS)

MIS stands for Management Information Systems. MIS is a branch of


Management that involves the development, maintenance and understanding of
Computer Information Systems from a business perspective. It is difficult to say exactly
what this field of management entails because, like the computers and technology the
field involves a rapidly changing and expanding area. Much like accounting, marketing
and financing, the Management of Information Systems has become a crucial function
in the operation of modern day business. Now and in the foreseeable future, the success
of a company will increasingly depend on the quality of its Information Systems and
technologies. MIS is a unique field in that it is able to combine the uses of technology
while at the same time having an appreciation for the business environment as a whole.

A widely recognized device for organizing types of management information


systems uses the MIS Framework. This framework is based upon two dimensions, the
level of managerial responsibility and the degree of problem structure. Structured
problem solving is handled by a variety of reporting levels. Detailed operational reports
are produced by transaction processing systems [TPS]. These in turn provide data to more

Copyright © Genetic Computer School, 2002


Page 14

highly summarized managerial reports used at higher levels. Management information


systems [MIS] produce standard "MIS" reports in printed and on-line "soft copy'
versions to assist middle management in their tactical decision making roles involving
the allocation of resources and the oversight of company operations. Executives at the
top rung of the corporation may have executive information systems [EIS], also called
executive support systems [ESS] to supply their special needs for structured reporting.

For less structured problems, and most often at higher management levels,
decision support systems play a role in modeling from a wide variety of sources, both
internal and external. Such problems are often called unstructured problems although
that term can be misleading. The structuredness of a problem must be viewed as a
continuum running from the fairly unstructured to the highly structured types of
problems. No problem that can be analyzed by any technique can be considered as
totally unstructured.

Management Information Systems (M.I.S.) is a career area which focuses on two


related topics:

• Organizations - the business processes, and people as problem solvers and


decision-makers.

• Technology - current information technology so vital for any organization.

The MIS field deals with all the information and problem solving activity of a
modern, successful organization. The MIS discipline brings together the various
business areas, computer science, and quantitative analysis techniques. This program
provides the theory and methodology to analyze, design, implement, and manage an
organization's information technology and systems.

Management Information Systems (MIS) is a valuable new aspect of every


business organization. Basically, MIS is looking at what types of computer systems
organizations need and adapting the systems for their needs. The MIS field revolves
around the fact that information and knowledge is one of the most valuable business

Copyright © Genetic Computer School, 2002


Page 15

assets. Without proper information channels, a business would lose its advantage over
other companies.

In the past, many people were turned off by the idea of using a personal computer
in the everyday workplace. Many more people are starting to get comfortable around
computers. New programs are always emerging that combine powerful applications
with ease of usage. For instance, Microsoft, easily the biggest computer software
provider in the world, has numerous software programs that even a beginner can tackle
without much commotion. MIS specialists must take and apply their knowledge of both
computers and people to the common business. They need to design programs that
appeal to and motivate employees. MIS specialists must learn an overview of the
organizational, strategic, and technical issues surrounding the management of
information in businesses today. By building a solid base at businesses everywhere,
MIS specialists help prepare future managers to manage information as a resource and
to identify opportunities for using information as a competitive advantage.

Management Information Systems (MIS) are used in businesses to help make


decisions. They are an implementation of computer based processing and/or manual
procedures yielding useful and timely information for decision making.

A MIS is composed of procedures. In a MIS, procedures are sequences of steps to


co-ordinate processing. Some are performed by staff and some by computer specialists.
A business has many procedures, we refer to four of the common ones below:

• Payroll

• Personnel

• Accounting

• Inventory

A Management Information System (MIS) is designed to help Middle


Management decisions.

Copyright © Genetic Computer School, 2002


Page 16

It is an information reporting system, which receives data from all sections of the
business. The system is connected to a more complex network than a Transaction
Processing System.

A MIS uses internal data as well as some external data.

2.3 Decision Support Systems (DSS)

The group of tools and techniques, together with their definitions and descriptions
of their interrelations, that assist in the specification and recording of an analyst's
judgments about evidence, and inferences going into a hazard identification decision.

A DSS is a tool (e.g. computer programme) or process (such as a best


management practices procedure), whose utility is related to its ability to support
decision-making in the real world situation and not to provide the solution itself by
representing a system involving those real-world variables (physical, economic and social)
and their interrelationships, whose complexities can be simplified for inclusion in the
tool or process, where these variables and interrelationships can be customized to
comply with individual situations and, where these variables and inter-relationships can
be customized and/or interfaced with individual decision-makers, more particularly
with their objectives, resources and constraints.

Decision Support Systems can be differentiated from other types of information


system in that they are directed towards use by users who are skilled in their subject
area and who are supported rather than placed by the use of the computer. Decision
support system users have a good knowledge of the problems area of the system and
therefore need to be able to interact effectively with a flexible system which provides
the type of support required.

Decision Support Systems (DSS) is a dynamic and rapidly changing field which
touches on a wide range of computing topics. Decision Support Systems have been
defined as:

Copyright © Genetic Computer School, 2002


Page 17

Computer Based Systems are systems that help decision makers confront ill
structured problems through direct interaction with data and analysis models.

The emphasis is on problem solving tasks, which are semi-structured, ie. they
combine human judgement with the use of computing tools and techniques. DSS do not
replace managerial judgement but rather provide support for decision making - the final
agent remains the human.

Computer applications for management support are increasing and the availability
of microcomputers has dramatically increased the number of systems on managers
desks.

2.4 Office Automation Systems (OAS)

Office Automation involves the planned application of integrated information


handling tools and methods to improve the productivity of people in office operations.
Although the handling of information by office people is the focus of this new
technology, other aspects of the office will be affected. These include factors such as
the organization of functions and lines of reporting, training for new methods, work
space design, travel patterns, branch office location, home vs office work, hours of
work, employee morale, and job classifications. Organizations that harness office
automation products will need to deal with many more than just technological issues.

About 22% of the US work force is now in the office, with that percentage rising.
Labor costs account for about 70% of the total office costs in our economy and salary
costs are increasing about 6% each year. During the past 15 years there has been
relatively little increase in productivity of the office work force, in contras with the
manufacturing sector where the average productivity has more than doubled. The cost
of new technology aimed at increasing the productivity of office workers is going
down, while the capabilities of office automation systems have been rapidly increasing.

Office automation will impact industry and government organizations in very


significant ways with both COST DISPLACEMENT and VALUE-ADDED results.

Copyright © Genetic Computer School, 2002


Page 18

These two terms are now being used by people who are considering the potential
payoffs of office automation to their organizations.

Cost displacement applications have the objective of achieving overall reductions


in support staff costs or of increasing work volumes without adding support staff. Such
applications typically center around WORD PROCESSING and provide the base for the
more advanced value-added applications.

Value-added applications are viewed as being directed toward improving


managerial and professional staff productivity (and effectiveness) through use of more
integrated office automation systems that can directly affect their work.

The value-added approach deals with far more fundamental issues than the
replacement of some support staff positions with word processing pools. Its focus upon
individuals and groups of managers and professionals as targets for productivity
improvement brings with it opportunities for significant increases in organizational
effectiveness and major cost benefits in the largest segment of the office cost spectrum.

For any organization, management choices at many levels will affect the balance
between reducing total office costs (cost displacement) or increasing the total office
effectiveness (value-added effects).

Word processing applications have, until recently, been equated with the term
office automation. It is interesting to note that, on the average, typing tasks comprise
only 30% of the secretaries' and typists' work -- and thus account for only about 2% of
the total office salaries. The next few years will see a very rapid growth in the
introduction of advanced technology into offices and in applications with more impact
on managers and other non-clerical people, bringing with it broadened perceptions of
what office automation really includes.

Office Automation is likely to become one of the fastest-growing and most


significant new industries of the century. It will apply electronic technology to a broad
new set of applications and bring significant change to many of the ways in which

Copyright © Genetic Computer School, 2002


Page 19

people and organizations work -AUGMENTING their capabilities and increasing both
the quantity and quality of their contributions.

2.5 Executive Support Systems (ESS)

An enterprise management information system encompasses the information flows


in an entire organization. The applications for MIS fall into five categories; localized
applications, interdepartmental applications, business process redesign, business
network redesign, and business scope redefinition.

Interdepartmental information systems integrate the activities of different


departments into a single business system that produces appropriate coordinated
responses to the enterprise's environment. Both localized and interdepartmental systems
involve little redesign of the underlying business processes; they support the existing
processes.

Business process redesign requires the nature of business processes to be


redesigned as the information system development progresses; it becomes a vehicle for
change and improvement.

Business network redesign concerns itself with how multiple enterprises work
together; consideration is given as to how information is structured and handled as it
crosses organisational boundaries.

Business scope redefinition consists of applications that change the nature of the
business. Interdepartmental information systems use a shared or corporate database
facility. Procedures and practices are consistent and coordinated by the nature of the
database structure. Departments are provided with different views of the information
but the underlying data is consistent. Data captured in one department may be the
source of information for the activities of another department. Because such systems are
standardized, changes to data and the information produced have to be carefully
controlled. There is a need for enterprise systems management.

Copyright © Genetic Computer School, 2002


Page 20

Interdepartmental management information systems differ from workgroup


systems in several ways. Workgroups will know each other whereas users of an
interdepartmental system are distant and the only common factor may be the use of an
information source, they are separated both physically and organisationally.

Workgroups may number in the tens of members, interdepartmental systems could


supply the information needs of hundreds.

Data contained in interdepartmental systems tends to be more complex and


heterogeneous in nature.

Whilst workgroup systems tend to be highly functional, indepartmental systems


cross functional boundaries coordinating the information needs of several process
related departments.

Business process design is emerging as a necessary step to optimize the use of


information technology. Rather than automating existing processes, the underlying
business processes are redesigned to take advantage of the new technologies.

Business process redesign can have dramatic impact for an enterprise; This is attributed
to the change of culture, the redesign of work, the retraining or the loss of staff.

Business network redesign refers to the use of information systems to enable


groups of enterprises to interact more productively. Four types of network redesign can
be identified; electronic data interchange (EDI), in which organisations agree on a
common set of data standards, interenterprise system access in which organisations use
each other's systems, interenterprise system integration in which enterprises develop
shared information systems and knowledge networks use information technology.

Business scope redefinition produces applications, which changes the nature of an


enterprise's activities. New technologies can make it possible to enter new markets,
enable new products or extend or enhance the capabilities of existing ones.

Copyright © Genetic Computer School, 2002


Page 21

CHAPTER 3: ORGANIZATIONS AND INFORMATION SYSTEM

Introduction

Organization

MIS Framework - Nolan & Wetherbe

Figure 3.1 Effectiveness of Organizations

It could be said that the role of the organization is to support creative individuals
while providing an environment where they can create knowledge.

But why is it important for organizations to provide this supportive environment?


Well, one only has to look at the changes that have taken place on a macro-economic
scale, in the creation of wealth; the reduced contribution by the manufacturing-based
industry sector and corresponding expansion of the service and information sectors. This
shift was identified in the early 1960's, and the term "knowledge worker" was born. Since
that time, knowledge has increasingly been viewed as a resource with knowledge workers
playing a pivital role in many organizations.

An organization is defined as a systematic arrangement of people to accomplish


some specific purpose. Organizations share three common characteristics.

Copyright © Genetic Computer School, 2002


Page 22

a. Has a distinct purpose.

b. Is composed of people.

c. Develops a systematic structure that defines and limits the behavior of


organizational members.

All effective organizations are not necessarily efficient. Effectiveness is concerned


with goal attainment, while efficiency is concerned with resource usage (refer to figure 3.1).
An organization can achieve its goals, but do so by being wasteful and using an
inordinate amount of input resources.

3.1 Workgroup Information Systems

A workgroup is a collection of people who work together to achieve a common goal.

In homogeneous workgroups, everyone in the workgroup fulfills the same role. As


such workroups grow so does productivity, in 'a" linear fashion. Providing information
systems for these workgroups can be very cost effective since everyone carries out
similar work.

In heterogeneous workgroups there are several roles and jobs covered by


individuals within the group. For this reason productivity does not increase in a linear
progression when additional people join the group. The increase in productivity is more
related to the previous knowledge of the individuals joining the group. Training and
communication becomes an important issue. Since there are many roles and jobs
covered by a heterogeneous group the provision of information systems can be
expensive and difficult to support.

Workgroup norms can play an information role in the acceptance or rejection of


certain information systems. The influence of workgroup attitudes has a large part to
play, peer pressure can persuade or deter acceptance, fears can be allayed or reinforced.

Workgroups can be permanent or temporary. They can exist on single sites or be


distributed throughout an organization.

Copyright © Genetic Computer School, 2002


Page 23

Workgroup effectiveness can be measured by output, personal satisfaction of the


members, and by the group capacity for further cooperation. Effectiveness is
determined by group effort, knowledge and skill. The strategies and approaches used to
perform work are important.

The major difference between personal and workgroup information systems is the
need for workgroup systems to control the shared use of resources. This must be
achieved without inconvenience to individual members of the group. Granularity refers
to the size of the shared information resource. Large granularity means that group
members share large information sources with the possibility of delays due to
contention problems but with small information systems adminstration overheads.
Conversely, small granularity produces few contentions but administration overheads
rise accordingly.

The major categories of a workgroup information system are hardware sharing and
data sharing facilities. Hardware sharing applications allow members of the workgroup
to share many hardware devices, some of these devices may be uneconomical for a
single user system.

Systems that focus on communications are applications such as group E-mail


systems that provide a number of mailboxes that can be for individual mail or group
and project mail. These systems can be passive or dynamic, the user or group has to
access incoming mail in the static system, whereas, in the dynamic system the user is
told that mail has arrived. Group E-mail systems are designed to enhance group
coordination. Group conferencing including electronic bulletin boards,
videoconferencing with white boarding, help to ensure coordination and often support
group dynamics in problem-solving settings.

Collaborative writing systems let users share the writing of documents. These
applications coordinate and control the users efforts. The tedious copying and altering
of documents in a single user environment is avoided.

Copyright © Genetic Computer School, 2002


Page 24

Analytical support for workgroups is typically provided by workgroup


spreadsheets. These are often composed of spreadsheet templates, the coordination of
these spreadsheets can be complex. To make these work effectively it is essential that
good documentation and training is provided.

Group Decision Support Systems are applications which allow individual users to
use group workboards in their own decision making process. Group decision making
can be facilitated by a controller with the group using support applications.

Another important category in workgroup information systems is the retrieval of


information, including group database management systems, workflow automation
systems, group scheduling systems, group project management systems and shared
textbase systems.

Group database management systems have similar functionality to personal


database systems but are more complex and tend to concentrate on workgroup
coordination.

Workflow automation applications provide automatic routing of forms used in


routine work processing. They also track the progress of forms. Groups can often
improve processes by looking at the output from workflow systems.

Trucking and monitoring is provided by similar systems provided for personal


information systems but they are more complex. Project management has more
complexity with the focus on group coordination.

Group textbase systems are similar to personal systems but shared documents are
indexed for group or individual retrieval.

3.2 Personal Information Systems

Personal computer systems are used to help an individual facilitate their work.
This facilitation can be considered as support for communication, analysis and tracking
and monitoring.

Copyright © Genetic Computer School, 2002


Page 25

Typical communication support is provided by word processors, which can be


used to produce simple text documents or more sophisticated documents containing a
wide range of colours, fonts and even graphics. The functionality of these programs can
often provide templates, boilerplate and automated document production. They also
feature mailmerge and support scripting and macro languages to extend the
customization of document production. Within the scope of communication support lies
the desk-top publishing application and the presentation application. Both can produce
professional quality output that can be shown on computer slideshows or be printed.

Analytical support is given by such applications as the electronic spreadsheet


capable of modelling quite complicated business activities. Many such programs are
supported by scripting and macro languages and can support graphics and database type
functionality. These applications are often enhanced to provide connectivity to various
internal and external data sources to enable data extraction. Other types of analytical
programs are the operational research ones providing linear programming and other methods.

Tracking and monitoring support is given by database and project management


applications. A database application includes a database, some means of data entry,
reporting facilities and possibly applications programs. Users can interface with many
databases using default, customized or programmed access. A common interface
standard is SQL, structured query language. Project management applications keep
track of a project's progress with regard to the allocation of resources and the
management of tasks. A variety of techniques are used, typically, Gantt charts, critical
path graphs and the like.

Integrated application provide an environment in which all applications can work


statically or dynamically with each other.

Changes made to database application will dynamically update speadsheet models.


In the Microsoft Windows environment this is achieved using DDE, dynamic data
exchange and OLE, object linking and embedding.

Copyright © Genetic Computer School, 2002


Page 26

3.3 The System Development Process

Systems Development Life Cycle

Define Problem

Develop/ Code

Figure 3.2 The System Developement Life Cycle

System Development Life Cycle

From concept to production, you can develop a database by using the system
development life cycle, which contains multiple stages of development. This top-
down, systematic approach to database development transforms business information
requirements into an operational database. (Refer to figure 3.2)

Strategy and Analysis

Study and analyze the business requirements. Interview users and managers to
identify the information requirements. Incorporate the enterprise and application
mission statements as well as any future system specifications.

Build models of the system. Transfer the business narrative into a graphical
representation of business information needs and rules. Confirm and refine the model
with the analysts and experts.

Copyright © Genetic Computer School, 2002


Page 27

Design

Design the database based on the model developed in the strategy and analysis
phase.

Build and Document

Build the prototype system. Write and execute the commands to create the tables
and supporting objects for the database.

Develop user documentation, help text, and operations manuals to support the
use and operation of the system.

Transition

Refine the prototype. Move an application into production with user acceptance
testing, conversion of existing data, and parallel operations. Make any modifications
required.

Production

Roll out the system to the users. Operate the production system. Monitor its
performance, and enhance and refine the system.

The SDLC procedures ensure that timely and accurate information concerning the
progress of system development is available to stakeholders and others in the university
community. System Development Life Cycle refers to a methodology for developing
systems. It provides a consistent framework of tasks and deliverables needed to develop
systems.

The SDLC methodology may be modified to include only those activities


appropriate for a particular project, whether the system is automated or manual,
whether it is a new system or an enhancement to existing systems.

The SDLC methodology tracks a project from an idea developed by the user,
through a feasibility study, system analysis and design, programming, pilot testing,

Copyright © Genetic Computer School, 2002


Page 28

implementation and post-implementation analysis. Documentation written during


project development is used in the future when the system is reassessed for its
continuation, modification, or deletion.

Alternative Quality Process - an alternative quality process is a systematic


sequence of tasks for planning, organizing, and monitoring for quality performance
improvement. It begins with customer requirements, focuses on root causes or barriers
to improvements, and ensures that decisions and actions are based on real data; it
includes problem solving processes such as the plan-do-check-act model. Detailed
descriptions are available in the materials located in the quality center or available
through team training.

System - an organized collection of independent tasks and processes that is


designed to work together in order to accomplish specific objectives. The processes and
tasks typically receive input(s) from and provide output(s) to other processes and tasks
and even other systems. The tasks and processes may or may not be supported by
automation.

The system development life cycle (SDLC) is an organized approach to obtaining


an information system. The SDLC is formalized in many organizations, with detailed
instructions outlining reporting requirements, specific tasks that must occur in each
phase, and individual responsibilities. Major objectives of system development are to
create a system: with the desired capabilities, within budget, and on time.

The SDLC assessment process involves three components:

1. Understanding the current level of achievement (baseline assessment),

2. Defining the goal or achievement objective, and

3. Developing a plan to achieve the defined goal.

The System Development Life Cycle (SDLC) is intended to provide a set of


guidelines for the successful completion of application system development projects.
The SDLC consists of seven distinct Phases as shown in Figure 3.3 below:

Copyright © Genetic Computer School, 2002


Page 29

Phase 1 Phase 2 Phase 3 Phase 4 Phase 5 Phase 6 Phase 7

Planning Definition Analysis Design Build Transition Warehouse

Each Phase contains one or more individual deliverables associated with the Phase.

Figure 3.3

Planning

This Phase of the SDLC is required to determine the feasibility of a particular


project proceeding, or not. This Phase will produce a high-level overview document of
the proposed project. It will contain information relating to the project's requirements
and will enable the formalization and definition of the scope of the project.

This is the first Phase in the SDLC. The Project Charter deliverable is the first
major deliverable in the project. The Project Charter is to be produced prior to the
commencement of a project.

Due to fiscal year budget allocations and fiscal contract restrictions the Project
Charter will typically refer to project activity up to the end of the fiscal year in which
the project starts. There may however be exceptions and special circumstances when a
Project Charter will span more than one fiscal year. In these cases, this should be
specifically stated in the Project Charter document.

For projects spanning more than one fiscal year, an updated Project Charter will
be required for each year that the project is undertaken, prior to the start of the fiscal
year (normally done in March of each year).

The updated Project Charter for a subsequent fiscal year(s) will mainly focus on
identifying the scope and deliverables of the project for that particular fiscal year. The
updated Project Charter will also include revisions (as appropriate) to project team,
budget, schedule and project status. The Business Champion is responsible for
producing the Project Charter. However, the Business Champion may delegate this

Copyright © Genetic Computer School, 2002


Page 30

activity for completion. Together with appropriate members of the User Team, the
Business Champion will provide information for the completion of the Project Charter.

The Project Sponsor is responsible for reviewing and approving the Project
Charter and securing the necessary funding for the project

The draft Project Charter must be submitted to the Information Systems Branch
(ISB) for Business Analyst review and QA against standards.

The completed Project Charter will be submitted to the Executive Management


(Steering _ Committee) for approval and budget allocation authorization.

Definition

This Phase of the SDLC defines exactly what, who, when and how the project will
be carried out. This Phase will take the deliverable from the previous Phase (Project
Charter), expand on the high-level project outline and provide a specific and detailed
project definition.

This Phase is the first activity of the project after obtaining approval and funding
to proceed.

The description of this Phase here assumes that the following associated activities
have been completed:

• the preparation and distribution of an RFP;

• the selection of a contract development team; and

• the appointment of a Project Manager.

This Phase provides an effective way of communicating to project stakeholders,


the project scope and schedule as well as any risks or constraints related to the project.

This is the second Phase in the SDLC but the first Phase of the project itself. The
Project Statement deliverable from this Phase must be completed and signed-off prior
to commencing the next phase of the project.

Copyright © Genetic Computer School, 2002


Page 31

Agreement to the Project Statement ensures that everyone involved in the project
is clear on the project scope, objectives, goals and outcomes.

The Project Manager is responsible for producing the Project Statement. However,
the Project Manager may delegate this activity to the Development team for
completion.

The Business Champion, Business Analyst and appropriate members of the User
Team will provide project information which will support the completion of the Project
Statement.

The draft Project Statement must be submitted to the Information Systems Branch
for Business Analyst review and QA against standards.

The completed Project Statement will be submitted to the Business Champion and
Project Sponsor for approval and sign-off.

Problems can be viewed as unsatisfied needs. They can be seen as opportunities to


develop new and better methods, procedures and if appropriate computer systems. The
definition phase should be used to examine all aspects of the problem and to come up
with a clear definition of the problem. The scope of problem definition varies. In an
informal setting simply specifying business application requirements may be sufficient.
In a more formal case the product of the definition phase is a FUNCTIONAL
SPECIFICATION - a precise description of the functional requirements of a computer
system.

The definition phase can be viewed as a series of steps;

• Getting to know the application.

• Deciding whether a computer system can meet the application's requirements


and how it will do so.

• Defining processing requirements constraints, costs and benefits.

• Developing functional specifications and prototypes.

Copyright © Genetic Computer School, 2002


Page 32

• Reviewing the specifications and prototypes.

In some instances the steps leading to the functional specification and evaluation
might be quite informal. The systems analyst uses the simple techniques of questioning
and comparison. By asking questions, the analyst can consult references that have
already addressed the problem or can talk to people with more experience in the
problem area under consideration. By making comparisons the analyst will be able to
recognize and isolate those parts of the problem that are familiar.

In other instances, as in a large organization where many projects are in


competition for management's attention more formal analytical techniques, such as
feasibility studies, top-down design and structured analysis are required.

The more people that need to interact in the decision-making process the more
formal the analysis and specification of the functional requirements will have to be.

One important aspect of functional specifications is their value as a communication


tool between .w

Analysis

This Phase of the SDLC is required to understand and document the users' needs
for the system. This Phase will document, in significantly more detail than the Project
Statement, the scope, business objectives and requirements of the current/proposed
system.

The emphasis throughout this Phase is on what the system is to do. During the
analysis and specification, the technical aspects and constraints should be considered,
but should not be influenced by implementation characteristics. The technical aspects of
the system are addressed in the Design Phase.

During this Phase the Data Conversion requirements, at a high level, will become
known. This will commence a parallel set of SDLC Phases for the Data Conversion
associated with the system. Data Conversion will follow Phases 3 to 6 of the SDLC.
Depending on the size and complexity of the total project (system and data conversion)

Copyright © Genetic Computer School, 2002


Page 33

the Data Conversion Application could be either incorporated as a section or


component of the main system Design deliverable, or a separate Data Conversion
deliverable of its own.

In addition, the Data Warehousing requirements will also be identified during this
Phase and a parallel set of Data Warehouse SDLC Phases should commence.
Depending on the business requirements of the system, it is possible that the Data
Warehousing aspects will only be considered once the system has been operational for a
while.

These parallel streams of SDLC are shown in more detail in Figure 3.4 below. I

Figure 3.4: SDLC Parallel Processing Streams

System Development Data Conversion Data Warehouse

Life-Cycle Phases SDLC Phases SDLC Phases

The Detailed System Analysis is to be stated in a language that reflects the


background of the user base. Most requirements are written as natural language
sentences supplemented with diagrams, graphics and tables of detailed information.

The Project Manager is responsible for producing the deliverables associated with
the Detailed System Analysis. However, the Project Manager usually delegates this to

Copyright © Genetic Computer School, 2002


Page 34

the Business Analyst. This deliverable has significant input from the Business
Champion and User Team. In cases where the production of some or all of the Phase's
deliverables have been delegated, the Project Manager will still maintain overall
responsibility for the production of quality deliverable(s) submitted to the Business
Champion for review and sign-off and ISB for Quality Assurance.

The Project Manager will provide initial Quality Assurance of the deliverable
prior to review by ISB QA, User Team and the Business Champion.

The draft Detailed System Analysis must be submitted to the Information Systems
Branch for Data Administration review and QA against standards.

The completed Detailed System Analysis will be submitted to the Business


Champion and Project Sponsor for approval and sign-off.

Design

This Phase of the SDLC continues on from the Detailed System Analysis and
describes how the proposed system is to be built. The Design is specific to the technical
environment that the system will be required to operate in and the tools to be used in
building the system. The results of this Phase will significantly impact the Build and
Transition Phases of the system.

The Project Manager is responsible for producing the deliverables associated with
the Detailed System Design.

However, the Project Manager usually delegates responsibility for some or all of
these deliverables to the Development Team. In cases where the production of some or
all of the Phase's deliverables have been delegated, the Project Manager will still
maintain overall responsibility for the production of quality deliverable(s) submitted to
the Business Champion and ISB for Quality Assurance.

The Project Manager will provide initial Quality Assurance of the deliverable(s)
prior to review by ISB QA, User Team and the Business Champion.

Copyright © Genetic Computer School, 2002


Page 35

The draft Detailed System Design must be submitted to the Information Systems
Branch for Database Administration review and QA against standards.

The completed Detailed System Design will be submitted to the Business


Champion and Project Sponsor for approval and sign-off.

Build

This Phase of the SDLC deals with the development, unit testing and integration
testing of the system (application) modules, screens and reports. In addition, this Phase
will address the preparation and establishment of the technical environment for
development, testing and training of user representatives.

This Phase is usually carried out in parallel with the development of user
procedures and user documentation from the Transition Phase. Both of these will be
required for module testing, upon the completion of the Build Phase. Coordination of
the activities of the Build and Transition Phases is a key responsibility of the Project
Manager at this time.

Any special procedures for data conversion and/or data warehousing are also
developed and tested. The processes of developing and testing of data conversion and
data warehousing modules is no different from those required for the system itself.

This is the fifth Phase in the SDLC.

The Project Manager is responsible for producing the deliverables associated with
the Build Phase. However, the Project Manager usually delegates responsibility for
some or all of these deliverables to the Development Team. In cases where the
production of some or all of the Phase's deliverables have been delegated, the Project
Manager will' still maintain overall responsibility for the production of quality
deliverable(s) submitted to the Business Champion, User Team and ISB for Quality
Assurance.

The Project Manager will provide initial Quality Assurance of the deliverable(s)
prior to review by ISB QA, User Team and the Business Champion.

Copyright © Genetic Computer School, 2002


Page 36

The draft System Build must be submitted to the Information Systems Branch for
technical review and QA against standards.

The completed System Build will be submitted to the Business Champion and
Project Sponsor for approval and sign-off.

Transition

This Phase of the SDLC is to prepare for and carry out the transition of the
developed system through user and acceptance testing to a full production system.

This is the sixth (and in some cases last) Phase in the SDLC. This will be the last
Phase only for those Business Systems that (for specific documented reasons) will not
make its data available in the Data Warehouse.

This Phase will provide users with the documentation and training to effectively
use the system. Although the Data Conversion will only to be done once, user
documentation will also be required.

Individual system components that successfully completed unit and integration


testing during the Build Phase are now subjected to a more rigorous system and
acceptance testing as defined by the testing plans. In addition user and operation
procedures are to be tested for the system.

As appropriate, the Data Conversion will be performed prior to the finalization of


the system into Production. The detailed Data Conversion and Transition Plans will
define in more detail exactly how this will be accomplished.

The Project Manager is responsible for producing the deliverables associated with
the Transition Phase. However, the Project Manager usually delegates responsibility for
some or all of these deliverables to the Development Team. In cases where the
production of some or all of the Phase's deliverables have been delegated, the Project
Manager will still maintain overall responsibility for the production of quality
deliverable(s) submitted to the Business Champion, User Team and ISB for Quality
Assurance.

Copyright © Genetic Computer School, 2002


Page 37

The Project Manager will provide initial Quality Assurance of the deliverable(s)
prior to review by ISB QA, User Team and the Business Champion.

The draft deliverables must be submitted to the Information Systems Branch for
technical review and QA against standards

Warehouse

This Phase of the SDLC addresses the publication of the system's data into the
Ministry's Data Warehouse for business manipulation and decision support. Although
described as one Phase here, the Warehouse Phase actually comprises, as appropriate,
all the deliverables associated with SDLC Phases 2 [Definition] through 6 [Transition].

This is the final Phase in the SDLC.

The Project Manager is responsible for producing the deliverables associated


with the Warehouse Phase. However, the Project Manager usually delegates
responsibility for some or all of these deliverables to the Development Team. In cases
where the production of some or all of the Phase's deliverables have been delegated,
the Project Manager will still maintain overall responsibility for the production of
quality deliverable(s) submitted to the Business Champion, User Team and ISB for
Quality Assurance.

The Project Manager will provide initial Quality Assurance of the deliverable(s)
prior to review by ISB QA, User Team and the Business Champion.

The draft deliverables must be submitted to the Information Systems Branch for
review and QA against standards.

Copyright © Genetic Computer School, 2002


Page 38

CHAPTER 4: EVOLVING APPLICATIONS OF


INFORMATION TECHNOLOGY

The application of IT has evolved and is continuing to evolve through 3


stages: Automation of work

Information management Business transformation

This evolution involves major leaps in the complexity of tasks that it is being
designed to perform. As we review this evolution, a consistent pattern of change
emerges in the business application of IT. As we evolve automation of work
through information management to business transformation, the strategic
importance of IT applications increases and that amount of organizational change
required to realize the benefits of an application is also greater. Specifically, an
increasing number of changes are being made to elements of the business system
beyond IT such as business processes, organisational structure and even business
culture. At the same time, the number and complexity of applications (or potential
applications) also increases. The three stages of evolution are summarised in the
table 4.1 below.

4.1 Stages of IT Evolution

Stage Impact Benefit Examples

Automation of work • Getting work Operational • Payroll


Done efficiency • Check processing
• Doing the • Basic order
same things more processing
efficiently • Basic airline
reservation
systems

Copyright © Genetic Computer School, 2002


Page 39

Information • Restructuring Operational and • Customer


management work and work Tactical effectiveness information systems
processes • Airline yield
• Doing things management systems
differently • Executive
information systems
Business • Defining the Strategic effectiveness • JIT inventory
transformation business and positioning systems
• Doing different • Electronic
things commerce
• Changing the • OLAP
business/industry
rules

Table 4.1

• Automation of work

The first application of IT in business involved the automation of work tasks


such as census data calculations, check processing and payroll as well as basic
order processing and reservation systems. An automation application such as
payroll, is not a strategic application, and while it is certainly necessary, and
failure to pay employees would have serious consequences, it is not an application
that provides strategic advantage. The benefits were largely in the area of
operational efficiency. A few new jobs were created to program, operate and
support the technology itself and some manual jobs, such as pay calculation and
check processing, were replaced. There was also limited change to people's jobs or
to business processes, but the overall change to the nature of work was not
significant. Learning requirements were relatively simple and narrow, focused on
how to use the technology. Change was generally limited to one or a small number
of functional areas. In the case of payroll, little if any change was experienced

Copyright © Genetic Computer School, 2002


Page 40

outside of the payroll department. The most important thing was that the payroll
application ran correctly.

The self-sufficiency of computerization was reflected by the physical reality


of the mainframe, where computers were isolated behind data center walls - and
operated invisibly by IT experts. Applications were often limited to those
conceived by same experts, with little understanding of technology by broader
business community, or how it could be applied.

• Information Management

Automation applications created information as a by-product of automating


work. In the early years of the automation stage, this information was not generally
used, certainly not in widespread format way. As we moved into the information
stage, the opportunities to use this information began to be recognized. With the
wider distribution of desktop computer terminals, IT was increasingly applied to
provide information to support improved decision making, to move it "close to the
customer" and to support new service and product design. Here the introduction of
advanced order processing systems, airline yield management systems, customer
information systems as well as the start of so-called executive information systems
(EIS). Benefits moved beyond operational efficiency to operational and tactical
effectiveness. Information could be used to make tactical, and in some cases,
strategic decision. Initially, information was used to enable workers to do their jobs better.
Their jobs changed somewhat, but primarily they were required, and had to be
trained to take largely predetermined action based on the information provided.

As the information stage advanced, benefits were premised on workers


improving how they analyzed and applied information to their work. In the case of
order processing, seasonal variations in demand might be noticed, and adjustments
made to order levels. In the case of customer information systems (or customer
information files in financial institutions), information was used to increase the
value of "customer moments" by cross-selling and target marketing of certain
services. Airline systems moved beyond basic reservation systems to sophisticated

Copyright © Genetic Computer School, 2002


Page 41

yield management systems. In the later steps of the information stage, automation
information bases provided opportunities to design new products, such as today's
multitude of mutual funds and numerous volume-based discount plans for valued
customers. It was no longer sufficient to simply provide the application and make
sure that it worked as specified. For these benefits to be realized, the nature of
people's work had to change. Business processes had to be restuctured and better
integrated. Reward systems had to change. Significant learning was required. The
changes crossed functional boundaries, and in some cases, changed or eliminated them.

Physically, personal computers emerged from behind the walls of - the central
data center. PCs began to appear everywhere in organizations and to be operated
by nonexperts. The number of potential applications of technology increased
dramatically. Many of these were conceived outside of the IT world, by the
broader community of business managers and front-line technology users.

• Business transformation

Information management applications enable organisations to rethink and


redesign their business processes and how they carry out their business. As more
and more computing power is distributed, and as advanced communications
capabilities continue to erase the constraints of time and distance, the very nature
of businesses, and even entire industries, is being redefined. Benefits have moved
beyond operational and tactical effectiveness to strategic effectiveness and positioning.

Business transformation applications, such as just-in-time (JIT) inventory


systems and advanced electronic commerce, enable organizations to rethink not
just how they do things, but also what they choose to do. For example, JIT
ordering and inventory management systems are fundamentally changing the
value/supply chain and shifting the balance of power among stakeholders.

The emergence of Internet and virtual banking is redefining the financial


industry by removing century-old barriers to entry and blurring financial product
boundaries. Airlines are now offering passengers direct access to reservations

Copyright © Genetic Computer School, 2002


Page 42

systems and more fighting for ownership of the client with travel agents, and thus
redefining the travel agent business. Amazon.com is helping to redefine the book
industry. It is not only selling books electronically and offering a wider selection
than is possible in physical bookstores, it is using the power of computers to
repackage - and eventually transform - a range of services that were historically
spread across multiple businesses, including the reference capabilities of libraries,
the retail display and selection expertise of bookstores, the efficiency of volume
discount distributors and the knowledge of professional book reviewers.

All these applications have significant strategic implications. While


technology enables these benefits, most of the required change is beyond the realm
of IT. Changes are required in the mission, the very raison d'etre of organizations.
The organizations carrying out these changes will, in many cases, be redrawing
traditional industry boundaries or, at minimum, changing industry structures and
rewriting industry ground rules. In doing so, they will harness IT with the aim of
"competing for the future".

The potential of transformational applications is tremendous, but to realize it


will present organizations with new and significant challenges. Automating payroll
processing was primarily an engineering question, whereas creating a virtual bank
branch or bookstore is primarily a business one.

As we moved beyond automation of work to information management and


business transformation applications, sound management of IT projects remains
necessary but is no longer by any means sufficient. In the case of financial
customer information files, for example, employees have to learn new skills,
assume new responsibilities and accept different reward systems. Cross-selling to
banking customers means astutely interpreting customer profile information and
cultivating personal relationships, rather than efficiently handling transactions and
answering routine questions to shorten waiting time in a branch line-up. Such
changes in management practice and work place culture are just as important as

Copyright © Genetic Computer School, 2002


Page 43

new IT - though perhaps less visible - to any organization's ability to make a


successful transition to Knowledge Economy.

Organizations that have recognized this, and changed their management


practices accordingly, have been relatively successful with their IT investments. In,
those organizations that have not - the majority, unfortunately - the failure to
recognise these changes has caused management practices to lag behind the
evolving nature of the application of IT.

While the application of Information Technology has evolved significantly


over the past three decades, our approach to managing it has not. When the primary
application was automation, management thinking was still rooted in the industrial
age. The mind-set meant that you were ahead and built the system, plugged it in
and made sure it was running, like a new machine on a manufacturing assembly line.

Unfortunately, as the application of IT has moved beyond automation of work


all the way to business transformation, our management approach has remained
rooted in industrial-age thinking. Management thinking has failed to understand
the implications of the evolving role of IT in business system beyond technology.
Many still think in terms of payroll processing systems that can begin depositing
money in employee accounts on day one. In fact, we are delivering customer
information systems that will only produce results gradually on day 50, day 100 or
day 365, after people are trained and motivated to use the new application when
serving customers.

4.2 Management blind spots: Four critical dimensions of complexity

Current management practice fails to adequately address the impact and


resource implications of four critical dimensions of complexity. These blind spots
in traditional management mind-sets are: linkage, reach, people and time.

Copyright © Genetic Computer School, 2002


Page 44

Linkage

This refers to the linkage between the expected results of an IT investment


and business strategy, and between the IT investment and investments required in
other areas of the business in order to realise the benefit. Understanding and
addressing linkage requires a clear appreciation of the ultimate benefits and of the full
scope of the investment required to achieve the benefits.

Reach

Reach refers to the breadth of change required by an IT investment, meaning how


much of an organization is impacted. It also refers to the depth of change - the degree of
impact and of organizational change required to realize the benefit. Addressing each
requires understanding what areas of the organization, other organizations and
stakeholders will be affected, what the impact will be and how it will be managed.

People

A large number and diversity of people must be motivated and prepared to change.
This critical factor in business transformation is often underestimated. We need to
understand who these people are today, how they will have to change and what
interventions will be required to effect the change. We need to ask how these interventions
will be managed for people with different starting points, attitudes and motivations.

Time

In business transformation, time is always of essence, but realistic time frames are
notoriously hard to estimate. We need to ask - and ask again and again - what the realistic
length of time is for all the necessary changes to occur and for the full benefits to be
realized. We must base these estimates on understanding the previous three dimensions.

Copyright © Genetic Computer School, 2002


Page 45

Evolving complexity of IT applications

These dimensions of change have become increasingly complex as the applications


of IT have advanced through the stages of automation of work, information management
and business transformation.

In the automation stage, the four dimensions were fairly straightforward and posed
few problems. In the case of automated payroll systems, for example, there few linkages,
organizational reach was limited and few people were affected. Time was required to
deliver the benefits was short; or, at least, the time frames were easily predictable in
advance. Finally, benefits were easy to measure. As we moved through the information
stage, there were more linkages, not all of which were, obvious

4.3 Business transformation and the knowledge economy

The potential risks and rewards associated with such cases of business
transformation show, what is involved in engineering out transition to a knowledge
economy. The opportunities include expanding geographic scope, expanding electronic
commerce and creating virtual companies. We are, moving toward an economy that is
on-line, interactive, instantaneous, inter-networked and knowledge based. It is an
economy that will require new organizational forms and which will dramatically change
the nature of organizations and work.

The emergence of this new economy involves business transformation -


fundamental change in value chain management and the application of new technologies
to support "networked organizations that share knowledge, insight and experience
effectively. Some experts predict chief executives will become knowledge capitalists who
manage the knowledge assets of their organizations. Knowledge will not just be limited
to your organization; it will come from outside well. In addition to managing investments
in IT-enabled change in your own business system, you will have to manage change in an
extended business system which includes customers, supplier financial institutions,
regulators and many other intermediaries, all of whom will themselves be in state of change.

Copyright © Genetic Computer School, 2002


Page 46

While the opportunities created by business transformation are awesome the risks
can be daunting to investment decision-makers. Today's large-scale IT projects and
organizational change programs will be viewed as relatively simple initiatives
compared to the sophisticated business transformation ones that will be required in the
knowledge economy. These will raise significant new issues of linkage, reach, people
and time. To manage these dimensions of complexity successfully, business
transformation initiatives can no longer be viewed as traditional projects. They will
need to be treated almost like mid-size businesses within the business, as programs that
are managed continuously and proactively over long periods of time.

Copyright © Genetic Computer School, 2002


Page 47

CHAPTER 5: MANAGEMENT AND DECISION-MAKING

5.1 What is Management?

Introduction

Every organization, regardless of size, type, or location needs managers. The


characteristics of managers vary. Managers may come from any nationality or be
of either gender. They're just as likely to be women, particularly in middle
management and supervisory management positions, as they are men. For instance,
the manager described in the chapter opening Manager's Dilemma, Cynthia Glenn,
chief operating officer of Oxford Health Care, is a good example of a successful
manager in action. She's involved with a major change that will radically transform
her organization. As the COO, she epitomizes the planning, organizing, leading, and
controlling activities that must be performed in order for the company to meet its goals.

Management is the process of getting activities done efficiently with and


through other people. The process includes the functions or primary activities
performed by managers. Efficiency refers to the relationship between inputs and
outputs and refers to efforts to minimize resource costs. Effectiveness refers to
goal attainment. Managers seek to be both efficient and effective.

An organization is a systematic arrangement of people to accomplish some


specific purpose. Managers are important to an organization's success because they
direct and coordinate activities so the organization can reach its goals.

Management is important in our society today. Accordingly, there are two


reasons for studying management. We interact with organizations every day of our
lives. Every product we use and every action we take is provided or affected by
organizations. These organizations require managers.

Managers earn more than operators because their decisions have a significant
effect on organization's performance and because of the inadequate supply of
effective managers.

Copyright © Genetic Computer School, 2002


Page 48

Managers come in all shapes and sizes. First of all they exist at various levels
within the (hierarchic) structure of an organization. Secondly, they are found in a
variety of functional areas, such as marketing, finance and manufacturing.

Nevertheless, while there are obvious differences between the different levels
and functional areas, all managers perform the same functions and play the same roles.

A successful manager must possess many skills, of which communication and


problem solving are the most basic.

Managers transmit and receive information in both written and in oral form.
Written communications include reports, letters, memos, books and magazines,
electronic mail and electronic bulletin boards like the World Wide Web. Oral
communication comes from telephone calls, formal and informal meetings, and
voice mail.

Problem solving is the set of activities that leads from the recognition of a
problem to its solution, where a problem is a condition or event that is (potentially)
harmful or alternatively, a condition or event that is (potentially) beneficial. While
trying to find a solution to a problem, managers engage in a sub-process of
decision making, where a decision is a particular selected course of action. In
reaching a solution to a given problem, it is generally necessary to make decisions.

Management and Leadership

Management and leadership, they are often used to mean the same thing. But what
many people don't realize is that they are quite different. What is management and what
is leadership? How do they relate?

Management is composed of five major functions; planning, organizing, staffing,


directing and controlling. These five functions are used to define what is management
and the role of a manager. These functions are easily defined and match the definition
of management which is; an activity undertaken by those who are responsible for the

Copyright © Genetic Computer School, 2002


Page 49

success of organizations. Other key words that are used to define management are
coordinating, resources, accomplishment, and desired results.

Leadership is not very easy to define. One thing that is clear is that leaderships
function is to influence. The process of leadership is to influence. Influence in
leadership is defined as the ability to modify or change behavior of people. Influence
has five bases; legitimacy, coercion, rewards, expertise and referent. All these bases
help to exert influence. Management and leadership are the arts of influence over
people. They both are influencing people. Leadership is dealing solely with influence
over others. Leadership is dealing exclusively with making people do something they
otherwise would not do unless influenced to do. Management is much the same but is
not so aggressive as leadership is. Management is a way of making sure people stay on
task when they may not be on task.

The role of management is to guide an organization toward goals and management


is the process of reaction to goals by working with or through people and other
resources. Leadership is guiding people in a certain direction. They both sound alike,
but there is a difference that is important. Management is a function where as leadership
is a process which is not structure as well as management.

In management, the idea is to accomplish a goal that is set forth and agreed upon.
There isn't any real persuasion that goes into this because the goals are set and decided
on and the only thing need is to just do the work. In leadership, influence must be
exercised. Influence is a process of making people do something they otherwise would
not do. In management, influence is not very great because the goals are established and
agreed on. Influence is not a major part of management like it is in leadership.

Management and leadership are not the same. Management is a way to accomplish
goals that have been agreed upon. Leadership is influencing people. The two have many
areas that seem the same but they are very different because of the level of influence involved.

Copyright © Genetic Computer School, 2002


Page 50

Management Functions

Henri Fayol, a French industrialist from the early part of the 1900s, proposed that
managers perform five management functions: POCCC (plan, organize, command,
coordinate, control). These functions still provide the basis around which popular
management textbooks are organized. However, the functions have been condensed to four.

1. Planning includes defining goals, establishing strategy, and developing plans to


coordinate activities. Formulation of long and short term plans for goals.

2. Organizing is determining what tasks are to be done, who is to do them, how


the tasks are to be grouped, who reports to whom, and where decisions are to be made.
Creation of structures and frameworks of standards, procedures and policies.

3. Leading includes motivating subordinates, directing others, selecting the


most effective communication channels, and resolving conflicts.

4. Controlling is monitoring activities to ensure that they are being


accomplished as planned and correcting any significant deviations. Monitoring
performance and the environment.

Although the functional approach is clear and simple, critics have suggested
that it does not provide an accurate description of what managers actually do.

Management Roles

In the late 1960s, Henry Mintzberg conducted a precise study of managers at


work. His findings challenged several long-held beliefs about the manager's job.
He concluded that managers perform ten different, but highly interrelated roles.
Management roles refers to specific categories of managerial behavior.

1. Interpersonal roles included figurehead, leadership, and liaison activities.


In Mintzberg's framework, there are three interpersonal roles that manager play.

Figurehead

Copyright © Genetic Computer School, 2002


Page 51

The manager performs ceremonial and other duties, such as greeting visiting
dignitaries and signing legal documents.

Leader

The manager maintains the unit by hiring and training the staff and providing
motivation and encouragement

Liaison

The manager makes contact with persons outside the manager's own unit -
peers and others in the unit's environment - for the purpose of attending to business
matters. In many organizations it is the practice for all mail to be addressed to or
signed by the manager.

2. Informational roles included monitoring, disseminating, and spokesperson


activities. In Mintzberg's framework, there are three informational roles that
manager play.

Monitor

The manager constantly looks for information bearing on the performance of


the unit, scanning both the internal activity of the unit as well as its environment.

Disseminator

The manager passes on valuable information along to others in the unit.

Spokeperson

The manager passes on valuable information along to those outside the unit-
superiors and persons in the environment. Common activities are issuing press
releases, giving media interviews, reporting to board meetings and so on.

3. Decisional roles included those of entrepreneur, disturbance handler,


resource allocator, and negotiator. In Mintzberg's framework, there are four
decisional roles that managers play.

Copyright © Genetic Computer School, 2002


Page 52

Entrepreneur

The manager makes rather permanent improvements to the unit, such as


changing the organizational structure.

Disturbance handler

The manager reacts to unanticipated events, such as a currency devaluation in


an overseas market.

Resource allocator

The manager controls the purse strings of the unit, determining which
subsidiary units get which resources.

Negotiator

The manager resolves disputes both within the unit and between the unit and
its environment.

Follow-up studies of Mintzberg's role categories in different types of


organizations and at different managerial levels within organizations have
generally supported the notion that managers perform similar roles. However, the
more traditional functions have not been invalidated. In fact, the functional
approach still represents the most useful way of classifying the manager's job.

Management Skills

Managers need certain skills to perform the varied duties and activities
associated with being a manager. Robert L. Katz found through his research in the
early 1970s that managers need three essential skills or competencies.

1. Technical skills are skills that include knowledge of and proficiency in a


certain specialized field. Technical skills became less important as a manager
moves into upper levels of management. However, top managers still need some
proficiency in the organization's specialty.

Copyright © Genetic Computer School, 2002


Page 53

2. Human skills include the ability to work well with other people both
individually and in a group. The importance of human skills remains consistent,
regardless of level within the organization.

3. Conceptual skills include the ability to think and to conceptualize about


abstract situations, to see the organization as a whole and the relationships among
the various subunits, and to visualize how the organization fits into its broader
environment.

All managers perform essentially the same functions, but lower-level managers
emphasize leading while upper-level managers spend more of their time planning,
organizing, and controlling. For the most part, the manager's job is the same in both profit
and not-for-profit organizations. Managers in small businesses tend to emphasize the
spokesperson role and are generalists. Also, the formal structure and nature of a
manager's job in a large organization is replaced by more informality in a small firm.
When managers work in different countries, they often need to modify their practices.

5.2 Managers and their Information needs

Figure 5.1 Management Structure

Management Structure

The model above (see figure 5.1) is rather simplistic but clearly, in any
substantial organization, there will be a top layer of senior management dealing
with high level, strategic and long-term issues for the organization. Below them
will be a middle management layer with responsibilities for performance

Copyright © Genetic Computer School, 2002


Page 54

monitoring and shorter time horizons. Finally, there will be an administrative level,
ensuring that the day-to-day operational activities are effectively managed.

Managers at all levels of the organization need information in order to ensure


that they are achieving the desired objectives or targets for their area of the
business. Those particular objectives will be different at the various levels and
business units in the organization (although they should all be geared to achieving
its corporate objectives) and, hence, their information requirements will also be
different. Nevertheless, there will be a flow of information both up and down the
organization structure as each level receives information for its purposes and
adapts it for the next level. We will be looking in more detail at these information
systems in a later session but for the moment the focus is on what information is
required.

It is common to classify managers in a hierarchic organization as operating at


one of three basic levels (refer to figure 5.2)

Figure 5.2 Hierarchic Organization of Managers

Copyright © Genetic Computer School, 2002


Page 55

1. The strategic planning level (Top management)

Upper management is usually comprised of the CEO and Vice Presidents of a


company. The decisions that they make have the broadest scope and cover the longest
time frame. Upper management decisions are very unstructured.

Upper Management uses a ‘Decision Support System’ to help make its decisions.

2. The management control level (Middle management)

Decisions by middle management, usually the Directors, cover a broader range of time
than operational, management but less broad than upper management. These decisions
involve experience using historical data to plan and control operations and implement
the policies of upper management. Middle management decisions are semi-structured.

Middle management uses a 'Management Information System', MIS, to produce


periodic reports.

3. The operational control level (Operating management)

Departmental or operating management is responsible for maintaining business


records and improving the flow of work on a daily basis. Decisions here cover a
narrow time-frame and are structured, i.e. predictable, and made by following a well-
defined set of procedures using the current status of activities.

Operating Management produces Daily Reports using a 'Transaction Processing


System'.

The characteristics of these three levels of management are quite different as


shown in the following table by Kanter, with top management being more involved
with planning and concentrating on a longer time horizon. Top management focuses
more on external information sources, and deals with complex, relatively unstructured
activities involving few people, but having broader scope than lower levels of
management.

Copyright © Genetic Computer School, 2002


Page 56

Characteristic Top management Middle management Operating management

Focus on planning Heavy Moderate Minimum

Focus on control Moderate Heavy Heavy

Time frame One to five years Up to a year Day to day

Scope of activity Extremely broad Entire functional area Single subfunction or task

Nature of activity Relatively Moderately structured Highly structured


unstructured

Very complex, Less complex, better


Level of complexity Straightforward
many variables defined variables
Job measurement Difficult Less difficult Relatively easy
Implementation
Plans, policies and
Result of activity schedules, performance End product
strategies
yardsticks
Type of information Internal, reasonable Internal historical, high
External
utilised accuracy level of accuracy
Responsible,
Creative,
Mental attributes persuasive, Efficient, effective
innovative
administrative
Number of people
Few Moderate number Many
involved
Departmental/
Intra-division Intra-department Inter-department
divisional interaction

Table 5.1

Table 5.1 shows Job content of management levels.

Copyright © Genetic Computer School, 2002


Page 57

Measuring Performance

As stated previously, effective information managers are those who have a clear
understanding of what information they require to tell them how they are performing
against their objectives. That is not always as easy as it sounds. Managing a busy
department, for example, is a 'messy' business, with problems coming from all
directions and the manger constantly having to reassess priorities to deal with them. It
is easy to lose sight of the real goals and then, as a consequence, having to take drastic
short-term actions to regain course.

Critical success factors (CSF's) are crucial areas of responsibility where 'things
must go right' requiring 'constant and careful attention from management'. For example,
responding to new business enquiries within two working days.

1. Identify the managers 'Critical Success Factors' (CSF's) in your organization.

2. Determine what information is required to show whether or not that CSF is


being achieved.

3. Set up the systems necessary to deliver that information as and when required.

This technique proved to be particularly effective in those organizations which


had clearly expressed corporate and departmental objectives as these ought to tie in
closely with CSF's. Additional benefits were also generated by applying the technique
throughout the organization whereby it helped to focus on improving communications
between levels of management.

This process has been applied and can be seen in many large multi-national
companies today, where subsidiaries operate with a high degree of autonomy, reporting
regularly to the parent company or Board on their performance in a number of 'key' areas.

‘Hard’ and ‘Soft’ Information

When we talk about information in the context of the business environment


there is a tendency to think in terms of facts and figures, reports and statements,

Copyright © Genetic Computer School, 2002


Page 58

print-outs or whatever appears on our computer terminal. It is relatively easy to


obtain this type of information, which may be generated as a by-product of
formal systems and processes already in place. While these forms of information
are important, there are other forms, less formal, which can also play an
important role in the decision making process.

Previous research has shown that the informal channels for collecting, what
is termed, 'soft' information can be very important, particularly in more senior
positions within the organization. Indeed, less regard is likely to be given to
'hard' factual information. This tends to reflect the way in which managers work
in the real world, where informal soft information (over the telephone or in the
canteen) is collected and stored until other items of information (either soft or
hard) reinforce or add to that information, like pieces in a jig-saw. The human
mind is extremely effective at this form of 'information management' far more so
than computers, where attempts to mirror this methodology have been relatively
unsuccessful.

So, an important lesson for the manager is not to focus solely on hard information
and associated systems but also to consider how to develop and nurture the informal
contacts and networks, which operate in and between organizations.

5.3 Problem Solving and Decision Making

Decision Making

Figure 5.3

Copyright © Genetic Computer School, 2002


Page 59

Management decision making can be organized according to the level of


managerial responsibility (refer to figure 5.3). At the top sits the strategic
management level. Their strategic decision making requires a wide variety of
information sources and the flexibility to model and present ideas in support of
their strategic mission. Information systems technology at the strategic level must
help to clearly communicate ideas and concepts. Information usually needs to be
summarized more highly than at lower levels of management support, and also
tends to use more "external" information coming from outside sources.

At the middle range, tactical matters require management to reach into


corporate records and obtain external data to implement and support company
strategy. This tactical management level of management decision making is
responsible for allocating resources to reach objectives and therefore requires
specific feedback concerning performance of operational units. Control of corporate
resources, such as performance monitoring and budget planning are part of their job.

At the lower management levels, daily operations are the chief concern. At
this operational management level, transaction processing and process control data
are the major concern, and information systems often highly automated and critical
for managing competitively. Beneath these systems, in the "basement" under the
management information system [MIS], the transaction processing and other
automated production processes churn out masses of data that must be stored,
processed, and converted into business intelligence that feeds into every level of
management.

Effective decision making requires an understanding of decision-making


processes including behavioral and communications aspects, the ability to identify,
obtain, store, and retrieve critical information necessary for making sound
decisions and the skill to use mathematical, statistical and simulation tools in
analyzing information in order to make decisions.

Structured decisions have a known and well-defined solution that uses fully
available data. The outcomes will always be the same. Examples of structured

Copyright © Genetic Computer School, 2002


Page 60

decisions are calculating net pay, calculating interest on a loan, etc. In nonstructured
(or unstructured and semi-structured) decisions, there is no agreement upon
procedure and not all the necessary data may be available. It requires judgement,
evaluation and intuition. There are few unstructured business decisions, many are
semi-structured or parts of the problem are structured and parts require judgement
and intuition.

Most business decisions are structured or semi-structured. Budget analysis is


structured and can be done by a spreadsheet program. Budget preparation is semi-
structured. Computers can analyze, past data and projections. Humans must judge
what they think is most likely to happen in future.

Each level has different types of decisions, therefore, measures the value of
information differently.

General Model of Decision Making

1. Intelligence [Finding out what is going, what the significant problems are]

2. Design [Identifying the decision(s) that need to be made]

3. Choice [Making the decision]

4. Implementation Go through an example of this process (Choosing classes


for the Fall)

Problem Solving

♦ Structured Problems

• solved by known methods and models

• all relevant data available

• require little judgement, evaluation or insight

• easily automated

Copyright © Genetic Computer School, 2002


Page 61

♦ Unstructured Problems ("Less structured" would be a better term)

• standard solution methods unknown

• some desirable data may not be available

• requires considerable human judgement, evaluation and creative insight

• difficult to automate

♦ Semi-structured Problems

• Some parts are structured, others are not

• Remember that there is a continuum from structured to unstructured.

Two important points to remember:

1. Structuredness is not the complexity of the problem -- complex problems may be


decomposed and approached in a highly structured manner.

2. All problems that can be approached in a rational manner are capable of some
degree of structuring.

MIS are often categorized according to the type of approach that they employ in
helping resolve managerial problems related to these activities. Managerial problem
solving range from the highly structured to the relatively unstructured approachs.
Structured problems are sometimes called "programmable" decisions because they are
easily formalized with models. Structured problems predominate in operations, where
data processing on transaction data can be coupled with process control data to permit
highly automated decision making. When such processes become very repetitive, they
may be fully automated, leaving humans to simply monitor the decision and outcomes.

On the other end of the range, unstructured problems require tools that permit
managers to model the dimensions of their problems, honing in on the best solutions to
difficult, and often unique, situations. However, no problem that confronts management

Copyright © Genetic Computer School, 2002


Page 62

is ever totally unstructured. If a problem can be identified, it must contain elements that
allow it to be structured to some degree. The amount of structuring that is done depends
upon the nature of the problem and the time and skill the manager has to devote to it. It
also depends upon the manager's own decision making style.

In the middle between the highly structured and the unstructured problems lie
many semi-structured problems in which parts, but not all of the problem can be
approached using formal decision making models. For example, the decision of
AquaPenn to acquire additional manufacturing facilities on the West Coast was based
partly upon a structured analysis of the logistics of supplying customers in that regoin
of the country. Solid data and knowledge to use it was available concerning the
logistics. However, parts of the problem, such as those involving the activities of their
competitors or the future demand for bottled water, required a less structured analysis of the
market for bottled spring water. Not all factors were known and not all data was "hard”.

5.4 Coping with Changes

How managers can help employee deal with change

Different employees have different needs. The most effective learned how to adapt
their style based on the individual's needs.

♦ "Surviving" (Innocent) Perspective:

• Employees in this level are very fragile and should be protected from change
as much as possible. Keep as much of their routine the same since any type of
changes - even what seem to be small ones - will be terrifying to them and
their reactions may be extreme.

♦ "Learning" (Rule-based) Perspective:

• Employees at this level need a "parent" figure to help them deal with the
changes - someone who can help them feel safe and secure with the changes as
well as help them adjust to the changes at a pace that they can handle.
Otherwise, they will create all sorts of problems for management.

Copyright © Genetic Computer School, 2002


Page 63

♦ "Competing" (Striving) Perspective:

• People at this level need to maintain their sense of self-worth through the
changes. If they feel threatened, they will fight back in whatever ways they can
find. If they can be recruited to have a feeling of personal importance in the
change, they will find it much easier to deal with. It is important to help them
understand that fighting the changes are not in their best interest, and that
working with others is the way they can benefit them the most.

♦ "Relating" (Partnership) Perspective:

• The challenge for people at this level is to feel that everyone's needs are being
taken care of. They will feel the pain of their co-workers' struggles to cope
with change, yet they may not know how to deal with their own feelings of
insecurity. They most value being part of a team and will usually want to help
others adjust to the changes. They may participate actively in the gossip about
changes. People at this level can assist by being asked to "partner" with
someone who is expected to have more difficulty in coping with changes.

♦ "Teaching (Integration) Perspective:

• People at this level will be able to understand the reasons behind the changes
and can be very helpful as a calming presence for people at other levels. They
can be most helpful to management if they are told about upcoming changes
and have time to help prepare others. They will be most effective in this role
when they can see the change(s) as a positive one and translate that message to
those who are more fearful. They will usually take a more philosophical
attitude toward the changes when they are told the larger picture and overall
strategy behind the changes. As with people in the "relating" perspective,
employees at this level can be very helpful if they are asked to help others cope
with the changes through a "partner" or "buddy" arrangement, whether formal
or informal.

Copyright © Genetic Computer School, 2002


Page 64

Note: "employee" can be any person who works in an organization - line, staff,
management or executive

Organizational Responses to Change

♦ Top Management:

Top management has a hard time coming to grips with the direct implications of
the change. They often underestimate the impact that change has on their employees.
They tend to isolate themselves. Often they engage in strategic planning sessions and
gather information in survey reports. They avoid communicating or seeking "bad
news," because it is difficult for them to admit "they don't know." They expect
employees to "go along" when a change is announced and they blame their middle
managers if people resist or complain about the change. They often feel betrayed when
employees don't respond positively.

♦ Middle Management:

Managers in the middle feel the pressure to "make organization change" according
to the wishes of top management. They feel pulled in different directions. Middle
managers often lack information and leadership direction to focus on multiple priorities.
They are caught in the middle, and often fragmented because they don't have clear
instructions. They feel besieged with upset, resistant or withdrawn employees who no
longer respond to previous management approaches, and deserted, blamed or
misunderstood by their superiors.

♦ Employees/Workers/Associates:

Workers often feel attacked and betrayed by changes announced by management.


They are often caught off guard, not really believing that "my company could do this to
me." Many respond with resistance, anger, frustration and confusion. Their response
can solidify into a wall of "retirement on the job." They become afraid to take risks, be
innovative or try new things. They experience a loss of traditional relationships,
familiar structure and predictable career advancement patterns.

Copyright © Genetic Computer School, 2002


Page 65

CHAPTER 6 - KNOWLEDGE REPRESENTATION

6.1 Introducton

Artificial intelligence techniques do not necessarily reproduce human thought


processes. Rather, they seek to apply computers to problem solving tasks requiring
intelligence. These techniques need to produce results consistent with human activity if
they are to be useful, however, they do not necessarily reproduce the reasoning of an
expert in their analysis. Expert systems represent the logical extreme of the trend away
from large generalized artificial intelligence models. By adopting a specific problem
domain, the tasks of knowledge representation and problem description are rendered as
simple as possible. This enables a more powerful model to be developed, since the
knowledge base applicable to the chosen problem domain can be made much more
comprehensive.

A detailed and specific knowledge base is the source of the problem solving
power of an expert system (Waterman [1986]). The manner in which the knowledge
base is used will be determined by the system's meta knowledge. Meta knowledge can
be seen as a set of rules governing how the knowledge base is applied. Meta knowledge
determines what case specific knowledge is obtained from the user and what knowledge
from the knowledge base is incorporated in the problem solving process. It also
provides the means to combine these two sources of knowledge for an expert system to
generate its output.

Meta knowledge and knowledge complement one another. Knowledge cannot be


applied in the absence of meta knowledge. However, reliance on meta knowledge may
be offset to the degree of reliance on a knowledge base. An advanced expert systems
shell facilitates the construction of expert systems based models that embody a high
degree of declarative meta knowledge. Such a model solves problems while making
"economical" use of the system's knowledge base. It limits its use to those pieces of
knowledge pertinent to the problem.

Copyright © Genetic Computer School, 2002


Page 66

A model with a less advanced meta knowledge structure may produce the same
conclusion, but it would be less discriminating in its use of the knowledge base. As a
result, it may incorporate many pieces of knowledge that were not strictly necessary in
its analysis. A procedurally coded computer model will necessarily place a heavier
reliance on its knowledge base than it will on a meta knowledge structure. This is not
necessarily a failing. Knowledge is typically easier to capture and represent than meta
knowledge. Expert systems are commonly constructed using proprietary shell packages.
These packages provide the tools to construct the knowledge base required by an expert
systems model and the inference engine and meta knowledge required to access that
knowledge base. The construction of an expert system requires the skills of a
knowledge engineer, as opposed to a conventional programmer. In contrast, decision
support systems can be constructed using only a conventional program code compiler.
A compiler embodies less functionality than an expert system shell, but as a result,
needs to be much less structured and can therefore allow the developer much more
freedom over the form of a decision support system. The programming skills required
to use a conventional program compiler are much more readily available and therefore
potentially more cost effective than the skills of a knowledge engineer.

The remainder of this paper is organized as follows. The second section discusses
the distinction between expert systems and decision support systems. The third section
reviews the application of expert systems to audit risk assessment. The fourth outlines
the constructiort of two computer models, one expert systems based and the other
procedural. The final section provides a summary and conclusion.

A well-known definition of artificial intelligence (AI) is the following: AI is


the science of making machines do things that would require intelligence if done by
humans. An important branch of Al research is expert systems (ES) or more
generally, knowledge based systems (KBS).

In the 1980s, there were a number of highly-publicized, successful expert


systems put into operation around the world (e.g., MYCIN, XCON, Prospector)
Despite these successes and the many operational expert systems in service today,

Copyright © Genetic Computer School, 2002


Page 67

the field is widely seen as having arrived, failed, and disappeared. However, it is
important to keep in mind that much this perceived failure is due to unrealistic
expectations fueled by hype rather than a lack of achievement by ES researchers
and developers. In fact, such a fate is not uncommon in information technology.

The basic idea behind expert systems is to create small, practical "intelligent"
systems by eliciting a knowledge base of rules from experts in a particular field
and providing non-expert users with means of accessing this knowledge easily. For
example, an early (circa 1975) expert system called MYCIN helps physicians to
diagnosis infectious blood diseases and prescribe medication. The knowledge base
of MYCIN was elicited from a large number of experts in the field and contained
an enormous amount of information about different types of bacteria associated
with blood diseases. Although MYCIN provided diagnosis as good or better than
any human expert, the technology itself never achieved widespread acceptance in
the medical community.

6.2 Knowledge Representation

Knowledge representation is crucial. One of the clearest results of artificial


intelligence research so far is that solving even apparently simple problems
requires lots of knowledge. Really understanding a single sentence requires
extensive knowledge both of language and of the context. For example, today's
(4th Nov) headline "It's President Clinton" can only be interpreted reasonably if
you know it's the day after the American elections. (Yes, these notes are a bit out
of date). Really understanding a visual scene similarly requires knowledge of the
kinds of objects in the scene. Solving problems in a particular domain generally
requires knowledge of the objects in the domain and knowledge of how to reason
in that domain - both these types of knowledge must be represented.

Knowledge must be represented efficiently, and in a meaningful way.


Efficiency is important, as it would be impossible (or at least impractical) to
explicitly represent every fact that you might ever need. There are just so many
potentially useful facts, most of which you would never even think of. You have to

Copyright © Genetic Computer School, 2002


Page 68

be able to infer new facts from your existing knowledge, as and when needed, and
capture general abstractions, which represent general features of sets of objects in
the world.

Knowledge must be meaningfully represented so that we know how it relates


back to the real world. A knowledge representation scheme provides a mapping
from features of the world to a formal language. (The formal language will just
capture certain aspects of the world, which we believe are important to our
problem - we may of course miss out crucial aspects and so fail to really solve our
problem, like ignoring friction in a mechanics problem). Anyway, when we
manipulate that formal language using a computer we want to make sure that we
still have meaningful expressions, which can be mapped back to the real world.

6.3 What is Artificial Intelligence?

Artificial intelligence (AI) is a broad field, and means different things to


different people. It is concerned with getting computers to do tasks that require
human intelligence. However, having said that, there are many tasks,. which we
might reasonably think require intelligence - such as complex arithmetic - which
computers can do very easily. Conversely, there are many tasks that people do
without even thinking - such as recognizing a face - which are extremely complex
to automate. Al is concerned with these difficult tasks, which seem to require
complex and sophisticated reasoning processes and knowledge.

People might want to automate human intelligence for a number of different


reasons. One reason is simply to understand human intelligence better. For
example, we may be able to test and refine psychological and linguistic theories by
writing programs, which attempt to simulate aspects of human behaviour. Another
reason is simply so that we have smarter programs. We may not care if the
programs accurately simulate human reasoning, but by studying human reasoning
we may develop useful techniques for solving difficult problems.

Copyright © Genetic Computer School, 2002


Page 69

AI is a field that overlaps with computer science rather than being a strict
subfield. Different areas of Al are more closely related to psychology, philosophy,
logic, linguistics, and even neurophysiology.

6.4 Is AI Possible?

Artificial intelligence research makes the assumption that human intelligence


can be reduced to the (complex) manipulation of symbols, and that it does not
matter what medium is used to manipulate these symbols - it does not have to be a
biological brain! This assumption does not go unchallenged among philosophers
etc. Some argue that true intelligence can never be achieved by a computer, but
requires some human property which cannot be simulated. There are endless
philosophical debates on this issue (some on comp.ai.philosophy), brought recently
to public attention again in Penrose's book.

The most well known contributions to the philosophical debate are Turing's
"Turing test" paper, and Searle's "Chinese room". Very roughly, Turing considered
how you would be able to conclude that a machine was really intelligent. He
argued that the only reasonable way was to do a test. The test involves a human
communicating with a human and with a computer in other rooms, using a
computer for the communication. The first human can ask the other
human/computer any questions they like, including very subjective questions like
"What do you think of this Poem". If the computer answers so well that the first
human can't tell which of the two others is human, then we say that the computer is
intelligent.

Searle argued that just behaving intelligently wasn't enough. He tried to


demonstrate this by suggesting a thought experiment (the "Chinese room").
Imagine that you don't speak any Chinese, but that you have a huge rule book
which allows you to look up chinese sentences and tells you how to reply to them
in Chinese. You don't understand Chinese, but can behave in an apparently
intelligent way. He claimed that computers, even if they appeared intelligent,

Copyright © Genetic Computer School, 2002


Page 70

wouldn't really be, as they'd be just using something like the rule book of the
Chinese room.

Many people go further than Searle, and claim that computers will never even
be able to appear to be really intelligent (so will never pass the Turing test). There
are therefore a number of positions that you might adopt:

• Computers will never even appear to be really intelligent, though they


might do a few useful tasks that conventionally require intelligence.

• Computers may eventually appear to be intelligent, but in fact they will


just be simulating intelligent behaviour, and not really be intelligent.

• Computers will eventually be really intelligent.

• Computers will not only be intelligent, they'll be conscious and have


emotions.

Computers can clearly behave intelligently in performing certain limited tasks, full
intelligence is a very long way off and hard to imagine. However, these philosophical
issues rarely impinge on Al practice and research. It is clear that AI techniques can be
used to produce useful programs that conventionally require human intelligence, and that
this work helps us understand the nature of our own intelligence.

6.5 Some AI Tasks

Human intelligence involves both "mundane" and "expert" reasoning. By mundane


reasoning I mean all those things which (nearly) all of us can routinely do (to various
abilities) in order to act and interact in the world. This will include:

• Vision: The ability to make sense of what we see.

• Natural Language: The ability to communicate with others in English or another


natural language.

• Planning: The ability to decide on a good sequence of actions to achieve your goals.

Copyright © Genetic Computer School, 2002


Page 71

• Robotics: The ability to move and act in the world, possibly responding to new
perceptions.

By expert reasoning I mean things that only some people are good at, and which
require extensive training. It can be especially useful to automate these tasks, as there
may be a shortage of human experts. Expert reasoning includes:

• Medical diagnosis.

• Equipment repair.

• Computer configuration.

• Financial planning.

Expert Systems are concerned with the automation of these sorts of tasks.

AI research is concerned with automating both these kinds of reasoning. It turns out,
however, that it is the mundane tasks that are by far the hardest to automate.

6.6 Knowledge Engineering

Having decided that your problem is suitable you need to extract the knowledge
from the expert and represent it using your expert system shell. This is the job of the
knowledge engineer, but involves close collaboration with the expert(s) and the end user(s).

The knowledge engineer is the AI language and representation expert. He/she


should be able to select a suitable expert system shell (and other tools) for the project,
extract the knowledge from the expert, and implement the knowledge in a correct and
efficient knowledge base. The knowledge engineer may initially have no knowledge of
the application domain.

To extract knowledge from the expert the knowledge engineer must first become at
least somewhat familiar with the problem domain, maybe by reading introductory texts or
talking to the expert. After this, more systematic interviewing of the expert begins.
Typically experts are set a series of example problems, and will explain aloud their

Copyright © Genetic Computer School, 2002


Page 72

reasoning in solving the problem. The knowledge engineer will abstract general rules
from these explanations, and check them with the expert.

As in most applications, the system is wasted if the user is not happy with it,
so development must involve close collaboration with potential users. As
mentioned in the introduction, the basic development cycle should involve the
rapid development of an initial prototype and iterative testing and modification of
that prototype with both experts (to check the validity of the rules) and users (to
check that they can provide the necessary information, are satisfied with the
systems performance and explanations, and that it actually makes their life easier
rather than harder!).

In order to develop the initial prototype the knowledge engineer must make
provisional decisions about appropriate knowledge representation and inference
methods (e.g., rules, or rules+frames; forward chaining or backward chaining). To
test these basic design decisions, the first prototype may only solve a small part of
the overall problem. If the methods used seem to work well for that small part it's
worth investing the effort in representing the rest of the knowledge in the same form.

Expert system development was very trendy around 5-10 years ago, with
unrealistic expectations about the potential benefits. Now some cynicism has set
in. Expert system shells are in fairly wide use, but are often used to solve fairly
simple problems, and are chosen as much for their user interface and development
environments as for their inferential abilities.

Copyright © Genetic Computer School, 2002


Page 73

CHAPTER 7: PLANS AND STRATEGIES

7.1 Strategy Planning for Information Systems

History

Perhaps, because of the relative 'youth' of computing in the business


environment and coupled with its associated pace of change, the track record of IS, as
far as effective strategic planning is concerned, has been patchy. There are many well-
documented examples of poor strategic decisions and costly failures. Surveys indicate
that, until comparatively recently, a high percentage of organizations failed to plan
strategically in this area and those that did were not particularly effective. The
strategic planning process demands regular review and reassessment. However, as
stated in the text, major corporate changes i.e. acquisitions, alliances, etc. or radical
changes in the market which the organization operates in, will hasten these.

Consider the following questions:

• Where does an IS strategy fit within the wider set of strategies?

• What has been the history of IS strategy planning?

• What circumstances demand major reassessment of IS strategy plans?

• Who might be employed to do the actual planning?

• What might an IS strategy plan contain?

Firstly, in its supporting role, the IS strategy has to be clearly linked with the
corporate and business goals of the organization. Its constituent parts will consider
specific issues, often viewed as sub-strategies, in considerable detail. As an example,
the software sub-strategy may include a detailed software replacement programme,
identifying which systems will be replaced and when, together with the resources
required at each stage. This sub-strategy will have been determined by reference to
the overall business strategy, as well as determining factors at the IS strategy level

Copyright © Genetic Computer School, 2002


Page 74

such as 'a decision to source future systems from off the self rather than developed in-
house'. This would also be reflected in the sub-strategy dealing with I.T. staffing.

Key Players

Much has been written about who the key players are in IS strategy formulation
and a preliminary stakeholder analysis may give the clue as to who should be
involved. The culture of the organization - centralized or devolved - will also act as a
determining factor. IT directors have seen their fortunes wax and wane over the years
as more and then less were afforded Board level status. The concept of the 'hybrid
manager' has been adopted in many organizations.

Components of an IS Strategy

Undoubtedly, there is no one arrangement, which best suits all. Similarly, the
content of an IS strategic plan may very well be a 'mix and match' most appropriate to
the organization concerned - although elements of 'best practice' are well described in
the text and include the following:

• A clear statement of the IS objectives.

• An inventory and assessment of current organizational capabilities.

• An implementation plan.

7.2 IS and Bisiness Strategy- Alignment and Intergration

Rationale

There is little contrary argument to the proposition that working to a well-


developed plan offers greater chance of achieving desired -objectives than working in
an intuitive, reactionary mode. Not least, a structured plan provides a map to refer to
when the actual position is different to that anticipated or desired. Information systems
are costly to develop and manage, in common with the other resources the organization
has to use. More importantly perhaps than with other assets, its life-cycle is more

Copyright © Genetic Computer School, 2002


Page 75

indeterminate and potentially has a greater impact on the well-being of the parent
organization.

Consider an overview of the strategic planning process as it relates to the


organization and you will (will have) encountered many of the models elsewhere in the
Programme. Determining the business strategy is fundamental to the subsequent
development of an appropriate supporting IS strategy. As a consequence, before
developing an organization-wide IS strategy, there must be a clear understanding of the
external business environment together with an assessment of the organization's internal
'condition'. Having established the credence for organizations adopting an IS strategy,
we need to consider which planning process is the most appropriate for a particular
organization. The basic components of an IS strategy will be common to most
organizations. However, the methodology or framework by which it is developed is not.

Framework

The framework or methodology adopted to develop the IS strategic plan will


contain the following elements:

• A structure that gives guidance on what to do and when to do it.

• A definition of techniques to do what needs to be done.

• Advice on how to manage the quality of the results.

• Tools to automate the process.

It has to be emphasized at this point that frameworks are not always mutually.
However, a particular approach may lend itself to the style and culture of an
organization.

Copyright © Genetic Computer School, 2002


Page 76

7.3 Analytical Tools for IS Strategic Planning

Introduction

There is a real danger in attempting to use too many tools and techniques in
formulating an IS strategy and the predominant framework, as discussed in the previous
session, ought to give a firm steer as to which ones may be most appropriate.

Tools

The value of each of the tools described is in their ability to clarify, to a greater or
lesser extent, a complex situation. Often, they also force the planner to consider a range
of issues from a different perspective. As a consequence, 'Beauty' - the value of a
particular tool - is very much 'in the eye of the beholder'. For example, the layered
structure and inter-connectedness of Rockart's Critical Success Factors is immediately
attractive to someone coming from a highly stratified and bureaucratic organizational
background but would be alien to someone from an informal and innovative
organization such as 3M.

7.4 MIS Strategic Planning

Strategic Planning and Information Technology

Strategic planning is the first step in initiating organizational change, as well as an


essential element of business process reengineering. The strategic planning process
includes an organizational assessment and the development of strategic foundations:
mission, future vision, and guiding principles. Analyzing the gap between the current
state of the organization and its future vision provides essential information for
developing strategic goals, specific strategies, and objectives.

Each strategy may have several objectives. It is at the level of objectives that it is
reasonable to introduce performance measures to gauge progress toward achieving the
objectives and, therefore, the strategies and goals.

Copyright © Genetic Computer School, 2002


Page 77

Projects are typically the links between plans and the budgeting process. As they are
used in this document, projects are undertakings directed at the accomplishment of an
objective. Projects, with cost and schedule estimates, may implement new information
systems or major improvements to existing systems. The completion of projects should
mark progress toward reducing the gap between an organization's current state and its
future vision.

Strategic Planning Process

The literature on the strategic planning process is extensive; the references provide
some recommended starting points. The ITSC has found two documents to be especially
helpful. Wells, [1995] have evolved a structured strategic planning process that is
appropriate for public sector organizations. While many variations of this process are
possible, it does provide a basic framework.

An overview of the planning process, as described by Wells, is presented in figure 7.1.


The process begins with pre-planning activities designed to prepare senior management
for strategic planning. These activities focus on gathering data about the organization and
assessing the current state of the organization. In a workshop setting, senior leaders
develop, through a consensus building process, the mission, the future vision, and the
guiding principles of the organization. These elements represent the strategic foundation,
but the elements by themselves are insufficient as a plan. Strategic goals, strategies, and
objectives are necessary to guide deployment and implementation.

To develop the goals, strategies, and objectives, managers work through a gap
analysis to identify what needs to be changed to move the organization from its current
state to the new desired state. This gap analysis further illustrates the relationships
between strategic planning and typical business process reengineering (BPR) activities,
during which the "as is" and "to be" states are described.

Copyright © Genetic Computer School, 2002


Page 78

Adapted From Denise Wells, et al., Strategic Management: Using the


Strategic Plan to Build Organizational Budgets, Workshop on Performance
Measuring and Monitoring in Government San Francisco, 1995.
Figure 7.1

Figure 7.1 shows that developing goals, strategies, and objectives marks the
completion of the strategic plan. Understandably, the calendar time and effort to
develop the plan will vary among organizations. In the Wells approach, the strategic
foundations -- the vision, mission, and guiding principles -- are developed in a 3-day
workshop with trained facilitators.

After the workshop, top management usually seeks the assistance of other
members of the organization to complete the goals, strategies, and objectives, based on
the strategic foundations.

After completion of the plan, it is published and distributed throughout the


organization. Implementation of the strategic plan occurs when action plans (i.e.,
projects, with cost and schedule estimates) are developed to accomplish the objectives.
The projects are developed and implemented by personnel throughout the organization.

Copyright © Genetic Computer School, 2002


Page 79

Implementation efforts are monitored and measured so that the organization can
evaluate progress toward achievement of its goals, strategies, and objectives.
Information from this evaluation is used as input to the next iteration (see Figure 7.1) of
the strategic planning process where the strategic plan is validated and updated based
on changed conditions. An important by-product of this process is team-building and
the breaking down of functional barriers. Team building is a natural by-product because
senior managers are discussing the future of the organization and how they can co-
operate towards achievement of that future. The managers have agreed to work together
for the overall aim of the organization rather than their departments. This leads to
organizational alignment and the breaking down of barriers among departments.

Strategic Goals

The strategic goals are broad statements about where the organization wants to be
at some point in the future. These goals work towards achieving the overall mission of
the organization and help achieve the vision of the future. The strategic plan usually
contains goals in multiple strategic areas.

The strategic goals are sometimes confused with the future vision. Whereas the
future vision is a single statement or paragraph that describes a desired organizational
end-state at some point in the future, the strategic goals are general statements of
activities that if pursued would lead to the attainment of the future vision. In this
context the future vision is a statement about where to go, and the strategic goals are
statements about how to get there.

Progress in achieving these goals should be measurable, hence enabling


quantitative performance measurement.

Objectives and Performance Measures

Objectives represent specific courses of action that are bounded by and support the
strategies. They contain a target and a performance measure. In the absence of targets
and performance measures, it is impossible to measure progress towards achieving the
strategic goals.

Copyright © Genetic Computer School, 2002


Page 80

Objectives follow from the "top-down" approach of IM strategic planning. Goals


describe the end state to be achieved; strategies outline how to get there; and objectives
describe who does what and when. Once the outcomes and the methods for achieving
those outcomes have been established, indicators to measure performance must be
developed. Implementation, however, is a "bottom-up" activity; it is around the
achievement of objectives that data are, collected and performance is measured.

Measuring progress in achieving strategic goals is never easy; however, for


government to be accountable, performance measurement is important. Performance
measures are quantifiable expressions that measure the achievement of program
objectives. Multiple indicators are often needed to capture the intent of a program. The
measures must reflect the desired result intended, and they must possess the qualities of
validity and sufficiency.

A valid performance measure is one that reasonably represents the program


objective. A valid performance measure should not measure workload data, and it
should minimize the measurement of change not attributable to the program.
Sufficiency implies that the number of indicators adequately reflects the intent of the
program objective. A sufficient measure provides a comprehensive measure of
effectiveness. Additional indicators are needed if they provide new information that is
pertinent to the assessment of program effectiveness.

7.5 Strategic Information Systems

Introduction

The topic of 'strategic information systems' is concerned with systems which


contribute significantly to the achievement of an organisation's overall objectives. The
body of knowledge is of recent origin and highly dynamic, and the area has an aura of
excitement about it. It is risky to attempt a historical exposition of such a recently
emerged topic. On the other hand, the line of development which the conventional
wisdom has followed is itself interesting and instructive. This paper is prepared as an
introduction to the literature, but embodies interpretation in both its structure and its

Copyright © Genetic Computer School, 2002


Page 81

expression, and should therefore be read with at least as critical a disposition as any
other paper in the area. It may also be compared with other critical interpretations such
as Swatman & Swatman (1992), Galliers (1993) and Ciborra (1994).

The notion and its origins are first discussed. The emergence of the key ideas is
then traced. The process whereby strategic information systems come into being is
assessed. Finally, areas of weakness are identified, and directions of current and future
development suggested.

Origins

The role of Information Systems (IS) has developed during the years. The original
conception was of the automation of existing manual and pre-computer mechanical
processes. This was quickly succeeded by the rationalisation and integration of systems.
In both of these forms, IS was regarded primarily as an operational support tool, and
secondarily as a service to management.

During the 1980s, an additional potential was discovered. It was found that, in
some cases, information technology (IT) had been critical to the implementation of an
organisation's strategy. The dominant sense in which the term is used is that a strategic
information system (SIS) is an information system which supports an organisation in
fulfilling its business goals.

An alternative interpretation of the term is that it is not necessarily a particular IS,


but rather the combination of those parts of an organization's cluster of information
systems which provide information into its strategic planning processes (Higgins &
Vincze 1993.p.93). The functions involved include the gathering, maintenance and
analysis of data concerning internal resources, and intelligence about competitors,
suppliers, customers, government and other relevant organisations.

A variety of interpretations of strategy exist, most of which have a great deal to do


with competition between corporations. Chamberlin's theory of monopolistic
competition sees corporations as being heterogeneous, and competing on the basis of
asset differences, such as technical knowledge, reputation, ability for teamwork,

Copyright © Genetic Computer School, 2002


Page 82

organizational culture and skills, and other 'invisible assets' (Chamberlin 1933, Itami
1987). Competition therefore means cultivating unique strengths and capabilities, and
defending them against imitation by other firms. Another alternative sees competition
as a process linked to innovation in product, market, or technology (Schumpeter 1950).

Porter's Strategic Theory

The context within which SIS theory emerged was the competitive strategy
framework put forward by Porter (1980, 1985), which was based on industrial
organisation economics. For developments along that path, see Kaufmann 1966,
Kantrow 1980, Pyburn 1981, Parsons 1983, EDP Analyzer 1984a, 1984b, McFarlan
1984, Benjamin et al 1984, Wiseman & Macmillan 1984, Ives & Learmonth 1984, Cash
& Konsynski 1985, Porter & Millar 1985, Keen 1986, King 1986). This first section
outlines the basis of that theory. It will then be shown how Strategic information
systems theory is concerned with the use of information technology to support or
sharpen an enterprise's competitive strategy.

Competitive strategy is an enterprise's plan for achieving sustainable competitive


advantage over, or reducing the edge of, its adversaries. In Porter's view, the
performance of individual corporations is determined by the extent to which they cope
with, and manipulate, the five key 'forces' which make up the industry structure:

• the bargaining power of suppliers;

• the bargaining power of buyer;

• the threat of new entrants;

• the threat of substitute products; and

• rivalry among existing firms.

Porter's classic diagram representing these forces is reproduced in Figure 7.1.


Enterprises, through their strategies, can influence the five forces and the industry
structure, at least to some extent.

Copyright © Genetic Computer School, 2002


Page 83

There are two basic strategic stances that enterprises can adopt:
• low cost; and
• product differentiation.
In the long run, firms succeed relative to their competitors if they possess
sustainable competitive advantage in either of these two, subject to reaching some
threshhold of adequacy in the other.
Another important consideration in positioning is 'competitive scope', or the
breadth of the enterprise's target markets within its industry, i.e. the range of product
varieties it offers, the distribution channels it employs, the types of buyers it serves,
the geographic areas in which it sells, and the array of related industries in which it
competes.
Under Porter's framework, enterprises have four generic strategies available to
them whereby they can attain above-average performance. They are:
• cost leadership;
• differentiation;
• cost focus; and
• focused differentiation.
Porter's representation of them is reproduced in Figure 7.2.
Figure 7.2

Copyright © Genetic Computer School, 2002


Page 84

According to Porter, competitive advantage grows out of the way an enterprise


organises and performs discrete activities. The operations of any enterprise can be
divided into a series of activities such as salespeople making sales calls, service
technicians performing repairs, scientists in the laboratory designing products or
processes, and treasurers raising capital.

By performing these activities, enterprises create value for their customers. The
ultimate value an enterprise creates is measured by the amount customers are willing to
pay for its product or services. A firm is profitable if this value exceeds the collective
cost of performing all of the required activities. To gain competitive advantage over its
rivals, a firm must either provide comparable value to the customer, but perform
activities more efficiently than its competitors (lower cost), or perform activities in a
unique way that creates greater buyer value and commands a premium price
(differentiation). (Refer to figure 7.3)

Figure 7.3

Many differentiation bases exist, classified into four major groups (Border 1964,
quoted in Wiseman 1988):

• product (quality, features, options, style, brand name, packaging, sizes,


services, returns);

• price (list, discounts, allowances, payment period, credit terms);

• place (channels, coverage, locations, inventory, transport); and

Copyright © Genetic Computer School, 2002


Page 85

• promotion (advertising, personal selling, sales promotion, publicity).

IT can be used to support or sharpen the firm's product through these various attributes.

Of especial importance is 'product differentiation'. This is the degree to which


buyers perceive products from alternative suppliers to be different, or as it is expressed
by economic theory, the degree to which buyers perceive imperfections in product
substitutability. The buyers of differentiated products may have to pay a price when
satisfying their preference for something special, in return for greater added-value. The
connection between the producer and buyers may be reinforced, at least to the level of
customer loyalty, and perhaps to the point of establishing a partnership between them.
Such a relationship imposes 'switching costs' on the buyer, because its internal
processes become adapted to the beneficial peculiarities of the particular factor of
production, and use of an alternative would force internal changes. Hence product
differentiation also serves as an entry barrier. In addition, a continuous process of
product differentiation may produce an additional cost advantage over competitors and
potential entrants, through intellectual property protections, such as patents, and the
cost of imitation.

The activities performed by a particular enterprise can be analysed into primary


activities, which directly add value to the enterprise's factors of production, which are
together referred to as the 'value chain', and supporting activities. Figure 7.4 reproduces
Porter's
diagram.

Figure 7.4

Copyright © Genetic Computer School, 2002


Page 86

The primary, value-adding activities include those involved in the production,


marketing delivery, and servicing of the product. They are linked, generally in a
chain. Support activities include those providing purchased inputs, technology,
human resources, or overall infrastructure functions to support the primary activities.

By co-ordinating linked activities, an enterprise should be able to reduce


transaction costs, gather better information for control purposes, and substitute less
costly operations in one activity for more costly ones elsewhere. Co-ordinating
linked activities is also an important way to reduce the combined time required to
perform them. Hence co-ordination is increasingly important to competitive
advantage. Gaining competitive advantage requires that an enterprise's value chain
be managed as a system rather than as a collection of separate parts. Reconfiguring
the value chain, by re-locating, re-ordering, re-grouping, or even eliminating
activities is often at the root of a major improvement in competitive position.

An enterprise's value chain for competing in a particular industry is embedded


in a larger stream of activities that Porter terms its 'value system', but which might
be more usefully referred to as the 'industry value-chain'. This includes suppliers
and distribution channels. Figure 7.5 reproduces Porter's representation.
Competitive advantage is a function of how well a company can manage the entire
industry value-chain. A corporation can create competitive advantage by co-ordinating its
links in that chain.

An enterprise's activities are subject to influence from:

Figure 7.5 Porter’s Industry Value-Chain

Copyright © Genetic Computer School, 2002


Page 87

• New technologies. These may alter the path of the value chain, e.g. the
invention of semiconductors forced many vacuum-tube producers out of
business, and the printing and publishing industries are currently confronted by a
major upheaval;

• New or shifting buyer needs. Customers are demanding the convenience and
consistency offered by fast-food chains. This in turn influences related
market segments;

• Change in industry segmentation. The disappearance of old intermediaries and


the emergence o new ones create the potential to substantially reconfigure the
value chain. Enterprises that fail t adjust will be forced out;

• Shifts in the costs or availability of factors of production. Competitive


advantage can be gained b optimizing based on current conditions. On the
other hand, enterprises saddled with assets and approaches tailored to outdated
modes of operation suffer;

• Change in government regulations. Changes in product standards,


environmental controls, restrictions on entry to the market, and trade barriers
all affect an enterprise's performance.

7.6 Three key success factors common to effective business

First is integrating the network into the organization's business strategy thus
ensuring the flexibility that today's businesses require. New business opportunities
require new application that need network support. An integrated planning process
ensures that the network can meet new demands and support new directions.

The second factor is the homogeneity of services the network provides. An


enterprise network facilitates information exchange among all authorized subscribers.
Homogeneity refers to the ease and commonality of use wherever a customer connects
to the network, which is especially important to subscribers who travel, use devices at
other locations, or connect via remote services

Copyright © Genetic Computer School, 2002


Page 88

Finally, an effective enterprise network is a strategic, resource used for


competitive advantage. Telecommunication industry publications feature numerous
examples of organizations using networks to gain competitive advantage.

Competition in the last half of the 1990's and beyond requires that all business
investment provide benefits greater than their cost, networking is no different. Planning
any major business investment is challenging; however, planning the enterprise network
is especially complex. Network planners often, face several constraints that limit their
progress. For example, most organizations large enough to consider data networks have
already committed to a computing strategy, which may confine network planning
options. Many have already implemented data networks, and the financial inertia of
existing systems limits planning mobility. Additionally, new computing architectures,
machine platforms, and business reengineering make network planning little like
building an ocean liner in mid-voyage. Network planners must consider the constraints
as they develop their plans and evolve towards an effective enterprise network.

Today's businesses put much greater demands on the network than those of just a
few years These increasing demands are coupled with unprecedented technological
advances, service offerings, and vendor selection options. Uniting the business and IT
plans with an appropriate enterprise network is becoming increasingly difficult.

Copyright © Genetic Computer School, 2002


Page 89

CHAPTER 8: WOrKGROUP INFORMATION SYSTEMS

8.1 Introduction

A major class of new information systems is emerging. These systems are


group productivity systems designed to enable workers to collaborate on a series of
information-based tasks. Workgroup computing offers the potential to dramatically
improve the effectiveness and productivity of organizations by radically improving
the coordination of complex tasks which require cross-functional teams to work
together. These systems must balance the need to structure and coordinate work
with the seemingly conflicting need to support flexible communication between
knowledge workers.

Workflow systems are the flagship information system within the class of
group productivity software. Business processes are dynamic in nature. As the
market changes, as the organization changes, as regulations change, the business
process must change in response. Workflow systems are the essential tools which
an organization must have to respond to in this changing environment.

Workflow systems must have several attributes to enhance the effectiveness


of workgroups:

• Rapid Application Generation: Rapid application development strategies


enable organizations to quickly define a new workflow, to simulate its
operation, to visualize the new workflow graphically, and to generate new
workflow process definitions through a series of iterations.

• Implication Independence & Extensibility: Workflow products must


support comprehensive line-of-business applications from such wide-
ranging areas as insurance and bank loan processing to engineering change
management and process safety management. To achieve this breadth of
support, the workflow product must be completely independent of

Copyright © Genetic Computer School, 2002


Page 90

application requirements for data definition, process routing, and


applications interfaces.

• Robust Workflow Engine: The essence of a workflow product is the


underlying, which supports process definition, scheduling, and routing. To
completely support application requirements the workflow engine must
support several key functions such as job definition, security, parallel and
ad hoc routing, drill-down workflows, event monitoring, and simulation.

• Modular Product Design: A modular product architecture with well-


defined interfaces between subcomponents ensures that compliance with
emerging workflow standards can be achieved at reasonable cost.

• Environmental Independence: The underlying technology environment in


which workflow products must operate continues to change rapidly. Well-
designed workflow products must operate independently of key technology
infrastructures such as database, server, and network.

Workflow systems operate not in isolation but rather within an information


systems infrastructure. With the Internet and especially internal corporate Intranets
increasing in popularity, the emerging battle between traditional Client/Server
configurations and the Internet/Intranet is likely to last for a few years with no
conclusive winners. Each technology infrastructure has its own significant
advantages. Workflow products which work across these major groupware environments
will become essential elements of the emerging information infrastructure required
to. manage the global enterprise. Workflow products which work in both
traditional client-server environments as well as with the Internet/Intranet will
provide the best of both worlds, taking full advantage of each world while offering
the buyer the opportunity to use a common workflow product across the corporation.

The basic idea behind groupware is said to be moving information, not people by
using software supported "intentional group processes" aimed at what one commentator
defines as "proactive analysis, compression and automation of information based tasks

Copyright © Genetic Computer School, 2002


Page 91

and activities". Group Decision Support Systems (GDSS) are defined as interactive
computer systems supporting a group's formulation and solution of "unstructured
problems". Workflow is defined as a "process driven" way of managing a series of tasks
defined by procedures related to the flow of documents through organizations. It is
noted that one of the biggest problems in the workflow world is the lack of a "succinct
definition", although a March 1996 report identifies four types of workflow:
Production, Collaborative, Administrative and Ad hoc.

At the beginning of 1994, Groupware is seen as being challenged by


multifunctional bundles of integrated applications software. A distinction is made
between workflow and groupware on the basis that the former are designed for "simple
and repetitive activities", while the latter tried to address more complex and variable
organisational processes, although comparisons in the field are felt to be problematic.
Notes is portrayed as being good at dealing with "free form and semi-structured"
information and is praised especially for its "extended cross platform support" and
flexibility, but is said not to be "robust enough" for transaction processing or activity
management. News of Notes add-on products, especially products enabling seamless
connections between Notes and other applications is regularly reported, while Notes
upgrades feature automatic document versioning, full text search, "security features"
and an object store for version 3.x, and promises of "sleeker" user interface, outside
company connections, agent technology and an object oriented scripting language for
version 4.0. Collabra described as a "discussion database" and "an excellent,
inexpensive approach to groupware" is reported being integrated with Novell's
GroupWise suite. IBM and Oracle are said to be releasing workgroup applications
based on enhancements to DB2 and Oracle relational databases, while BeyondMail and
ClipBook are portrayed as extending messaging functions to encompass groupware
applications. As workflow applications are characterised as becoming "increasingly
sophisticated", GUI and OO-based, object-orientation is said to be enabling the creation
of new generation of more easily customisable groupware. At the end of 1994, reports
begin to appear on multimedia groupware --initially with voice, then video.

Copyright © Genetic Computer School, 2002


Page 92

8.2 The Evolution of Workgroup Computing

Information systems initially evolved for the purpose of processing transactions.


Transaction processing systems emphasized the efficient and reliable recording of
business events such as customer order receipt and invoicing, inventory control, etc.
Mainframe-based computer systems grew from this need, and even today they manage
the majority of business transactions. The dramatic improvements in processing power
and the emergence of the personal computer gave rise to systems that expand personal
productivity. Significant software innovation began to be focused not just on
transaction management, but also on task automation for individual workers. The use of
desktop publishing packages by graphics designers and computer-aided design systems
by product designers are merely two examples of personal productivity systems.

Today, a third major class of information systems is emerging. These systems


increase group productivity. They are designed to enable workers to collaborate on a
series of information-based tasks. Workgroup computing offers the potential to
dramatically improve the effectiveness and productivity of organizations by radically
improving the coordination of complex tasks, which require cross-functional teams to
work together. Such tasks as product design, customer support, and regulatory
compliance require systems to enable cross-functional teams to communicate,
collaborate, and coordinate their work. These systems must balance the need to
structure and coordinate work with the seemingly conflicting need to support flexible
communication between knowledge workers.

A convergence of technologies is shining the spotlight on workgroup computing. -


Technology advances such as reliable networking, relational databases, powerful
workstations, and internal corporate Intranets have provided a backbone to make
enterprise-wide workgroup computing possible. At the same time, dramatic
marketplace changes are flattening organizations. These changes are forcing rapid
product development cycles, and accelerating the demand for global information
networks. As the need for communication and coordination across organizational
boundaries accelerates, so does the requirement for developing workgroup computing

Copyright © Genetic Computer School, 2002


Page 93

systems respond to these needs in dramatic fashion. Indeed, workgroup computing


offers the potential to finally deliver the organizational productivity gains so long
promised by the information age.

8.3 The Evolution of Workflow Market

Early workgroup computing systems began to appear in the mid-1980s. The focus
of these systems, not surprisingly, was on the automation of mission-critical business
processes, which are repetitive in nature. These systems became known as "workflow
systems", with FileNet Corporation becoming the clear market leader in automating
repetitive business processes. In reality, these systems are simply transaction processing
systems that automate paper processing in addition to data processing. Early
applications have been bank loan processing, insurance claims processing, and records
management archiving.

These systems streamlined the process of managing paper-based transactions.


However, they were not without their limitations:

1. Transaction-based workflow systems are focused on streamlining a single


business transaction, which flows through an organization in a sequential and
predictable order. They place great focus on automating routine tasks at the user
workstation, but are not well suited to business processes which are less orderly. As an
example, the processing of a loan application in a bank is reasonably predictable. On
the other hand, the processing of an engineering change order in a manufacturing firm
is much less predictable and may in fact need to be routed to different groups based on
the type of product, the nature of the change, or the financial magnitude of the project.
Thus, transaction-based workflow systems optimize the automation of repetitive tasks
and "paper mills" characterized by clerical workers. They are not well-suited for the
automation of collaborative businesses processes driven by knowledge workers.

2. Most transaction-based workflow systems provide little or no document


management capability. This is of little consequence when the basic unit of work is a
piece of paper being routed through an organization; however, most collaborative

Copyright © Genetic Computer School, 2002


Page 94

business processes are document-intensive rather than transaction-intensive. In


reviewing engineering contract changes or large insurance policy changes, there is a
constant need to check documents into and out of archives in a controlled manner.
Thus, the workflow process must be fully integrated with a system which provides
document management and revision control capabilities. This can be provided by
integrating the workflow system with a document management system. However,
ideally this should be accomplished through one system providing both document
management and workflow, so that workers performing collaborative tasks can share
one user interface in an easy-to-administer system- a system where the user
organization does not have to maintain coordination between two systems when each
one comes out with a new release.

3. Transaction-based workflow systems require significant customization, often


through high-level programming (4GL or "C" type), to "add the application to the work
flow". This effort was certainly productive when automating mission-critical, repetitive
business processes, which rarely change from year to year. However, for less
predictable, more collaborative business processes the process of system definition and
"business process reengineering" must be more iterative. This results in an absolute
requirement for a new class of "collaborative workflow" system, which supports the
rapid prototyping of new business processes with little, or better yet, no programming.
Of course, such rapid application development facilities would also be beneficial in
repetitive application building as well.

4. Finally, transaction-based systems typically use server-based architectures,


which maintain much of the workflow logic on the system server. These systems have
proven difficult to migrate into the latest advances in client-server technology, and
therefore have often lagged other systems in their support of advances in networking
environments, client workstation technologies, and application development
environments. Newer generation workflow systems have placed much of the logic in
the client environment, thus enabling the system to take advantage of the increasing
processing power of the client workstation while using the SQL and ODBC standards to
support a wide variety of relational databases.

Copyright © Genetic Computer School, 2002


Page 95

As the workflow market has evolved, systems have tended to specialize in one of
three types of applications:

• Transaction-based workflow systems for automation of repetitive, paper-based


business transactions.

• Collaborative workflow systems for the automation of the more fluid, mission-
critical, knowledge-based business processes.

• Ad hoc workflow systems which are enhanced e-mail systems to route


documents "on the I fly", but typically lack the ability to structure, coordinate,
and manage the business process.

A critical weakness of Ad hoc systems is that they are completely user-driven


rather than being driven by the organization as a whole. With a repetitive or
collaborative workflow model, the corporation can establish a new process definition,
modify the process as new knowledge is generated, and capture the benefits of process
re-engineering. With an ad hoc model, all process knowledge is "user-dependent" and
leaves the organization if the user departs.

This market structure is depicted in Figure 8.1 below:

Workflow Market Structure


(Exhibit 1)

Figure 8.1

Copyright © Genetic Computer School, 2002


Page 96

8.4 The Essential Attributes of Workflow as a Strategic Technology

To enhance the effectiveness of workgroups, workflow systems must include


several attributes which are often not apparent from a superficial analysis of product
functions and features. These; attributes are crucial to ensuring that the workflow
product can support the underlying business process which it is automating over the life
cycle of the business process.

8.5 Rapid Application Generation

Business processes are dynamic in nature. As the market changes, as the


organization changes, as regulations change, the business process must change in
response. The workflow system must be a tool to support the underlying results, which
the organization is attempting to achieve, rather than the other way around. To support
business processes, which are subject to rapid changes, workflow systems must support
rapid application generation.

The workflow product must provide an easy to use, visual, point and click
interface for configuring screens and building workflow scripts. Screen building, script
building, and workflow should be integrated into a single user interface.

The workflow product must be application independent, that is, it must not impose
a particular data or routing structure. The rules, which define the organization structure,
document structure, and process routing must be completely customizable by
authorized users. Furthermore, the user must be able to change any of these workflow
definition structures through an intuitive interface, which does not require programmers
to execute.

The workflow product must support application frameworks, which provide


building blocks that enable the user organization to quickly generate new applications
as well as modifications to previously defined applications.

The workflow product should be extensible from market-leading application


development environments such as Visual Basic. Developers operating in these

Copyright © Genetic Computer School, 2002


Page 97

environments should be able to define workflow objects with all of the key capabilities
of the underlying workflow product.

Rapid application development strategies enable organizations to quickly define a


new workflow, to simulate its operation, to visualize the new workflow graphically, and
to generate new workflow process definitions through a series of iterations.

8.6 Application Independence & Extensibility

Workflow products must support comprehensive line-of-business applications


from such wide-ranging areas as insurance and bank loan processing to engineering
change management and process safety management. Figure 8.2 depicts common line-
of-business applications, which are greatly enhanced by collaborative workflow:

Common Line-of-Business Applications


(Exhibit 2)

Engineering Change Management


Commercial Loan Processing
Legal Contract Management
Process Safety Management
Regulatory Product Approval
Insurance Policy Underwriting
ISO 9000 Compliance
Construction Project Management
Collaborative Workflow Systems

Figure 8.2

To achieve this breadth of support, the workflow product must be completely


independent of application requirements for data definition, process routing, and
applications interfaces. The objective of the workflow product is to provide an
underlying framework for screen building, document routing, process definition,

Copyright © Genetic Computer School, 2002


Page 98

and process measurement without imposing requirements on the application. The


workflow product must contain a full set of Application Program Interfaces (APIs)
to enable third-party applications to easily integrate with all of the functions of the
workflow product.

OLE automation is a key technology for providing integration with a variety


of Windows based document processors. An innovative approach to APIs is to use
OLE to enable the workflow vendor's product to work within OLE enabled
applications. Most leading development environments, such as Visual Basic, allow
developers to make calls directly from their scripts to any other application that
provides an OLE API.

8.7 Robust Workflow Engine

The essence of a workflow product is the underlying engine which supports


process definition, scheduling, and routing. To completely support application
requirements the workflow engine must support several key functions:

• Job Definition: The ability to define the workflow job as a series of tasks
and link each task to a set of task completion rules.

• Security: Each workflow task must be linked to the organizational groups


and individuals who can view, modify, or complete the task. The product
should provide multiple user-definable levels of security at the database
record level. The security rules should also be definable at the workflow
step level so that security rules can be modified as the workflow proceeds.

• Process Modification: Authorized users must be able to change workflow


jobs in process without altering the baseline job definition.

• Parallel routing: Collaborative workflow can be streamlined by routing the


workflow task to multiple users at once, - taking advantage of the existence of
electronic rather than paper documents. To achieve this the workflow engine

Copyright © Genetic Computer School, 2002


Page 99

must support parallel document routing and the ability to re-synchronize the
workflow process based on managing the document approval cycle.

• Ad hoc routing: In addition to parallel routing, authorized users must also be


able to define flexible routing paths, as many workflow routings are
"conditional" on business events and user decision-making.

• Approval rules: Complex business processes often require the workflow


engine to support sophisticated approval rules including approval delegation,
limits `to approval authority, and various approval voting formulas.

• Drill-down workflows: The organization must be able to define "workflows


within a workflow" so that the organization can take full advantage of the
process definition work already completed when establishing a new workflow.

• Event Monitoring: The engine must monitor significant events such as when
a user does not approve or reject a document within a specified time frame, and
generate appropriate system actions based on such events.

• Status Reporting: The workflow system must provide detailed status


reporting on all jobs in process. Graphical displays should be provided to allow
the user to quickly comprehend the status of the workflow job.

• Audit logging: Audit logs and reports should be provided to allow managers
and regulators to review historical actions and provide proof of compliance
with mandated processes.

• Simulation: There should be a method of simulating the newly defined


workflow process, using event-based probabilities, prior to actually putting the
new process into operation.

In providing these functions the workflow engine is supporting both the need to
structure the workflow process as well as to enable authorized users to generate ad hoc
events within the workflow process. If developed properly, the workflow product will

Copyright © Genetic Computer School, 2002


Page100

support completely structured workflow processes, completely ad hoc functions, as well


as combinations of the two approaches.

8.8 Modular Product Design

As the workflow market matures, different vendors may specialize in different


areas of the classical workflow product. The Workflow Management Coalition, in
recognition of this trend, has begun to develop interface standards between the major
subcomponents of a workflow product. The Coalition has been well accepted by both
the vendor and user community of the workflow market.

Mission-critical workflow products must be designed to conform to these


standards. A modular product architecture with well-defined interfaces between
subcomponents can ensure that compliance with the ongoing development of standards
can be achieved at reasonable cost.

8.9 Environmental Independence

The underlying technology environment in which workflow products must operate


continues to change rapidly. Well designed workflow products must operate
independently of key technology infrastructures such as database, server, and network.
It is possible today to utilize standards such as SQL and ODBC to develop database-and
network independent workflow products. The workflow product should support
market-leading databases such as Oracle, Sybase, and Informix. For large Lotus Notes
installations, the workflow product should also be able to use the Lotus Notes
database as a back end data store as well as provide an API for integration with the
Notes client interface.

Even as the migration to client-server technologies continues, there are still


many mainframe-based line-of-business applications. Products which can extract data
from legacy systems, and utilize that information to define and drive the workflow
process, will add significant value to the, business process. Client/server based
workflow products which support relational databases, butt do not force organizations

Copyright © Genetic Computer School, 2002


Page101

to convert data from legacy systems into relational databases, will provide the best of
both worlds.

8.10 Workflow and the Emerging Information Infrastructure

Workflow systems operate not in isolation but rather within an information


systems infrastructure (see Figure 8.3 below).

Figure 8.3

As workgroup computing has increased in importance, a new class of systems


infrastructure called "groupware" has emerged. Groupware environments are a
collection of software which facilitate document delivery and collaboration between
multiple users. Groupware makes corporate information available to anyone on the
network enabled with the groupware product. Groupware capabilities may include
electronic mail, routing, foldering, and document replication.

Groupware environments are the backbone upon which users will expect
document management systems and workflow systems to operate. Major commercial
groupware products include Lot Notes, Novell GroupWise, and Microsoft Exchange.
Lotus Notes is the current groupware market leader with approximately 4.5 Million

Copyright © Genetic Computer School, 2002


Page102

installed seats. Groupware products include built-in electronic mail and database
replication where data is distributed based on the user group. Each user group
contains an object database which may include groupware forms as well as other
document types.

The rapidly growing popularity of the Internet and web (HT-FP) servers` threatens
the market position of the major commercial groupware products. Many analysts in fact
anticipate that the Internet as well as major corporate Intranets will become the
dominant groupware environment of the future. The Internet provides several
significant advantages over the currently installed technology infrastructure:

• Real-time information access to massive data servers providing large volumes


of information, including information external to the corporation.

• Low incremental cost per user relative to current user workstations. The
Internet relies only on a "thin client" to retrieve information from back-end
servers, and may in fact open up new opportunities for lower cost client
hardware.

Low cost multi-platform development environment.

Very simple, "point and click", standard user interface

The potential of an easier distribution strategy for new software and


software updates, where users can subscribe to software for periods of
time, rather than purchasing software which must be installed on each client.

Major corporations are establishing internal Intranets, in an effort to realize the


advantages of the Internet but "firewalling" their internal environments from the array
of possible intruders on the vast public Internet. According to a recent article in PC
Week, over 60% or Fortune 1000 corporations either currently have, or plan to install, a
corporate Intranet.

The Internet has no built-in database structure, as one would find in a groupware
product. However, major relational database vendors such as Oracle, Informix, and

Copyright © Genetic Computer School, 2002


Page103

Sybase have announced their own Web servers so that Web pages can be dynamically
generated "on-the-fly" to serve as a front-end to these relational databases. This
potential convergence between the Internet front-end and the relational database back-
end threatens the position of many commercial groupware products.

The emerging battle between traditional Client/Server configurations and the


Internet/Intranet is likely to last for a few years with no conclusive winners. Each
technology infrastructure has its own significant advantages. Traditional client/server
has the advantage of being proven, mature technology with a large installed base. The
Internet/Intranet provides a platform independent, information publishing environment
at low incremental cost per user.

While these infrastructures will provide a "groupware backbone", none of them


provide the key capabilities of a robust workflow product:

The ability to streamline a mission-critical business process through a


structured workflow definition, task scheduling, task automation, and
process metrics.

The ability to establish and manage a complex process definition


through an intuitive graphical interface.

The ability to seamlessly integrate with key technologies such as


relational databases, windowing environments, document management,
and imaging.

The ability to rapidly generate sophisticated and yet secure


applications from application frameworks.

Workflow products will support a number of integration strategies to groupware


technologies. For integration with groupware products, the groupware database should
be supported at the back end, while the Workflow client should be seamlessly
integrated with the groupware client via OLE automation. For the internet and

Copyright © Genetic Computer School, 2002


Page104

intranets, one integration strategy will be to allow a web browser to serve as the client
for the workflow engine.

Workflow products, which work across the major groupware environments will
become essential elements of the emerging information infrastructure required to
manage the global enterprise. Workflow products which work in both traditional
client-server environments as well as with the Internet/intranet will provide the best of
both worlds, taking full advantage of each world while offering the buyer the
opportunity to use a common workflow product across the corporation.

8.11 Teleprocessing systems

This is a classic method of supporting a multi-user database system, which uses


one computer and one CPU. Users operate dumb terminals that transmit transaction
messages and data to the centralized computer. Since there is little intelligence at the
users' end, all commands for formatting the screen must be generated by the CPU and
transmitted over the communication lines. This means the users interface is generally
character oriented and primitive. All inputs and outputs are communicated over a
distance to the centralized computer for processing.

Copyright © Genetic Computer School, 2002


Page105

Chapter 9: groupware

Introduction

A workgroup support system (WSS) is a system that is designed specifically to


improve the performance of teams by supporting the sharing and flow of information.
The foundation of any WSS is groupware - the popular term for the software
component that supports the collaborative efforts of a team. SSs primarily support the
information-processing tasks of conveying, creating, and communicating. WSSs also
work with all types of information - internal, external, objective, and subjective.
Groupware is among the newest IT-based tools that you'll find in an organization.

Below we've listed the three functions that groupware supports and the software
that facilitates that support.

1. Team Dynamics

♦ Group Scheduling Software

♦ Electronic Meeting Software

♦ Videoconferencing Software

♦ Whiteboard Software

2. Document Management

♦ Group Document Databases

3. Applications Development

♦ Workflow Automation Software

Comparing those design options across applications yields interesting new


perspectives on well-known applications. Also, in many cases, these systems can be
used together, and in fact, are intended to be used in conjunction. For example, group

Copyright © Genetic Computer School, 2002


Page106

calendars are used to schedule videoconferencing meetings, multi-player games use live
video and chat to communicate, and newsgroup discussions spawn more highly-
involved interactions in any of the other systems.

Consider how these systems can be integrated in other ways. We are still quite far
from developing the grand groupware system that encompasses every type of
communication, and we will probably never get there since the possibilities are
constantly evolving with changes in both our patterns of social interaction and the
technology we have available.

The features of each of the above technology tools will be considered: -

9.1 Group scheduling software and technologies

Group scheduling software

♦ is a component of groupware

♦ is a part of the electronic meeting support component of groupware

♦ provides facilities for maintaining the day-to-day electronic calendars of team


members and evaluating those calendars to schedule optimal meeting times

♦ may even let you reserve a certain room for a meeting and any equipment that
you may need

Office Tracker

♦ Can schedule rooms and other facilities

♦ Supports both personal and group scheduling

♦ Finds available times for groups

♦ Displays side-by-side group schedules

♦ Automatically posts notifications and reminders

Copyright © Genetic Computer School, 2002


Page107

ResSched

♦ Can schedule by fraction of the hour

♦ Can schedule equipment, space, people, and other resources

♦ Supports schedule reviewing by participant, meeting type, etc.

♦ Provides alternatives for booking conflicts

♦ Restricts who can schedule meetings

WallCHART

♦ Can schedule teachers and classrooms

♦ Used by United Nations for scheduling meetings

♦ Provides good reporting features (e.g., room utilization)

♦ Works on intranets and the Internet

♦ Large international presence

Livelink OnTime

♦ Personal calendars can be dispersed across multiple servers

♦ Works on Windows '95, '98, and NT and Mac clients

♦ Scalable to tens of thousands of users

♦ Supports browser access

COMSEC Scheduling Software

♦ Rooms, resources, activities, and office appointments

♦ Attach notes and service requests for schedules

Copyright © Genetic Computer School, 2002


Page108

♦ Use FM WebView to check schedules via the Internet

♦ Works under Windows 3.1/Nt/95/98 and OS2/Warp

Visto Briefcase

♦ Primarily a personal information manager with group-support functionality

♦ Enables relationship identification between workers on the road administration

♦ Supports group appointment setting and e-mailing

♦ Shows private appointments as blocked out times

ScheduleSoft

♦ Rules-based schedule generation system

♦ Runs across all operating system platforms

♦ Supports use of SQL-Server

♦ Runs on LAN or WAN

Meetingmaker

♦ Scalable to tens of thousands of users

♦ Works under DOS, Windows, and Mac

♦ Provides for laptop, palmtop, and remote access

♦ Provides instant access to free/busy time information for both users and
resources

♦ Runs in a standard browser

Copyright © Genetic Computer School, 2002


Page109

Electronic meeting software:

♦ is a component of groupware

♦ lets a team have a "virtual" meeting through IT

♦ supports both synchronous and asynchronous meetings

MeetingWorks for Windows

♦ Supports brainstorming, idea organization, ranking, and voting

♦ Generates high quality reports and graphs of meeting information

♦ Supports joint application development (JAD)

♦ Free Internet download and use for up to 8 participants

♦ Also available in MeetingWorkds Internet Edition and MeetingWorks Connect

Groupsystems

♦ Supports brainstorming, list building, information gathering, voting,


organizing, prioritizing, and consensus building

♦ Supports strategic planning activity-based costing, business process


reengineering, and knowledge management

♦ Includes utility for on-line surveying

♦ Supports graphical business process analysis and redesign

GroupSystems for Windows

♦ Supports synchronous and asynchronous meetings

♦ Supports anonymous input by participants

♦ Can be integrated with whiteboard software

Copyright © Genetic Computer School, 2002


Page110

♦ Facilitates ranking of issues by secret ballot

Microsoft NetMeeting

♦ Allows people to share audio, video, and computer files

♦ Can be integrated with whiteboard software

♦ Supports Chat features which allows geographically dispersed participants


to communicate messages quickly and easily

♦ Supports simultaneous application sharing

Meeting Builder

♦ For organizing and delivering special events

♦ Triggers further investigation of details so nothing is left out

♦ Can be used in conjunction with Hotel personnel package

Meeting Center 2000 and MyPlaceWare Professional

♦ Real time Web conferencing

♦ Powerful presentation features such as the ability to demonstrate a live


application

♦ Combines real-time visual content through a Web browser

Videpconferencing software

Videoconferencing software:

♦ is a component of groupware

♦ is part of the electronic meeting support component of groupware

Copyright © Genetic Computer School, 2002


Page111

♦ allows teams to have meetings and see each other when the team members
are geographically dispersed

NetMeeting

♦ Made my Microsoft

♦ Works on Windows 95, 98, and NT

♦ Works over the Internet or an organization's intranet Includes application


sharing whiteboard capabilities

EVX

♦ Made by Avistar Systems Corporation

♦ Allows full-window or mosaic view of team members

♦ Connects to NTSC (North American TV video standard) or PAL (TV


standard in many other parts of the world)

♦ Has enterprise messaging capabilities

Bright Light

♦ Made by Avalon Information Technologies

♦ Includes an information management system

♦ Allows multimedia-rich content delivery

Omega Product Line

♦ Made by VSI Enterprises

♦ Various products for all levels of videoconferencing

Copyright © Genetic Computer School, 2002


Page112

♦ Controls all aspects of conference via on-screen icons and mouse


movements

♦ Includes diagnostic software allowing problem diagnosis or


reconfiguration from remote sites

CineVideo/Direct

♦ Made by CINECOM Corp

♦ Works over phone lines using 28.8 KB or faster modems

♦ Supports one-on-one videoconferencing and live broadcasts

♦ Can be used with video capture devices

PictureTel Live 200

♦ Made by PictureTel

♦ Compatible with Windows 95 and Windows NT

♦ Allows the sharing of applications

♦ Allows transfer of files

CU-SeeMe

♦ Videconferencing for the Internet

♦ Can be used on any TCP/IP network

♦ Has a CU-SeeMe Contact List to allow contact with just an e-mail


address

Intel PC Camera w/ Video Phone Software

♦ Enjoy special occasions with family and friends

Copyright © Genetic Computer School, 2002


Page113

♦ With Pro pack video-in feature, share video from your camcorder or VCR
while you talk

♦ Hold a point-to-point videoconference with clients of colleagues

MyPlaceWare

♦ Present any type of information in a virtual conference

♦ Integrates all types of media into a presentation

♦ Select a time for the meeting and send e-mail addresses an invitation with
the URL

QVIX/CU30

♦ High quality, low bandwidth, real-time full video

♦ Videoconferencing for LINUX

Whiteboard software

Whiteboard software:

♦ is a component of groupware

♦ lets team members interactively edit and share documents

♦ can capture hand-written notes on an electronic whiteboard

SMART Board

♦ Captures notes written, on an electronic whiteboard

♦ Displays those notes on larger screens and individual computer screens

♦ Supports geographically dispersed video and data conferencing

♦ Even recognizes pen color and saves in that color

Copyright © Genetic Computer School, 2002


Page114

InPerson Whiteboard

♦ Allows you to view, mark up, scale, and manipulate 2D and 3D models

♦ Allows you to share and manipulate information from Windows-based


applications

♦ Can capture images from video streams

♦ Can import files and images

GroupSystems Whiteboard

♦ Can display images for review and annotation by participants

♦ Works with text as well

♦ Facilitator can control who has update or view only privileges

♦ Saves whiteboard contents for later use

TeamBoard

♦ Supports mouse point & click concept with your finger

♦ Captures notes and drawings electronically

♦ Can communicate via LAN or the Internet

♦ Can also distribute notes and drawings via e-mail

Soft Board

♦ Supports remote collaboration

♦ Supports simultaneous access to spreadsheets, the Internet, and presentation


tools

♦ Electronic pen acts as on-board mouse

Copyright © Genetic Computer School, 2002


Page115

♦ Available in plasma display technologies

Group Logic Imagexpo

♦ Interoperable with Windows and Mac

♦ Adds color slides and documents to phone calls

♦ Lets you create softproofs from all common graphic arts formats

♦ Annotation can be made in different, password-protected layers with unique


names

TearnWave

♦ Use whiteboard for meetings, note taking, design, discussion, and more

♦ Share files, URL links, images, etc.

♦ Communication through text and audio-video based synchronous


communication

MicroTouch Electronic Whiteboard Ibid 2.1

♦ Simplifies whiteboard teleconferences by detecting use of NetMeeting

♦ E-mail, print, drag-and-drop, fax, or post meeting notes to a Web site

Work Flow Automation Software?

Workflow automatic software

♦ is a component of groupware

♦ supports application development

♦ is designed to automate the flow of business documents in a specific work


process or procedure

Copyright © Genetic Computer School, 2002


Page116

Work Expeditor

♦ Designed for use with Microsoft Exchange

♦ Supports SQL server for data access and control

♦ Uses Microsoft Outlook as the CUI

♦ Supports security down to the document level

♦ Provides an O-O business repository

MQSeries Workflow

♦ Provides business process reengineering tools

♦ Helps automate processes for the cyber-corp

♦ Uses process diagrams to implement work flows

♦ Allows you to assign staff to data and processes

♦ Works with Lotus Notes and Web browsers

Staffware

♦ Supports a Java-based Web client

♦ Allows and organization to map, control, and improve business processes

♦ Supports flowcharting of business processes

♦ Includes Executive Information Service, allowing managers to track a


graphical view of a business process

PowerFlow

♦ Enterprise-wide workflow solution

Copyright © Genetic Computer School, 2002


Page117

♦ Allows organizations to graphically design and control processes

♦ Supports SQL database server

♦ Integrates with a variety of third party tools

METEOR Enterprise Application Suite

♦ Supports rapid application and deployment of process management


applications

♦ Works simultaneously on different servers

♦ Supports Web, CORBA, and Java-based development

♦ Integrates with current legacy systems and information

Ultimus Workflow

♦ Includes graphical workflow design tools

♦ Integrates easily with relational databases

♦ For both Internet and intranet applications

♦ Includes automated form designing tools

Dolphin

♦ Doesn't require programming skills to model work flows

♦ Works with NT and Exchange

♦ Processes can be modeled in list or diagram form

♦ For small, medium, and large departments

Copyright © Genetic Computer School, 2002


Page118

Cabinet NG

♦ Written in C and compatible with most other languages

♦ Interfaces with most known databases

♦ Works with word processing, spreadsheet, CAD, etc.

♦ Can produce reports and write them to laser disc

Action Workflow System

♦ Workflow software based on Action's workflow methodology

♦ Available for Lotus Notes and SQL databases

♦ Metro Application Center brings the workflow technology to the


Internet/intranet

♦ DocRoute provides workflow and automated routing to the leading


document management systems

9.2 GroupWare suites

GroupWare suites are software packages designed to provide a full range of


functionality for the complete workgroup environment. These packages often
include facilities for group scheduling, group communication, group meetings,
whiteboard work, group document management, and applications development.

Lotus Note/Domino R5

♦ Delivers messaging and collaborative solutions

♦ Supports Web applications and integration

♦ Integrates with DB2, Oracle, Informix, and Sybase databases

♦ Integrates with SAP's R/3 enterprise software

Copyright © Genetic Computer School, 2002


Page119

Microsoft Exchange

♦ For messaging and collaborative solutions

♦ Unlimited storage capacity, hosting thousands of users

♦ Supports a variety of security controls and encryption methods

♦ Incorporates built-in Internet standards

♦ Works with Lotus Notes and Domino

GroupWase

♦ Includes calendaring, workflow, and much more

♦ Includes images support (e.g., scanning and archiving)

♦ Supports publishing documents on the Web

♦ Y2K compliant

LinkWorks

♦ Document-based business oriented

♦ Supports imaging software such as PowerScan and StorageWorks

♦ Supports O-O architecture

♦ Runs under Windows, Mac, and Internet browsers

♦ Scalable for enterprise-wide deployment

TeamWARE Office

♦ Supports enterprise-wide discussion forums

♦ Provides team calendaring functions

Copyright © Genetic Computer School, 2002


Page120

♦ Accesses you mobile phone when an important e-mail arrives

♦ Supports multi-level access control

Web-4M

♦ Transforms traditional chat rooms into multimedia forums

♦ Collaboration is integrated with mail and Web documents and a


distributed file system

♦ Supports platform independence

♦ Supports full-duplex phoning

Netscape - SuiteSpot

♦ Manages large-scale, business-critical intranets and extranets

♦ Web-based Express clients for mail and calendar

♦ Deploys extranets globally in many languages and character sets

♦ Integrated software set that forms the basis of the networked enterprise

9.3 Asynchronous GroupWare

Email is by far the most common groupware application (besides of course,


the traditional telephone). While the basic technology is designed to pass simple
messages between 2 people, even relatively basic email systems today typically
include interesting features for forwarding messages, filing messages, creating
mailing groups, and attaching files with a message. Other features that have been
explored include: automatic sorting and processing of messages, automatic routing,
and structured communication (messages requiring certain information).

Newsgroups and mailing lists are similar in spirit to email systems except
that they are intended for messages among large groups of people instead of' 1-to-1

Copyright © Genetic Computer School, 2002


Page121

communication. In practice the main difference between newsgroups and mailing


lists is that newsgroups only show messages to a user when they are explicitly
requested (an "on-demand" service), while mailing lists deliver messages as they
become available (an "interrupt-driven" interface).

Workflow systems allow documents to be routed through organizations


through a relatively-fixed process. A simple example of a workflow application is
an expense report in an organization: an employee enters an expense report and
submits it, a copy is archived then routed to the employee's manager for approval,
the manager receives the document, electronically approves it and sends it on and
the expense is registered to the group's account and forwarded to the accounting
department for payment. Workflow systems may provide features such as routing,
development of forms, and support for differing roles and privileges.

Hypertext is a system for linking text documents to each other, with the Web
being an obvious example. Whenever multiple people author and link documents,
the system becomes group work, constantly evolving and responding to others'
work. Some hypertext systems include capabilities for seeing who else has visited
a certain page or link, or at least seeing how often a link has been followed, thus
giving users a basic awareness of what other people are doing in the system -- page
counters on the Web are a crude approximation of this function. Another common
multi-user feature in hypertext (that is not found on the Web) is allowing any user
to create links from any page, so that others can be informed when there are
relevant links that the original author was unaware of.

Group calendars allow scheduling, project management, and coordination


among many people, and may provide support for scheduling equipment as well.
Typical features detect when schedules conflict or find meeting times that will
work for everyone. Group calendars also help to locate people. Typical concerns
are privacy (users may feel that certain activities are not public matters),
completeness and accuracy (users may feel that the time it takes to enter schedule
information is not justified by the benefits of the calendar).

Copyright © Genetic Computer School, 2002


Page122

Collaborative writing systems may provide both realtime support and non-
realtime support. Word processors may provide asynchronous support by showing
authorship and by allowing users to track changes and make annotations to
documents. Authors collaborating on a document may also be given tools to help
plan and coordinate the authoring process, such as methods for locking parts of the
document or linking separately-authored documents. Synchronous support allows
authors to see each other's changes as they make them, and usually needs to
provide an additional communication channel to the authors as they work (via
videophones or chat).

9.4 Synchronous or Realtime Groupware

Shared whiteboards allow two or more people to view and draw on a shared
drawing surface even from different locations. This can be used, for instance,
during a phone call, where each person can jot down notes (e.g. a name, phone
number, or map) or to work collaboratively on a visual problem. Most shared
whiteboards are designed for informal conversation, but they may also serve
structured communications or more sophisticated drawing tasks, such as
collaborative graphic design, publishing, or engineering applications. Shared
whiteboards can indicate where each person is drawing or pointing by showing
telepointers, which are color-coded or labeled to identify each person.

Video communications systems allow two-way or multi-way calling with


live video, essentially a telephone system with an additional visual component.
Cost and compatibility issues limited early use of video systems to scheduled
videoconference meeting rooms. Video is advantageous when visual information is
being discussed, but may not provide substantial benefit in most cases where
conventional audio telephones are adequate. In addition to supporting
conversations, video may also be used in less direct collaborative situations, such
as by providing a view of activities at a remote location.

Chat systems permit many people to write messages in realtime in a public


space. As each person submits a message, it appears at the bottom of a scrolling

Copyright © Genetic Computer School, 2002


Page123

screen. Chat groups are usually formed by having listing chat rooms by name,
location, number of people, topic of discussion, etc.

Many systems allow for rooms with controlled access or with moderators to
lead the discussions, but most of the topics of interest to researchers involve issues
related to unmoderated realtime communication including: anonymity, following
the stream of conversation, scalability with number of users, and abusive users.

While chat-like systems are possible using non-text media, the text version of
chat has the rather interesting aspect of having a direct transcript of the
conversation, which not only has long-term value, but allows for backward
reference during conversation making it easier for people to drop into a
conversation and still pick up on the ongoing discussion.

Decision support systems are designed to facilitate groups in decision-


making. They provide tools for brainstorming, critiquing ideas, putting weights and
probabilities on events and alternatives, and voting. Such systems enable
presumably more rational and even-handed decisions. Primarily designed to
facilitate meetings, they encourage equal participation by, for instance, providing
anonymity or enforcing turn-taking.

Multi-player games have always been reasonably common in arcades, but are
becoming quite common on the internet. Many of the earliest electronic arcade
games were multi-user, for example, Pong, Space Wars, and car racing games.
Games are the prototypical example of multi-user situations "non-cooperative",
though even competitive games require players to cooperate in following the rules
of the game. Games can be enhanced by other communication media, such as chat
or video systems.

9.5 Groupware: Design Issues

As with all user interface design, the method used for designing a groupware
system is more significant than specific design suggestions. This introduction thus

Copyright © Genetic Computer School, 2002


Page124

begins with the groupware design process. The remaining sections address some of
the most common` issues that face groupware designers.

The Groupware Design Process

It's best to start by gaining a solid understanding of your prospective users,


what their goals are, and how they go about their work. For broadly-targeted
groupware applications, such as videophones or email, understanding users can
boil down to understanding how human beings communicate in the first place. A
design is also best informed by conducting user studies on system prototypes. In
these cases user testing is often significantly more difficult than with single-user
systems for the following reasons:

♦ Organizing and scheduling for groups is more difficult than for


individuals.

♦ Group interaction style is hard to select for beforehand, whereas individual


characteristics are often possible to determine before a study is conducted.

♦ Pre-established groups vary in interaction style, and the length of time


they've been a group affects their communication patterns.

♦ New groups change quickly during the group formation process.

♦ Groups are dynamic; roles change.

♦ Many studies need to be long-term, especially when studying asynchronous


groupware.

♦ Modifying prototypes can be technically difficult because of the added


complexity of groupware over single-user software.

♦ In software for large organizations, testing new prototypes can be difficult or


impossible because of the disruption caused by introducing new versions into
an organization.

Copyright © Genetic Computer School, 2002


Page125

When designing groupware, it is often best to begin with field studies. The goal is
to understand a particular type of group or organization that will be using the
groupware system. A number of different studies can be conducted: interviews,
surveys, analysis of artifacts used in the work process, examination of processes and
workflows, etc. In all cases, the object is to identify the users' tasks and goals,
understand how the group communicates and determine the power structures and roles.

One key challenge is to appear non-threatening and objective to the users in order
to obtain accurate information and to insure that they will accept any design that results.
Another challenge is translating the findings from one organization to others -- this is
especially a concern when the groupware is intended for organizations which are truly
unique or too large to effectively study.

Adoption and Acceptance

Many groupware systems simply cannot be successful unless a critical mass of


users chooses to use the system. Having a videophone is useless if you're the only one
who has it. Two of the most common reasons for failing to achieve critical mass are
lack of interoperability and the lack of appropriate individual benefit.

Interoperability

In the early 90s, AT&T and MCI both introduced videophones commercially, but
their two systems couldn't communicate with each other. This lack of
interoperability/compatibility meant that anyone who wanted to buy a videophone had
to make sure that everyone they wanted to talk to would buy the same system.
Compatibility issues lead to general wariness among customers, who want to wait until
a clear standard has emerged.

Perceived Benefit

Even when everyone in the group may benefit, if the choice is made by
individuals, the system may hot succeed. An example is with office calendar systems: if
everyone enters all of their appointments, then everyone has the benefit of being able to

Copyright © Genetic Computer School, 2002


Page126

safely schedule around other people's appointments. However, if it's not easy to enter
your appointments, then it may be perceived by users as more beneficial to leave their
own appointments off, while viewing other people's appointments.

This disparity of individual and group benefit is discussed in game theory as the
prisoner's dilemma or the commons problem. To solve this problem, some groups can
apply social pressure to enforce groupware use (as in having the boss insist that it's
used), but otherwise it's a problem for the groupware designer who must find a way to
make sure the application is perceived as useful for individuals even outside the context
of full group adoption.

Avoiding Abuse

Most people are familiar with the problem of spamming with email. Some other
common violations of social protocol include taking inappropriate advantage of
anonymity, sabotaging group work, or violating privacy.

The Commons Problem

If a village has a "commons" area for grazing cattle then this area can be a strong
benefit to the community as long as everyone uses it with restraint. However,
individuals have the incentive to graze as many cattle as possible on the commons as
opposed to their own private property. If too many people send too many cattle to the
commons, the area will be destroyed, and the whole village is worse off as a result.
There are a couple of straightforward solutions to the Commons Problem: an
appropriate fee can be charged for each head of cattle or a limit can be imposed on the
number of cattle any individual may bring. These solutions are an appropriate starting
point for solving problems of abuse in GroupWare.

CSCW (Computer-Supported Cooperative Work) refers to the field of study which


examines the design, adoption, and use of GroupWare. Despite the name, this field of
study is not restricted to issues of "cooperation" or "work" but also examines
competition, socialization, and play. The field typically attracts those interested in
software design and social and organizational behavior, including business people,

Copyright © Genetic Computer School, 2002


Page127

computer scientists, organizational psychologists, communications researchers, and


anthropologists, among other specialties.

9.6 How is GroupWare Design Different from Traditional User Interface Design?

GroupWare design involves understanding groups and how people behave in


groups. It also involves having a good understanding of networking technology and
how aspects of that technology (for instance, delays in synchronizing views) affect a
user's experience. All the issues related to traditional user interface design remain
relevant, since the technology still involves people.

However, many aspects of groups require special consideration. For instance, not
only do million-person groups behave differently from 5-person groups, but the
performance parameters of the technologies to support different groups vary. Ease-of-
use must be better for GroupWare than for single-user systems because the pace of use
of an application is often driven by the pace of a conversation. System responsiveness
and reliability become more significant issues. Designers must have an understanding
of the degree of homogeneity of users, of the possible roles people play in cooperative
work and of who key decision-makers are and what influences them.

Why Is GroupWare design worth paying attention to in the first place?

GroupWare offers significant advantages over single-user systems. These are


some of the most common reasons people want to use GroupWare:

♦ to facilitate communication: make it faster, clearer, more persuasive

♦ to enable communication where it wouldn't otherwise be possible

♦ to enable telecommuting

♦ to cut down on travel costs

♦ to bring together multiple perspectives and expertise

Copyright © Genetic Computer School, 2002


Page128

♦ to form groups with common interests where it wouldn't be possible to gather a


sufficient number of people face-to-face

♦ to save time and cost in coordinating group work

♦ to facilitate group problem-solving

♦ to enable new modes of communication, such as anonymous interchanges or


structured interactions

In addition to the benefits of GroupWare, another good reason to study usability


and design issues in GroupWare is to avoid a failed design. GroupWare is significantly
more difficult to get right than traditional software. Typically, a GroupWare system
can't succeed unless most or all of the target group is willing to adopt the system. In
contrast, a single-user system can be successful even if only a fraction of the target
market adopts it.

Copyright © Genetic Computer School, 2002


Page129

CHAPTER 10: PERSONAL INFORMATION SYSTEMS

10.1 Data Representation

Data and Information

Raw data, as it is initially collected, is of little value or use. The human brain
organizes audio/visual (and other) inputs and the mind interprets these inputs and
assigns meaning to them. Thus a pattern of vibrations in the air is interpreted as
conversation or as music, and a pattern of retinal impulses-is a rose or perhaps a lover.
One sees and hears, but there is much more to this than just a few organized sensory
inputs, for intelligent organization is required to give meaning to the stimuli. Likewise,
economic and scientific data consists only of raw numbers (symbols) until it has been
organized and interpreted. Once such higher levels of meaning have been attached to
data, it is termed information. This intelligent act of attaching meaning to raw data is
clearly an abstraction process. Indeed, assigning meaning may to some extent be
thought of as a synonym for the whole abstraction forming activity.

At a somewhat more concrete level, it is often possible to automate certain


repetitious calculating tasks that are part of the process of placing meaning on data
collections. These tasks are ideally suited to modern computing machinery and are
driven by sets of instructions that achieve the mechanical aspects of the data
organization. One can go farther than this and say that certain standard meanings are
collected and tabulated, then automatically assigned to the items in the data collection
by the computer program. The latter is then just a fast and reliable extension of human
intelligence bent to the abstraction task

Encoding (Representing) Data

However, there are practical issues to solve at a lower level than describing what
data processing is. These centre on how data is communicated. Whenever one writes a
symbol like "4" or “four," a potential for communication exists, based on the fact that

Copyright © Genetic Computer School, 2002


Page130

these symbols encode a certain idea. By convention, everyone encodes the same idea
with these symbols, so communication is possible.

Computation devices must also encode data in some consistent way. There are two
categories of such codes:

1. External codes. These are usually human-readable characters that are input into
a computer or output from it using a keyboard, screen, printer, or other device. The
most common form for this data is a character such as "4," "a," "%," and so on.

2. Internal codes. Because computer storage is based on electronic circuits that


can be thought of as ON/OFF switches, internal storage is in a different format than that
used for human interface with the machine. In this form, it is not directly accessible by
a human user. The person using a computing machine does not usually need to know
what kind of internal representation is employed for data, because there are input/output
programs that translate the data between internal and external formats.

There are some issues relating to data representation that do make a difference to
programmers, however.

Bits and Bytes

The modern digital computer can only understand binary representations of data.
Binary means "two" -- that is to say, only two "states" can be represented. Over the half
century since the computer has been in existance, many ways have been used to
electrically or mecahically record binary representations. In the beginning, punched
tape and cards were used-- hole or no-hole indicated the binary state. With today's
magnetic media, the polarity (positive or negative) is used to indicate these two binary
states. Many other methods have been used in the past and presently. For humans (not
machines), we use the convenient notation of zero [0] or one [1] to write these two
binary states. We call these elemental states bits--the "zero bit" and the "one bit." The
term "bit" is derived from "BInary digiT".

Copyright © Genetic Computer School, 2002


Page131

Quite obviously, if zero and one were all we could represent in the computer, we
wouldn't have much data. Only rarely is business data represented by a single bit. Bits,
however, are similar to atoms. They serve to create more complicated representations--
to carry the analogy further, the "molecules" of data that humans can use to encode
business information. Bits are assembled into eight-bit patterns called bytes. The basic
storage unit in the computer's main memory and secondary storage is the byte. Bits are
not stored separately.

This technique permits up to 256 possible combinations, which can be assigned


any meanings humans desire. Most frequently, they are used to represent binary
numbers or alphanumeric codes needed to represent business data. However, they can
also be used to represent colors, positions on computer screens, or computer
instructions that the CPU can interpret and act upon, such as moving data to a disk or
printing a line.

10.2 Database Processing

One of the major functions of an organization is the processing of data to generate


the information needed to control and manage the organization. Historically, this
processing was done manually, but the advent of the computer allowed large companies
and government agencies, that could afford a computer and a supporting data
processing department, to computerize the processing of data. If a new report was
needed, a request would be submitted to the data processing department and a
programmer would be assigned to write the program to generate the report. The advent
of the microcomputer was the beginning of radical changes that are still evolving today.
New software was developed in the area of data base management that allowed the user
to manipulate data directly without going through the data processing department. This
database management software gave the average computer user a great deal of power in
working with a database. Data can now be handled in a way that was formally reserved
for the data processing department: the end user can create files, maintain files, select
information from the files, sort the files and print reports based on the information in

Copyright © Genetic Computer School, 2002


Page132

the files. More sophisticated application are still programmed frequently using the
programming language available with the data base management system.

A database is a single collection of data. It can hold the information about many
different entities related to the company (an entity being simply a person, place or thing).
Databases not only contain the information about these entities (known as attributes),
they also contain information about relationships.

A database is all the raw data needed to create and provide information that will
satisfy the varying needs of users in the organization. In some cases the database may
be a mixture of data held on computer and manual filing systems but for the purposes of
future sessions we will assume it is all held on computer. Inside the database are all of
the items of data that belong to the organisation but this data may have been stored on
different systems in different locations (a distributed database). In order to be able to
locate where the information is stored, the database requires a management system
which effectively provides a map of where data is stored and a programme to locate and
retrieve the data specified. In order to overcome the problem of data stored on different
systems, written in different 'computer languages', programming staff have had to
develop bespoke management systems specifically for the purpose. The Data Base
Management System is, in effect, the intermediary between raw information and those
who need the information, giving users the freedom to use the data in their own specific
way. This may not be the ideal solution - disparate databases often require a high level
of maintenance to maintain their effectiveness and there is a substantial cost associated
with this. A preferred solution may be to develop a corporate database with a
corporate Information Systems structure and framework.

The database is the primary means of integration and dissemination of data (refer
to figure 10.1)

Copyright © Genetic Computer School, 2002


Page133

The data that is stored or updated by one system is then available to the others. It
can also be used to produce specific management information.

Figure 10.2

The higher up the organisation, the less structured information tends to become
(refer to figure 10.2)

In a previous session, a figure similar to the one above was used to explain the
role of different levels of management within an organization. If we consider the
types of decision with, which these levels will be involved then there will be a
tendency to move from structured decisions (programmed) at the operational level, to
unstructured (non-programmed) decisions at the strategic level. Underpinning the
pyramid is a representation of the operational and functional systems of the business,
i.e. accounting, stock control, debtors/creditors, providing information for the operational
level of management. In the past, these systems have also provided the management
information for the tactical and strategic levels of management. In effect, the next
layer up has received information contextualized by the level below. This has three
major disadvantages:

i. The information coming from these systems is looking backwards


(historical), whereas the higher levels of the management structure are focused on
increasingly longer forward projections.

ii. Information being passed upwards is a reflection of what that level thinks
the next level up requires (but is often not the case).

iii. By the time information has passed up the structure, been acted upon and
decisions passed back down again, the situation will have altered in any case!

Copyright © Genetic Computer School, 2002


Page134

In summary, a single management information system, which attempts to


satisfy the needs of all levels of management, often satisfies none.

10.3 Personal Strategic Planning

What are some of the benefits of Personal Strategic Planning?

First of all, strategic planning helps provide a personal vision for the future. It
is so easy to get wrapped up in the present that we lose sight of where the future
might be leading us. If we intend to grow in any significant aspect of our lives, we
need to periodically focus on what the future will look like. Depending on where
you are in your career and your life, this future vision may be as little as two or
three years out or could be as much as twenty or thirty years in the future. The
ironic thing about focusing on a period in the future is that what you project for
that time is probably not what's going to happen. Specific circumstances,
opportunities, threats, and personal preferences may lead you in a distinctly
different direction from what you established in your initial planning effort. That
does not invalidate the planning process. By focusing on the future, we are able to
determine when it is appropriate to change a course of direction.

Career direction is a strong concern for most professionals, particularly


during the early stages of their careers. It is interesting to note that only a small
fraction of professionals end up in the career 'for which they initially prepared. Our
interests change; different opportunities present themselves, family obligations
impact the direction in which we should be moving. By looking forward, we have a
better chance of identifying some of the road blocks that may interfere with where
we want to go and identify some of the options that may become available to us.
Sometimes, by playing the "what if' game, it will open up new vistas that might not
otherwise become apparent to us. While this will not necessarily prevent us from
stumbling along the way, there is a greater likelihood that we will be able to
respond more effectively to things that can significantly impact where we are
going.

Copyright © Genetic Computer School, 2002


Page135

Creating and taking advantage of opportunities is generally much more


productive if we keep our peripheral vision open. These "opportunities" are not
limited to our profession. They can include such things as financial investments,
real and personal property, family and friends, or anything that has a significant
meaning in our lives. While there is always a risk in looking at opportunities, that
risk tends to be reduced when we have anticipated it and planned how to deal with it.

As a counterpart to the risk involved in opportunity, strategic planning can


also significantly enhance our personal and business security and safety. This
includes, of course, financial security, health and well-being, family concerns, and
any factors that might cause significant future trauma.

The subject of retirement can arouse a variety of different emotions,


depending on where we are in our chronological and professional lives. For many
professionals, total retirement may never be a reality. Many of us feel that, as long
as we have our physical and mental faculties, we will continue to practice our
professions indefinitely. The base and focus of our efforts may change, but we plan
to continue practicing our professions to some degree as long as we are able. On
the other hand, there are many professionals who look forward to retirement as an
opportunity for a significant change in lifestyle that will create new experiences
that their careers may have interfered with in the past. Regardless of your own
perspective on retirement, when you come to the point where it is appropriate for
you to back off from what you have been doing, it will be much more satisfying
and fulfilling if it comes as a result of planned effort rather than the closing of a door.

The building and maintaining of a balanced life represents another significant


benefit of personal strategic planning. Oftentimes, practicing professionals develop
"tunnel vision" which keeps them narrowly focused on their business and careers.
During certain periods of our lives, this may be justified. However, life is much
more than getting ahead professionally or in business. We need to achieve a
balance that includes family and friends, health and wellness, personal fulfillment
that may not directly relate to our profession, financial planning, spiritual

Copyright © Genetic Computer School, 2002


Page136

development, and service to others. Just as a chair will not function properly if one
of its legs is longer or shorter than the other, neither will our lives function
effectively without some sort of holistic balance. Giving significant attention to
another important aspect of our lives does not necessarily detract from our
professional focus. In fact, it is possible to achieve a true synergy wherein the
"whole" person can be even more productive professionally as well as personally.

A final significant benefit is the opportunity to involve others in making our


futures come alive. These could include our life partners, parents and siblings,
children, professional colleagues, employers, and friends who have either a vested
interest in our success or a genuine concern about our future well-being. Since
there are very few things in our lives that we can accomplish without the help of
others, it is much more meaningful to get them involved as early as possible in the
planning process. Not only are they able to offer significant contributions that may
help make our plans more effective, their active support in the pursuit of these
efforts can go a long way toward assuring their achievement.

Why do Professionals tend to resist Planning?

Thinking versus action frequently provides an interesting dilemma for many


professionals. We tend to be action-oriented. When we are doing something, we
have a feeling of being productive. When we take the time to think, we sometimes
feel guilty of being unproductive. While there is no question but that whatever we
do should lead to some sort of productive action, in reality very little such action
comes without some kind of realistic thinking. Therefore, it is not "either-or";
rather, it is "both-and." In fact, without investing the time and effort required to
think about where we are going, there is a very strong likelihood that we will never
get there. And, of course, we have to decide where "there" is. It is difficult to do
that without putting significant effort into thinking.

Time required frequently is a barrier to effective planning. How often have


we said "I just haven't got the time to do that right now". While that may be a
legitimate reaction occasionally, in many cases it turns out to be just an excuse. It's

Copyright © Genetic Computer School, 2002


Page137

interesting to note that we frequently don't have time to plan the job appropriately
ahead of time. However, when the job does not get done right, somehow or other
we always find the time to do it over again. Time is one of those exhaustible
commodities with which we have to deal as professionals. Consequently, the
planning process, while it does take time, could lead us to much more productive
use of what limited time we do have available.

Changing circumstances frequently is given as a reason for not putting


significant effort into planning. "How can we plan when things are changing so
fast!" is a frequently-heard expression. As we identified earlier, changing or
modifying plans is not only inevitable but actually desirable in many situations.
The key, of course, is to recognize when changes are taking place. There is nothing
wrong with changing direction, provided we know we are changing and there is a
valid reason for doing so. The purpose of a plan is to provide us with a foundation
from which we can make appropriate digressions which may be more productive at
the time.

When things are going well is a time when it is easy to fall into the trap of
thinking that they are going to continue to go well "forever." When we are on a
highly successful path is probably the time when strategic planning is most critical.
Despite the euphoria we may feel at such times, it is absolutely certain that it will
not go on "forever." Something will happen that will be other than what we had
anticipated, thus moving us, whether we like it or not, in a different direction.
Being better prepared for dealing with those situations that may run counter to our
desired direction is one of the single most important reasons for doing strategic
planning in the first place. While we may not be able to anticipate everything that
will have an impact on where we are going, we stand a much better chance of
dealing effectively with that if we have looked ahead and anticipated some of the
things we might be facing.

When things are not going well is probably a more legitimate reason for
postponing strategic planning efforts. When your house is on fire is not the time to

Copyright © Genetic Computer School, 2002


Page138

think about installing a sprinkler system. Really, when we are faced with survival,
that has to take precedence over where we are going in the long run. For, as one
wag said, "In the long run we are all dead." The problem that frequently faces us,
however, is that we tend to think we are always in a crisis situation. Unfortunately,
this frequently comes as a result of the way we function personally. If you find
yourself in a situation where you are moving from one crisis to another on a
continuing basis, perhaps that is when you need to take some time off, sit back and
really think about where it is that you want to go. We frequently find that these
crises come as a result of a lack of effective planning in the first place. At some
point, we may need to break that pattern in a way that is going to be more
productive for us.

Finally, not recognizing the difference between strategic or long-term


planning and operational planning may inhibit our efforts. Operational planning,
which is commonly what we have to apply in dealing with crisis situations or in
achieving our short term results, tends to be quite specific and very much action-
oriented. Strategic planning, on the other hand, is much more conceptual and
visionary. It is far more directional than specific. If you plan to commit yourself to
some form of strategic planning in your own professional life, it needs to be
addressed at a time when that is the principal focus. You may find it helpful to
have someone else assist you in the process who can keep you focused on the
strategic or long-term concerns. Think of strategic planning as helping you
determine where it is you are going and operational planning as how you will get there.

10.4 Survey of Personal Information Systems

Text Processing

A.M. Turing said that if a machine could impersonate a human being, then the
machine was thinking. A computer may be said to have artificial intelligence when
it exhibits intelligence ordinarily associated with human behaviour (reasoning,
learning, use of language, and so on). In the field of AI, there are opposing view
points as to whether computers should be programmed to imitate the way the

Copyright © Genetic Computer School, 2002


Page139

human mind works, or whether computers should simply simulate human


behaviour in any way practical, whether or not that is the way people think.

Computer scientists generally agree that an AI programme is designed to


produce knowledge and that such processing includes inferencing (the ability to
derive additional knowledge from the original knowledge base). Inferencing
capabilities are required, for instance, to fulfil the task of determining phonemes
from variable sound or allophones.

If "intelligent computers" are ever developed, if artificial intelligence ever


approaches human intelligence, it is clear that linguistics will play a major role in
these developments, because human linguistic ability is the single most prolific
manifestation of intelligence.

Computers are masters of the statistical analysis of language. They can be


programmed to reveal such properties of language as the distribution of sounds,
allowable word orders, permitted combinations of morphemes, relative frequencies
of words and morphemes, etc.

Existing texts, such as the Bible or Shakespeare's works, can be analyzed, and
so can a collection of utterances gathered from spoken or written sources. Such a
collection is called a corpus. At Brown University (U.S.A.), a corpus of written
American English (consisting of newspaper passages, articles from magazines, and
other literary material) was compiled. A corpus of spoken American English,
similar in size to the Brown corpus, was also collected. The Brown corpus and the
spoken American English corpus were analyzed and compared. This comparison
provided a contrast between written and spoken American English. The pronoun
"I" occurs ten times more frequently in spoken language. Profane and taboo words
are, as expected, more frequent in spoken language. Prepositions occur more
frequently in written than in spoken American English, so different syntactic
structures are used in written and spoken English.

Copyright © Genetic Computer School, 2002


Page140

Computers can also be used to produce a concordance of a literary text,


which gives the frequency of every word in a text and the line and page number of
each occurrence. Nowadays, concordances are used to help ascribe the authorship
of a text to a certain author. If the concordance of a text -- of which the author is
unknown -- is compared to the concordances of known works by two authors, it is
possible to say which of the two authors wrote the text.

A concordance of "sounds" by computer may reveal patterns in poetry that


would be nearly impossible for humans to detect. Poetic and prosaic features such
as assonance, alliteration, meter, and rhythm have always been studied by literary
scholars. These days, computers can do this mind-numbing work.

Electronic Spreadsheets

Information Systems oral history and some published newspaper interviews


report that in 1978 Dan Bricklin, an M.I.T. alum and then student at Harvard
Business School, invented the electronic spreadsheet. Bricklin supposedly was
preparing a spreadsheet for an HBS "case study" and had two alternatives: 1) do it
by hand; or 2) use a clumsy time-sharing mainframe program. Bricklin thought
there must be a better way. He wanted a program where people could visualize the
spreadsheet as they created it. His metaphor was "an electronic blackboard and
electronic chalk in a classroom."

By the summer of 1978, Bricklin had programmed the first working version of
his concept. The program would let users input a matrix of five columns and 20
rows. The first version was not very "user friendly" so Bricklin recruited Bob
Frankston to improve and expand the program. Frankston expanded the program
and "packed the code into a mere 20k of machine memory, making it both powerful
and practical enough to be run on a microcomputer".

During the summer of 1978, Daniel Flystra joined Bricklin and Frankston.
Flystra was also an M.I.T./HBS grad. Flystra was marketing oriented and
suggested that the product would be viable if it could run on an Apple computer.

Copyright © Genetic Computer School, 2002


Page141

The three formed Software Arts corporation in January 1979. In April 1979, the
company began marketing "VisiCalc", a compression of the phrase "visible
calculator".

VisiCalc became an almost instant success and provided many business


people with an incentive to purchase a personal computer. About I million copies
of the spreadsheet were sold during Visicalc's product lifetime.

What is a spreadsheet?

In the accounting world a spreadsheet was and is a large sheet of paper that lays
everything out for a businessperson. It spreads or shows all of the costs, income, taxes,
etc. on a single sheet of paper for a manager to look at when making a decision.

An electronic spreadsheet organizes information into columns and rows. The data
can then be "added up" by a formula to give a total or sum. The, spreadsheet
summarizes information from many sources in one place and presents the information
in a format to help a decision maker see the financial "big picture" for the company.

What came after VisiCalc?

The market for electronic spread sheets was growing rapidly in the early 1980s
and VisiCale was slow to respond to the introduction of the IBM PC that used an Intel
computer chip. Mitch Kapor developed Lotus and his spreadsheet program quickly
became the new industry standard. It not only made spreadsheet formula easier by using
the shorter, more intuitive Al referencing system (as opposed to Visicalc's R1C1
system) but also added graphics and set spreadsheets on the road to become major data
presentation packages, as well as complex calculation tools. Lotus was also the first
spreadsheet vendor to introduce naming cells, cell ranges and spreadsheet macros.
While at Visicorp, Kapor wrote Visiplot/Visitrend which he sold to Visicorp for $1 million.
Part of that money was used to start Lotus Development Corp. Ironically, Kapor offered
to sell Visicorp his Lotus 1-2-3 program. Supposedly VisiCorp executives declined the
offer because Lotus 1-2-3's functionality was "too limited". To date Lotus 1-2-3 is still
the all-time best selling application software in the world.

Copyright © Genetic Computer School, 2002


Page142

In 1985 Lotus acquired Software Arts and discontinued VisiCalc. A Lotus


spokeperson indicated at that time that "1-2-3 and Symphony are much better products
so Visicalc is no longer 'necessary."

The next milestone was Excel. Originally written for the 512K Apple Macintosh
in 1985 Excel was one of the first spreadsheets to use a graphical interface with true
pull down menus and a mouse point and click. The spreadsheet instantly became easier
to use than the "archaic interface" of PC-DOS products and many people bought Apple
Macintoshes simply so that they could use Excel. Excel never did come out in a DOS
version.

When Microsoft unveiled the original Windows in 1987, Excel was one of the first
products to be written for it, and even now many people still use Excel 2.1 which was
written to run under Windows version 2. When Windows finally took off with Version 3.0
in late 1989 Excel was its flagship product. It remained the only Windows spreadsheet
for nearly 3 years and has only received any real competition from other products since
summer 1992.

By the mid 1980s many companies had introduced spreadsheet products. Lotus
had acquired Software Arts and the rights to VisiCalc. Also, Microsoft had-joined the
fray with the Excel spreadsheet. By the mid-1990's IBM had acquired Lotus and
Microsoft Excel was the spreadsheet market leader.

Copyright © Genetic Computer School, 2002


Page143

Chapter 11: executive information system

11.1 Introduction

At the strategic level of the organization, a reliance upon operational, historical


information is like 'trying to drive a car forward by only looking in the rear view
mirror'. Of course, historical information can be useful in identifying patterns and
possible future trends but these will be affected by external factors, particularly over
long time spans. Executive Information Systems enable data and information to be
drawn out from any number of systems, including those external to the business, for
example, Stock Market Share information. Models using this information can be
constructed and 'what if?' questions posed to help evaluate future positions. For
example, a 'model' might be constructed of the potential profitability of a new product
over a five years life cycle. The question, "what if a similar product is brought onto
the market in year 3 at 5% below our selling price?", can be fed into the model to help
provide an answer to that particular scenario.

Executive Support System are graphics-oriented systems designed for senior


management athat provide generalized computing and telecommunication facilities
for monitoring and controlling a business. Executive Information Support System
software and communication package integrates virtually all of your information
sources into a cohesive network.

Functions such as manufacturing, purchasing, customer service, accounting,


which generate production schedules, inventory usage, income statements, balance
sheets, are connected to provide immediate access and dissemination of critical
information. Executive support systems are very important for information technology
to support the management of business.

Executive support systems (ESS) help managers make unstructured and semi-
structured decisions. They focus on the information needs of senior management and
combine data from both internal and external sources. A system is a generalized

Copyright © Genetic Computer School, 2002


Page144

computing and communications environment that can be focused on a changing array


of problems. ESS include tools for modeling and analysis.

ESS can and do change the workings of organizations. Executives are better able
to monitor activities below them, allowing them to push decision making further down
in the organization while expanding the executive's span of control.

ESS flexibility allows executives to shape the problems, using the system as an
extension of their own thinking. ESS offer executives the ability to analyze quickly
and to compare and highlight trends, freeing up executives -- and even more so their
staffs -- for more creative analysis and decision making.

An executive support system (ESS) is a decision support system especially made


for senior-level executives, including all the necessary hardware, software, data,
procedures, and people. Several commercial software packages are available for
specific modeling purposes.

An ESS must take into consideration:

♦ The overall vision or broad view of company goals

♦ Strategic, long-term planning and objectives

♦ Organizational structure

♦ Staffing and labor relations

♦ Crisis management

♦ Strategic control and monitoring of overall operations

Basic qualities of an EIS are:

♦ capability to look over the situation in your company “in half an hour"

♦ punctual taking steps in eventual deviation from planning activities

Copyright © Genetic Computer School, 2002


Page145

♦ reduced and merged data, interesting for an executive

♦ ability to communicate with employees by local Email

♦ it doesn't require any keyboard skills from user, executives can use just
mouse

♦ very adaptable in concrete user environment and satisfaction of his


demands

♦ graphical presentation of data

♦ many kind of printed reports and reviews due the different criteria

Executive Information Systems (EIS) have grown in interest and use over the
past 15 years. They are a response to inadequacies in Management Information
Systems (MIS) which, although capable of manipulating vast quantities of data, are
frequently difficult to use and not able to respond to managers' needs with any
degree of flexibility. An executive decision maker requires a precise understanding
of the current organisational situation and it has been observed (Dreyfus & Dreyfus
1986) that businessmen prefer concrete information, even gossip, speculation and
hearsay to the abstracted summary information contained in routine MIS reports.
EIS are relatively new tools which purport to provide executives with computer-
based information support for decision making. Not surprisingly, one of the
specified characteristics of an EIS is that it be ‘user-friendly and require minimal
or no training to use' (Watson et al, 1991). Due to the reluctance of executives to
adopt information technology for their own work, developers of EIS have been
particularly concerned with the user-friendliness of their products, with the result
that most EIS have attractive and ease to use interfaces. Despite this, or maybe
because of this, the majority of EIS projects are not successful. This concentration
on a generic interface between a user and the machine arises from work in HCI
(Human Computer Interaction) based on cognitive psychology which, although
adequate for lower level information systems, may account for a lack of EIS
success when applied to these more complex systems

Copyright © Genetic Computer School, 2002


Page146

This leads us to seek an optimal interface design for users who are making
executive decisions' and to question whether the current models of HCI are
adequate. Questions raised relate to the characteristics of this user population and
the way executives may be expected to approach and make use of such a system.
Do executives model things differently to others? Do they focus on different
information? Do they deal with information in a different way? Are executives a
homogeneous group in this regard? Have studies been conducted which examine
these questions? In this paper we look at the background to this problem and
suggest ways of discovering solutions to this most fundamental problem in EIS
interface design.

11.2 Components of an Executive Information System

Examining the components of an EIS is dependent upon the definition used.


Opinions differ on how much peripheral support, as opposed to amalgamated
information, should be provided. Taking the distinction made by Watson et al
(1991) between an EIS and an Executive Support System (ESS) we would exclude
consideration of support for electronic communication (e-mail, word processing,
etc) and organising tools (electronic calculator, automated rolodex, etc). In
Watson's view an EIS would be limited to providing information. Rockart & de
Long (1988) identify three broad ESS capabilities - Communication, Status
Access, and Query & Analysis. These are listed in ascending order of complexity
and skill required by executives. Watson excludes data analysis from the definition
of EIS but it is included here as an extra or adjunct function which an executive
can call from the status access function. Some executives will not want this. Some
will ignore this initially but take advantage of the capability as they gain
familiarity and develop confidence with the system. The interface for the
communication tools can possibly be ignored since these are specified by the off-
the-shelf packages adopted generally in an organization. Status Access largely
requires an interface to choose the displays which will be made available, the
information that will be included and how it will be formatted. Data Analysis will
require a more sophisticated interface with provision for input of various types as

Copyright © Genetic Computer School, 2002


Page147

well as more complex choices of what analysis should be done and which data will
be used. Status Access and Data Analysis may shade into each other depending on
how much control is given to the user in status access to vary displays. If a
standard data access tool is used then the system design may be constrained by the
tool specified. Excluding Communication/Organizing tools from a system leaves
Status Access/Query & Analysis as a single module, which can be looked at in
terms of a consistent interface.

Another aspect is the scope of the system. A common misconception of EIS is


that they are merely an easy to use graphical interface onto a small database
containing summarized information on the organization's progress. If the EIS
concept is limited to this, it is almost certainly doomed to failure. To implement a
successful EIS the system must be seen to include the whole organization and its
environment, from the executives who will use the information, to those
responsible for ensuring the integrity of the data in the main company databases
from which EIS database is fed. In most reports on EIS, developers of these
systems have emphasized the notion that "ease-of use" is their prime usability
constraint. However, in a study of executive use of EIS, Hasan & Gould (1994),
have found that no matter how `friendly' the interface, executives will only use the
system if they have confidence in the reliability of the information. Executives in
this study stated that it was far more important that information systems for
executive use, should have a consistent interface style, allow verification of the
accuracy of the information, allow annotation to explain exceptional cases and
permit sensible cross-functional comparisons of data from different systems.

11.3 Executive Cognitive

There are formal and abstract models of human cognition which deal with
‘language, inference and consciousness' such as Johnson-Laird (1983) but these are
not procedural or practical enough to be useful. Rockart & de Long (1988) outline
a number of models of the way executives operate but are founded on a rather
informal theoretical base lacking a solid psychological foundation. Allen (1994)

Copyright © Genetic Computer School, 2002


Page148

has studied the concept of usability which he defines as the non uniform effects of
system characteristics on user performance. He also explored the offering of choice
to users in the interface as a means of improving performance. This is important
because it may indicate that EIS need to be individually tailored to particular
cognitive styles. Er (1988) has classified the cognitive style of managers along two
dimensions; the preferred way of getting data and the preferred way of processing
it. In the former, an individual may be classified as a 'sensing' (S) type who prefers
hard data that deals with specific problems or an 'intuitive' (N) type preferring
holistic information that describes possibilities. On the processing dimension
individuals may be 'thinking' (T) using logic or other formal means for reasoning
or 'feeling' (F) where preference is given to personal terms in decision making.
Combining these dimensions gives four possible decision styles:

♦ Systematic (ST)

This decision maker prefers hard data and logic such as cost benefit analysis
and evaluation research.

♦ Speculative (NT)

This type prefers to speculate future possibilities with dicision trees and
sensitivity analysis.

♦ Judicial (SF)

This type prefers quantitative measures in conjunction with feedback from


decision making groups.

♦ Heuristic (NF)

Heuristic decision makers place great emphasis on current possibilities and


prefer reaching decisions through mutual adjustment.

The implications of this are that a decision making process needs to be


developed which takes into account the different cognitive styles of managers

Copyright © Genetic Computer School, 2002


Page149

making the decisions. For example 'what if analysis suits the speculative decision
maker and cost benefit analysis of hard data suits the systematic type. There is
some evidence (Phillips, 1984) that decision makers do not make decisions based
on the probability of success but rather on the basis of their aversion to failure due
to the more drastic penalties. Phillips goes on to suggest that decision technology
should be centred on problem solvers (with experience, intuition and knowledge)
supplemented by information technology (computers, software, databases,
networks and modelling) and preference technology (value judgements, time and
risk preferences and trade offs).

11.4 Psychological Profiling

There seems to be a role for psychological profiling of executives as a first


step in the design of an EIS. These profiles are currently used extensively in
organisations for providing management with information about the future
potential of employees. The most common form of administering this profiling is
by a video taped interview with a professional psychologist. The tape is then
viewed and responses by the subject transferred to profile data which is input to a
statistical package for analysis. The design of the structured interview questions is
usually determined by close consideration of company policy and desired
characteristics of future executives. This technique seems readily adaptable to the
analysis of executive requirements for EIS. For example the determination of the
class of executive along the lines suggested by Er (1988) would be an important
first step. Other factors to consider in this profile would be to determine whether
the executive focussed on key issues and preferred to skip details, whether they
were able to make decisions in the absence of real data and what are their reactions
to being forced to make a decision under pressure.

11.5 The Psychological Basis of EIS

To obtain user requirements and draw up user profiles for the development of
any computer system it is necessary to have all understanding of its prospective
users. EIS are no different in this respect except that the users are high powered

Copyright © Genetic Computer School, 2002


Page150

company executives. With conventional computer-based information systems,


cognitive science has been used with some success in defining user characteristics
as part of an "information processing loop" (Kaptelinin 1994). It is a form of
natural-scientific materialism that has evolved in the wake of the scientific and
technological revolution associated with the building and use of computers. The
human mind was seen as a special kind of information processing unit and, when
viewed in this light, the similarity to the newly emerging computer technology was
compelling. But, typically, it fails to distinguish between the psychic in the human
and the information processes of the machine. When cognitive psychology is
applied to defining user characteristics, the HCI system is seen as being composed
of two information processing units, the human beings and the computers, so that
the output of one unit is the input to the other. In this light, the role of the
computer is seen by many as "replacing" the human in intellectual spheres, so that
the two loops both human and computer are compatible and in fact identical so that
they can be considered as one extended loop. However, as Kaptelinin (1994) points
out, this combined information processing loop, made up of two other loops, is
itself closed, making it difficult to take into consideration any phenomena that
exist outside the loop. Dynamic, interactive systems such as EIS highlight the
severe limitations of cognitive science as a theoretical basis to explain the
interaction of people with the computer. It does not provide an appropriate
conceptual basis for computer use in its social and organisational context in
relation to the goals, plans and values of the user (Kaptelinin, 1994). A more
holistic approach such as that provided by Activity Theory may be more helpful.

Activity Theory is able to conceptualise humans in their context and positions


tools as the means of mediating work activities and developmental processes.
Taking on an Activity Theory approach means that activity is the core point of
departure for any kind of human research. It provides a "bottom line" from which to
start analysing any type of human interaction be it with a computer or any other tool.
This centralised approach in research is in stark contrast to cognitive science which by
virtue of its unsound basic premise of the human as an information processor, has been
forced into a piecemeal approach. It has also been argued that one of the main

Copyright © Genetic Computer School, 2002


Page151

weaknesses in traditional Western psychology is the establishment of so many


individual schools of psychology that none of them will ever be able to investigate any
single problem in enough detail to arrive at a satisfactory answer (Wertsch 1981).

Activity Theory, however, is not a monolithic "theory", but more a frame to be


filled out and a set of insights to be utilised - a set of basic principles evolved from a
dialectic materialistic approach to understand human life. All the principles that have
been delimited in Activity Theory are no more than premises that determine the general
direction of development of contemporary psychology (Asmolov 1987). An analysis of
human-computer activity would permit more complete use of the possibilities of the
development of human activities, which are revealed thanks to computers. The main
applied psychological problems revolve around how to make sure that people using the
computer are able to further improve their thinking.

An EIS itself does not have a clear cut goal as do most conventional computer
based information systems. It is rather a tool which has the flexibility to be used as a
source of information when needed. On the other hand executives are often very goal
oriented in their work on a specific task and rely on sets of intuitive procedures which
are known to have worked previously. This relates well to an Activity Theory approach
as outlined in Boedker (1991) where she maintains that all activity is bound to a goal
and/or an object and the characteristics of the goal or object partially determine and
structure the activity. The goals or objects of activity, undertaken by executives in their
day-to-day tasks, vary and cannot be anticipated when an EIS is being developed. This
concept is rarely appreciated by EIS developers. The relationship of goals, objects and
activities, as found in Activity Theory, provides a framework on which new
development. methodologies can be created for the building of flexible EIS where the
goals of the user can be determined as the system is used.

11.6 Conclusion

Mediated mental processes are the central thesis of Activity Theory. Just as the
use of language represented a new stage in the development of human higher mental
processes can we, as Tikhomirov (1981) poses, "distinguish a new stage in the

Copyright © Genetic Computer School, 2002


Page152

development of higher human processes" as a result of using the computer. If


computers were able to "think" we need to be able to distinguish how mediation of
mental processes by the computer differs from mediation by signs such as language. He
explains why it is "not a new stage in the path to internal mediation" but "is the further
development of external mediation". The use of computers in storing and searching for
information, in solving pre-programmed problems and other computer specialities, he
says, does not mean the "disappearance of human thought, but the reorganisation of
human activity". This leads to the appearance of new forms of mediation in which the
computer as a mental activity tool ultimately transforms the very mental activity which
it was devised to assist.

Nowhere is the concept of mediation more evident than in the use of EIS for
organisational strategic decision making. EIS is used as a flexible and exploratory tool
by the executive to identify problem areas and windows of opportunity thus
transforming the very activity it was designed to assist. In this way by continued EIS
use the executive's internal model of the organization is modified and new demands are
made of the EIS. The EIS must be flexible enough to adapt to these changing
requirements and to evolve along with the executive if it is to take its place as a useful
tool in a long line of other artefacts adopted by humans EIS over the centuries.

Much work has still to be done in this area in the adaptation of currently available
psychological profiles so that they provide information about executives, which can be
used as input to HO design of EIS interfaces. But the topic seems to be promising as a
future research area.

Copyright © Genetic Computer School, 2002


Page153

CHAPTER 12: DECISION SUPPORT SYSTEMS

12.1 Decision Support Systems

Introduction

A Decision Support System is used by Upper Management to make decisions for


the future of the company.

A DSS is a set of very special computer programs that are connected to all internal
networks but use mostly external Wide Area Networks, the information from which is
used to make corporate decisions.

A DSS uses mostly external data and is more hardware complex than the other
forms of MIS.

At the operational level of a business, decisions are often required to be taken by


relatively junior staff but with the minimum of delay. Decision Support Systems are
programmes, which allow staff to input a relatively small amount of specific
information, which dependent upon whether or not that data falls inside acceptable
parameters, provides a decision. This avoids the need to 'refer upwards' in all but the
most exceptional circumstances.

A Decision Support Systems (DSS) is a computer-based information system that


provides a flexible tool for analysis and helps managers with nonroutine decision-
making tasks. Some features of DSS are:

♦ Input and output: Inputs consists of some summarized reports, some processed
transaction data, and other internal data. They also include data that is external
to that produced by the organization. The outputs are flexible, on-demand
reports with which a top manager can make decisions about unstructured problems.

♦ Mainly for top managers: A DSS is intended principally to assist top managers,
although it is now mainly used by other managers, too. Its purpose is to help

Copyright © Genetic Computer School, 2002


Page154

them make strategic decisions - decisions about unstructured problems, often


unexpected and nonrecurring. These problems may involve the effect of events
and trends outside the organization.

♦ Produces analytic models: The key attribute of a DSS is that it uses models. A
model is a mathematical representation of a real system. The model allows the
manager to do a simulation-play a "what-if' game - to reach decisions.

The four capabilities of all DSS are:

♦ Representations: Conceptualization of information used in decision making,


such as graphs, charts, reports and symbols to control operations.

♦ Operations: Logical and mathematical manipulations of data.

♦ Memory aids: Databases, views of data, libraries and other capabilities to


update memory.

♦ Control Aids: Capabilities that allow the user to control the activities of the DSS.

Three basic components of a DSS:

♦ DSS database: a collection of current or historical data from a number of


applications or groups, organized for each access by a range of applications. The
DSS database is separated from active organizational data but contains data
extracted from corporate databases.

♦ Model base: a collection of mathematical and analytical models that can be


made easily accessible to the DSS user. Various kinds of models may be in the
model base, including libraries of statistical, optimization; sensitivity analysis
("what if”) and forecasting models.

♦ DSS software: permits easy interaction between users and the DSS database and
model base and delivers the end-user interface.

Copyright © Genetic Computer School, 2002


Page155

A DSS assists management decision making by combining data, sophisticated


analytical models and user-friendly software into a single powerful system that can
support semi-structured or unstructured decision making. These systems help end users
utilize data and models to discuss and decide semi-structured and unstructured problems,
but they do not solve the problems for the user.

Unstructured problems are novel, non-routine and have no predefined algorithms or


solutions. DSS combine data with models to produce various alternative scenarios that
can be used for making choices. In large organizations decision making is inherently a
group process, and DSS can be designed to facilitate group decision making.

DSS generally use smaller amounts of data than MIS, and do not need on-line
transaction data. DSS have a smaller number of important users, and tend to employ
more sophisticated analytic models than MIS. DSS are customized to specific users and
so require even more user participation than building MIS. Moreover, DSS are
continually evolving and changing, so they must use a flexible, iterative method of
development, usually prototyping.

Successful management depends on one's ability to make good decisions. Good


decision making relies on the availability of information. The Decision Support System
(DSS) can provide that information.

Today DSS uses operational files from Purdue's human resources and financial
information systems to create DSS tables. Business offices and academic units need, this
information to manage their operations. DSS allows access to an Oracle server containing
the administrative database. Query tools such as BRIO 3.5 and other software that
support Open Database Connection (ODBC) such as Microsoft Excel 5.0 are used to
retrieve the data. Information can then be processed in customized and ad-hoc reports,
graphics and tables for management decision-making.

Decision support systems help the decision making process by providing structured
sets of spatial information. Anything from a photograph to a map can be used to support

Copyright © Genetic Computer School, 2002


Page156

decision making and in developing a decision support system, experts are consulted to
ensure accuracy and understanding of the information being used.

A decision support systems is based on the requirements for making a specific range
of decisions and may be quite unsuited to a purpose other than that for which is has been
structured. When placed in context with environmental management and geographic
information systems, these tools may be described as highly specialised computer-based
systems which use expert knowledge and have a specific range of uses.

Geographic information systems may be used as decision support systems but not
necessarily vice versa. That is, we clan form late decision support systems from various
sources of spatial information using a geographic information overlay system which may
have no use other than for the range of questions it was designed to help answer. Of
course a decision support system may, have some dynamic modelling capabilities not
available in a geographic information system, particularly where non-spatial data sources
need to be incorporated in a model.

Challenge of Decision Support System

The challenge/purpose of DSS is to help people avoid mistakes in the various


different phases of decision making. The primary mistakes are:

1. Intelligence: Missing Important problems or pieces of information

2. Decision design:

♦ Focusing on the wrong decision

♦ Identify the wrong set of alternatives

♦ Not correctly identifying the outcome of the alternatives

3. Choice:

♦ Misapplying the criteria

Copyright © Genetic Computer School, 2002


Page157

♦ Using incomplete criteria (Mismatch between criteria and goals)

4. Implementation: Many (A whole, very different problem)

Decision Support Systems and Executive Support Systems focus on supporting the
fast 3 phases.

Components of a DSS

1. Data (Usually drawn from the organizations databases)

2. Models: A collection of models that can be used to determine the outcome of


various alternatives

3. Sensitivity analysis features - How do the results change if the values change (not
if the model changes)

4. DSS User Interface & Integration Software: The part of the system that lets the
users user the data in the model

Uses of DSS

They started off aimed as helping senior executives, but they have ended up being
used extensively by middle managers and other professionals (e.g. engineers).

[This is because the types of decisions faced by the two groups differ in terms of
their frequency, the kind of information that is needed, the amount of structure in the
problems]

General examples

Pricing decisions, route planning, budget planning, production planning, inventory


planning, site selection

Copyright © Genetic Computer School, 2002


Page158

Specific examples

• MRP - Manufacturing Resources Planning systems (and the DSS that go


with them) Problem is this:

• Orders -> (Some regular, some not regular)

• Production facilities (How things are produced is different for


different orders)

• Materials -> (Can take time to get there, buy lots - but not too
much, inventory costs money)

It is often hard to answer questions like:

Can you ship 10 batches of product X by next week?

How long will it take to produce a batch of Y?

Requires a complete model of the production process - DSS can help by providing
this model and detailed information about the current status.

• Police Crime map - An Intelligence System

Geographic information systems - Providing a useful representation of large data


sets (Useful for site planning, ATM placements, etc)

• Egyptian cabinet

• Complex decision

• Many alternatives

• Relatively unique decision

In this case the value of the system was probably the process of building it. To
build a DSS that was acceptable they needed to consider in detail the situation - quite
possible more detail and with more consistency than they otherwise would have.

Copyright © Genetic Computer School, 2002


Page159

12.2 Group Decision Support System (GDSS)

Introduction

People meet to create better, more efficient and stronger organizations in


Innovation Suites designed and built by Group Decision Support Systems. With an
ideal mix of technology and personal environment, furniture and open space, the
Innovation Suites foster learning, decision making and the formation of corporate visions.

These designs have been used by a variety of clients in a number of applications,


including strategic planning, procurement evaluation, business process innovation, and
large scale change management. The design principles are simplicity, productivity, and
an appropriate level of balance between technology and human interaction.

A group decision support system (GDSS) enables teams of co-workers to use


networks of microcomputers to share information and cooperate on projects. A GDSS is
also called workgroup computing.

In a GDSS environment, there is usually a big room with something like 40 seats,
which means that 40 people can be at the meeting at any one time. There are not only
40 seats but also 40 microcomputers. This enables every participant to have the use of
one microcomputer during the course of the meeting. The reason why each participant
needs a microcomputer depends on how GDSS works.

In the GDSS, with special computer software, the facilitator of each meeting will
first make the agenda of the meeting, which will be projected onto a big screen that
everyone can see. Then the participants will type simultaneously in their ideas of the
topic of discussion on the individual microcomputers next to them. Then the computer
will sort the ideas, and then the participants will then vote or comment on which ideas
they like or they dislike. In the course of the whole meeting, GDSS stores, categorizes
and prints out all the ideas, comments and vote tallies, so that each of the meeting
participants will get a summary of the meeting when it ends.

Copyright © Genetic Computer School, 2002


Page160

What so special about GDSS is that it enables meeting participants to


simultaneously "talk", when the computer sorts and sends ideas to each of the terminal,
all at the same time. That saves a tremendous amount of time, because all these are done
electronically instead of manually, and the time saved will enable participants to spend
more time manipulating and expressing their ideas. This can consequently increase the
productivity and efficiency of the group. The time-consuming benefit also has an added
bonus: when productivity and efficiency in meetings increase, it is likely that the team
spirit can be consolidated, resulting in an increase of the strength of binding among team
members.

Besides, under this GDSS, no one can dominate the meeting. This is because of
another feature of GDSS. GDSS provides an anonymous scheme, so that whatever you
type in the terminal (i.e. your opinion) will be protected. Under this circumstance, no one
really knows who is typing what. Because of this, not a single person can dominate the
meetings. In the worst case, we might say "some ideas" are dominating the meeting, but
this is perfectly fine because this is as a matter of fact an aim of the GDSS: to help
meeting participants voice their opinions from an idea-oriented mindset. For example,
simply because you have a prejudice against person A does not mean that you are going
to reject the idea being proposed in the meeting, because you do not know who is
proposing that idea!!

Besides, this anonymity scheme will also help those team members who are shy to
voice opinions. And with the anonymity, people are likely to be more honest, just as
you'll say more, and more honestly on the professor's evaluation form if you know
whatever you write will not affect your final grade on the course. This, of course„ is
because you know you don't have to worry about the consequences.

However, whether this anonymity is good or not can be very controversial. The
success of meetings supported by GDSS depend largely on the conduct of the
participants. If people are taking advantage of the anonymity system by typing obscene
words or foul languages, this system may be banned for the good of the organization.

Copyright © Genetic Computer School, 2002


Page161

The three underlying problems in group decision making that led to the
development of GDSS are:

♦ The explosion of the number of decision-maker meetings (managers spending


35 to 70 percent of their time in meetings).

♦ The growing length of decision-maker meetings.

♦ The increased number of attendees at these meetings--optimal size of traditional


meetings is only three to five participants.

Three elements of a GDSS:

♦ Hardware: including the conference facility itself, the room, tables, chairs, the
layout of the facility, electronic hardware, audiovisual equipment and computer
hardware;

♦ Software tools: including tools for organizing ideas, gathering information, and
ranking and prioritizing;

♦ People: including participants, a trained facilitator and hardware and software


support staff.

GDSS software tools:

♦ Electronic questionnaires: aid the organizers in pre-meeting planning by


identifying issues of concern and by helping to insure that key planning
information is not overlooked.

♦ Electronic brainstorming tools: allow individuals contribute ideas on the topics


of the meeting.

♦ Idea organizers: facilitate the organized integration brainstorming.

♦ Questionnaire tools: support the facilitators and group leaders as they gather
information before and during the prioritization process.

Copyright © Genetic Computer School, 2002


Page162

♦ Voting or prioritizing tools: make available a range of methods of prioritizing


and decision making, from simple voting, to ranking in order, to a range of
weighted prioritizing and voting techniques.

♦ Stakeholder identification and analysis tools: use structured approaches to


evaluate the impact of an emerging proposal upon the organization, and to
identify stakeholders and evaluate the potential impact of those stakeholders
upon the proposed project.

♦ Policy formation tools: provide structured support for developing agreement on


the wording of policy statements.

♦ Group dictionaries: document group agreement on definitions of words and.


terms central to the project.

♦ Additional tools might include: group outlining and writing tools, software that
stores and reads project files and software that allows the attendees to view
internal operational data stored by the organization's production computer
systems.

The Commander

Figure 12.1

The Commander (refer to figure 12.1) If the corporate boardroom is both a


showcase for beauty and the "heart" of an organization's decision making process, then
the Commander Innovation Suite is the ideal platform. Originally designed for the
Department of Defense's Air Staff Innovation Center, the Commander is built with the

Copyright © Genetic Computer School, 2002


Page163

highest quality materials and exceptional attention to detail. For your organization, the
Commander's classic "c" shape provides executives with a unique sense of team, while
the 40" of elbow room will deliver ample personal space and comfort.

The Executive

Figure 12.2

As the signature table of the IS line, the Executive is a high quality platform that
serves both as an exquisite center-piece and a powerful collaborative tool(refer to figure
12.2). Utilized by Learning Organizations across the country, including GDSS, the
Executive addresses the demands of an organization that needs to unleash the creativity
and innovation of its employees while maintaining an atmosphere of elegance.

The Diplomat

Figure 12.3

Simplicity often produces the greatest results. The Diplomat was forged from this
ideology. This table incorporates the basics of group collaboration into a simple, elegant
design that is perfect for teams' ongoing process work. (Refer to figure 12.3) Available
with or without the incorporation of collaborative technology, the Diplomat's compact

Copyright © Genetic Computer School, 2002


Page164

design is ideal for small conference rooms or break out areas. As a stand alone table, or
as a compliment to a larger Innovation Suite, the Diplomat will provide your organization
with a tool that empowers teams to produce quality results in less time.

The Two Top

Figure 12.4

The Two-Top design was born out of a demand for a flexible collaborative tool that
could meet a variety of needs in a taxing environment (refer to figure 12.4). Built with
the same attention to detail and quality as the full size Innovation Suites, the Two-Top
table takes advantage of sliding-covered laptop computers, self-contained networking
hardware, and unique "daisy-chaining" capabilities to produce a multitude of
configurations. This configurable platform increases the flexibility of boardroom
environments, serving as a traditional conference table, a roll-around work station, or a
powerful collaborative tool.

Advantages and Disadvantages of GDSS

Advantages

Group Decision Support Systems give groups several advantages over many
traditional, non-automated group meetings (Nunamaker, et al., 1991):

• More participation

Because a GDSS allows anonymity, group members may be encouraged to


participate because they do not feel as vulnerable to group censure for asking what may
be perceived as "foolish" questions or making unpopular comments. Similarly, the

Copyright © Genetic Computer School, 2002


Page165

participants will not be as subject to group think or conformance pressure (the reluctance
to criticize the comments of others due to politeness or fear of reprisals). In addition, each
group member will have more "air time" or time to contribute ideas. In non-automated
meetings, people must listen to others speak and pausing to reflect can cost a turn at
comment or response; a GDSS allows everyone to "speak" in parallel. In a typical
meeting, group members have only few minutes to express their ideas rather than the
entire meeting time. In some non-automated meetings, a few group members may exert
undue influence or monopolize the group's time; a GDSS makes every participant equal,
eliminating member status incongruities. Finally, more information will be presented to
the group as more participate.

• Group synergy

Other group members will be able to use an idea in a manner that the originator did
not because participants have different information skills. Also, the group as a whole will
be able to catch errors in a comment better than the individual who proposed the idea.
Reading a comment often gives creative stimulus to others in the group. Also, groups
may be more likely to consider an idea as the group's idea rather than an individual's
because ideas have been merged together.

• Automated record keeping

A GDSS can record all comments generated during the meeting, and consequently,
the group participants may not need to take notes. In a non-automated setting, group
members have to remember comments (rather than thinking of new ones) until they have
a chance to speak. Participants may also forget what has been said before. In vocal
meetings, some participants may not understand what was said or they may not be able to
process the information quickly enough. This automated log of the discussion supports
the development of an organizational memory from meeting to meeting.

• More structure

A GDSS also provides a certain amount of structure to the meeting. With this
structure in place, it is more difficult to deviate from the problem-solving cycle and make

Copyright © Genetic Computer School, 2002


Page166

incomplete or premature decisions. The group has a more concentrated discussion, and
they stay focused on the issues throughout the meeting. Lower levels of non-task
interactions (gossiping, for example) in such groups have been observed as compared
with traditional meetings.

• Other benefits

As a result of more participation, group synergy, record keeping, and structure,


many groups have been able to accomplish more in significantly less time necessary for
traditional, non-automated meetings. Also, these factors have contributed to higher group
satisfaction with the meeting process. Finally, the new technology has enabled larger
groups to meet, resulting in more information, knowledge, and skills that are brought to
bear to the task at hand.

• Disadvantages

There are some disadvantages to the technology, however, and they include:

• Slow Communication

Most people speak much faster than they type, and thus would usually prefer a
verbal environment (other things being equal). However, a GDSS allows participants to
review recorded comments (people may read and scan faster than they can hear and
process). Other advantages, including anonymity and parallel communication, may
override the slow typing speed. The breakeven point, where it is more efficient to type in
parallel rather than speak and listen in sequence, occurs at a group size of approximately
eight members (depending upon typing speed).

• Not all Tasks are Amenable to GDSSs

Group meetings, which involve "one-to-many" communication (for example, a


leader lecturing to the group) would not benefit from a GDSS. Only those tasks, which
require group members to exchange ideas or preferences efficiently ("many-to-many")
would benefit.

• Conclusion

A Group Decision Support System (GDSS) presents an efficient and effective


method for large groups to conduct meetings in which comments or preferences must be
exchanged.

Copyright © Genetic Computer School, 2002


Page167

Chapter 13: expert systems

13.1 Introduction

Expert systems are computer programs designed to review a set of facts (market
conditions) and apply a set of rules (knowledge base) to arrive at the same conclusion
that a team of experts would make if presented with the same set of facts. There are two
primary components to an expert system: knowledge base, and inference engine.

An expert system begins by capturing human expertise in the form of IF/THEN


rules. For example, the rule "IF the traffic light is red THEN step on the brake," captures
one small aspect of driving expertise. Often dozens (or hundreds) of such rules are
required to arrive at solutions to complicated, financial market problems.

The next step in the process involves applying the rules against a set of facts to
determine which rules apply. Embedded within the expert system is an inference engine,
which determines when and how to apply the rules. The unique order in which the rules
are applied to a given set of facts allows the expert system to be dynamic. As one might
expect given the features and capacity of these technologies, the number of potential
good solutions to a problem can be vast. The inference engine helps to quantify the
solution sets to provide our portfolio managers with consistent, unbiased, unemotional
problem analysis.

Recently, expert systems based decision models have been applied to various areas
within the auditing domain (McCarthy et al. [1992]). These models function adequately
when compared to human judgements. However, expert systems based techniques may
not be the only approach to decision problems within the auditing domain. Another
approach that is often available consists of conventionally coded decision support
computer models. It may be the case that these techniques provide a more efficient and at
least equally effective means of addressing problems within the auditing domain.

Decision support systems may have a potential for greater efficiency when applied
to some audit decision problems. This paper assesses whether decision support systems

Copyright © Genetic Computer School, 2002


Page168

are a match for the effectiveness of expert systems when applied to one particular audit
decision problem, namely, the assessment of inherent and control risk in the purchases
transaction cycle and the revision of the hours allocated to an audit program based on
those assessments. Two computer models were constructed, one based on an expert
systems shell (Clips Version 4.2, developed by the NASA Space Labs), and the other a
decision support model constructed using a conventional compiler (Microsoft QuickBasic
Version 4.5). Both models used the same data set as the basis of their knowledge base. This
data set was derived from a case study questionnaire administered to a sample of 80 auditors.

The objective of constructing these two models was to compare the output they
generated to determine if they were comparably effective. This comparison is primarily
concerned with the consistency of the output, but it also addresses other factors such as
the format in which the output is presented and the degree to which the computer models
were accessible to the user.

Before identifying the data used by decision makers and the significance they attach
to them, it is necessary to develop a framework within which to ft these data items. The
audit risk model provides a decision framework that is prescribed for use by auditors for
audit planning decisions. The way that auditor's risk assessments impact on audit
planning provides a readily accessible field for the development of computer models. The
application of the audit risk model by practicing auditors has attracted considerable
research attention (Peters, Lewis and Dhar [1989]) Mock and Wright [1990] examined
the link between audit planning judgement and observed levels of inherent and control
risk. They argued that risk assessments have a stronger impact on the extent of audit
work carried out than on the nature of work carried out and that risk assessments may
be subject to change over time. The models examined in this paper use a fixed set of
audit procedures; the models adjust the hours allocated for performing those
procedures, not the list of procedures.

13.2 Expert System Architecture

The user interacts with the system through a user interface, which may use menus,
natural language or any other style of interaction). Then an inference engine is used to

Copyright © Genetic Computer School, 2002


Page169

reason with both the expert knowledge (extracted from our friendly expert) and data
specific to the particular problem being solved. The expert knowledge will typically be
in the form of a set of IF-THEN rules. The case specific data includes both data
provided by the user and partial conclusions (along with certainty measures) based on
this data. In a simple forward chaining rule-based system the case specific data will be
the elements in working memory.

Almost all expert systems also have an explanation subsystem, which allows the
program to explain its reasoning to the user. Some systems also have a knowledge base
editor which help the expert or knowledge engineer to easily update and check the
knowledge base.

One important feature of expert systems is the way they (usually) separate domain
specific knowledge from more general purpose reasoning and representation
techniques. The general purpose bit (in the dotted box in the figure) is referred to as an
expert system shell. As we see in the figure, the shell will provide the inference engine
(and knowledge representation scheme), a user interface, an explanation system and
sometimes a knowledge base editor. Given a new kind of problem to solve (say, car
design), we can usually find a shell that provides the right sort of support for that
problem, so all we need to do is provide the expert knowledge. There are numerous
commercial expert system shells, each one appropriate for a slightly different range of
problems. (Expert systems work in industry includes both writing expert system shells
and writing expert systems using shells.) Using shells to write expert systems generally
greatly reduces the cost and time of development (compared with writing the expert
system from scratch).

13.3 Components of an Expert Systems

Expert systems operate in a fundamentally different manner from conventional


"procedural" programs. That is, rather than run through a set of commands in the same
way every time, an ES uses "declarative" knowledge and an inference engine to drive
the reasoning process. The basic components of an expert system are described below
in figure 13.1:

Copyright © Genetic Computer School, 2002


Page170

Figure 13.1

• knowledge base: knowledge from "domain experts" typically represented as IF-


THEN rules,

• user interface: provides a means for the non-expert to interact with the
knowledge contained in the knowledge base,

• inference engine: the core of the expert system--determines how the rules in the
knowledge base are processed,

• working memory: an area of memory containing (a) observed facts, and (b) new
facts deduced from observed facts.

13.4 An overview of deductive inference

Expert systems work by making deductive inferences. The main advantage of


deductive inference is that it allows specific facts to be derived from general information
(inductive inference, on the other hand, uses specific facts to generate general
hypotheses) To illustrate the basic concepts of deductive inference, consider the well-
known syllogism:

Copyright © Genetic Computer School, 2002


Page171

All men are mortal (general information)

Socrates is a man (observation)

Socrates is mortal (specific information)

From an ES point of view, the syllogism consists of three parts:

1. an observed fact: Socrates is a man

2. a rule: All men are mortal

3. a new fact deduced from the observed fact and the rule: Socrates is mortal

The only difference between this syllogism and the inference mechanism of an
expert system is that in expert systems, the rules are represented differently. Specifically,
expert systems cannot operate on universal quantifiers such as "all" and thus variables
have to be used instead. For example, the rule.

All men are mortal would have to be re-written as follows for inclusion in an ES
knowledge base:

RULE 1

IF x is a man

THEN x is mortal

This so-called "IF-THEN rule" has two parts: an antecedent (IF x is a man)
and a consequent (THEN x is mortal). A rule is said to fire whenever its antecedent is
satisfied. When a rule fires, its consequent is instantiated and added to working
memory. To illustrate this, consider the following "inference trace":

Step 0 working memory = observations: Socrates is a man

Step 1 RULE 1 is satisfied: IF x is a man THEN x is mortal

Socrates is a man
working memory = working memory + Socrates is mortal

Copyright © Genetic Computer School, 2002


Page172

Note that there is nothing magic here. The expert system simply has the ability to
match patterns. For instance, it considers the antecedent of the rule and identifies the
following pattern: <variable name> is a man. Then it searches working memory for a
fact of the form <concrete object> is a man. Since a match is found, the variable x is
bound to the concrete object "Socrates" and the consequent Socrates is a man is
added to working memory. This new fact may be used to satisfy some other rule, and
thus the chain of inference continues.

13.5 Inference Strategies

The deductive process itself is relatively straightforward. However, there are two
fundamentally different approaches to deciding how to navigate the knowledge base:
forward chaining and backward chaining. In either case, the problem to be solved is
posed as a question, for instance: "Is Socrates mortal?"

In forward chaining, deduction is data driven. In other words, the goal is not
used to drive the inference process. Instead, the inference engine simply checks the
contents of working memory from time to time to see if the question has been answered.
The basic flow of the forward chaining process is shown below in figure 13.2.

Figure 13.2

Copyright © Genetic Computer School, 2002


Page173

In backward chaining, deduction is goal driven. The inference starts with the
original question and seeks to answer it by matching it with the consequent of a rule. The
basic flow of the backward chaining process is shown below in figure 13.3

Put the original question on the back


burner and make the unresolved
variable the new question

Figure 13.3

There are two important features o± backward chaining. The first is that the goal (or
question to be answered) always drives the search strategy. The second is that the current
goal changes as new unresolved variables are encountered. The result is that the
backward chaining procedure is called recursively until the original question in answered
or a sub-goal cannot be resolved.

Copyright © Genetic Computer School, 2002


Page174

13.6 The Problem with Expert Systems

The dependence people place on cases poses a problem to those who treat human
cognition as being primarily rule-based. Much work in artificial intelligence, for
example, is done in "expert systems." These systems are based on the notion that expert
knowledge consists of a collection of rules. By determining the rules an expert in a
domain uses, the idea goes, we may then simulate expert behavior in that domain. Not
surprisingly, expert systems have run into a significant problem: they are brittle. When
faced with a problem which bends the rules, they are unable to cope. They fail because
they are not grounded in cases. They are unable to fall back on the details of their
experience, find a similar case, and apply it. Likewise, they are unable to use similarities
between tough problems and previous experiences to update their rules. Their failure to
retain cases cripples their ability to learn from their experiences.

13.7 Choosing a Problem

Writing an expert system generally involves a great deal of time and money. To
avoid costly and embarrasing failures, people have developed a set of guidelines to
determine whether a problem is suitable for an expert system solution:

1. The need for a solution must justify the costs involved in development. There
must be a realistic assessment of the costs and benefits involved.

2. Human expertise is not available in all situations where it is needed. If the


"expert" knowledge is widely available it is unlikely that it will be worth developing an
expert system. However, in areas like oil exploration and medicine there may be rare
specialized knowledge which could be cheaply provided by an expert system, as and
when required, without having to fly in your friendly (but very highly paid) expert.

3. The problem may be solved using symbolic reasoning techniques. It shouldn't


require manual dexterity or physical skill.

4. The problem is well structured and does not require (much) common sense
knowledge. Common sense knowledge is notoriously hard to capture and represent. It

Copyright © Genetic Computer School, 2002


Page175

turns out that highly technical fields are easier to deal with, and tend to involve relatively
small amounts of well formalised knowledge.

5. The problem cannot be easily solved using more traditional computing methods.
If there's a good algorithmic solution to a problem, you don't want to use an expert
system.

6. Cooperative and articulate experts exist. For an expert system project to be


successful it is essential that the experts are willing to help, and don't feel that their job is
threatened! You also need any management and potential users to be involved and have
positive attitudes to the whole thing.

7. The problem is of proper size and scope. Typically you need problems that
require highly specialized expertise, but would only take a human expert a short time to
solve (say an hour, max).

It should be clear that only a small range of problems are appropriate for expert
system technology. However, given a suitable problem, expert systems can bring
enormous benefits. Systems have been developed, for example, to help analyze samples
collected in oil exploration, and to help configure computer systems. Both these systems
are (or were) in active use, saving large amounts of money.

13.8 Rules and Expert Systems

Rule-based systems can be either goal driven using backward chaining to test
whether some hypothesis is true, or data driven, using forward chaining to draw new
conclusions from existing data. Expert systems may use either or both strategies, but the
most common is probably the goal driven/backward chaining strategy. One reason for
this is that normally an expert system will have to collect information about the problem
from the user by asking them questions - by using a goal driven strategy we can just ask
questions that are relevant to a hypothesized solution.

Anyway, in a simple goal-driven rule-based expert system there are often a set of
possible solutions to the problem - maybe these are a set of illnesses that the patient

Copyright © Genetic Computer School, 2002


Page176

might have. The expert system will consider each hypothesized solution (e.g., has_Cold
(fred)) and try to prove whether or not it might be the case. Sometimes it won't be able to
prove or disprove something from the data initially supplied by the user, so it will ask the
user some questions (e.g., "have you got a headache?"). Using any initial data plus
answers to these questions it should be able to conclude which of the possible
solutions to the problem is the right one.

The Distinction Between Expert Systems and Decision Support Systems

Expert systems, or knowledge based modelling, represent an application of the


artificial intelligence field of research. The objective of this field is to imitate human
mental activity. While computers can more than match human capabilities in activities
that do not actually require "intelligence" or reasoning ability, they are not a clear
match where reasoning abilities are required. The artificial intelligence field can be
seen as an attempt to address this deficiency.

One difficulty faced by researchers in the artificial intelligence field is to


develop a suitable test, and its attendant definition, to identify what constitutes
intelligence. Such a test is provided by the Turing Test (Turing [1950]). The test is
based on an operational definition where a device is accepted as possessing artificial
intelligence provided it performs as a reasoning human would when presented with
the same problem. The test is administered as a simple question and answer process.
The device to be tested and a human volunteer are both interviewed by an examiner in
an attempt to identify which is the artificial device. If the examiner is unable to do this
after a series of tests the device is deemed to have passed the test.

The essential point of this test is that it places no restriction on what techniques
are used to attain a successful result. Thus, if we accept this test, and its implications,
artificial intelligence need not be the same form of intelligence, or function in the
same way as human intelligence, so long as they provide results that are comparable
with those produced by a human expert analysing the same problem. An extension of
this position is to say that expert systems based models, and procedural models may

Copyright © Genetic Computer School, 2002


Page177

both be equally valid if they produce comparable results without necessarily


duplicating each other's analytical techniques.

The central component of an expert system is the inference engine. This is


software that combines the contents of the system's knowledge base with case specific
input obtained from the user in order to perform the problem solving tasks for which
the system was developed.

The knowledge base is composed of domain specific knowledge pertaining to the


chosen problem domain. An essential characteristic of expert systems is to have a
specifically, and often narrowly, defined problem domain. A detailed and specific
knowledge base is the true source of the problem solving power of an expert system
[Waterman 1986]. The manner in which the knowledge base is used will be
determined by the system's meta-knowledge. Meta-knowledge can be seen as a set of
rules governing how the knowledge base itself is applied. Meta-knowledge determines
what case specific knowledge is obtained from the user and what knowledge from the
knowledge base is incorporated in the problem solving process.

Meta-knowledge and knowledge complement one another. In the absence of


meta-knowledge, knowledge cannot be applied. However, the extent to which we rely
on meta-knowledge may also offset the extent to which we need to rely on the
knowledge base. An advanced expert systems shell may facilitate the construction of
expert systems based models that embody a high degree of meta-knowledge. Such a
model will be able to solve problems while making a relatively "economical" use of
the systems knowledge base. It will be able to limit its use of the knowledge base to
those pieces of knowledge that are most pertinent to the problem being addressed. A
model with a less advanced meta knowledge structure may produce the same
conclusion, but it would be less discriminating in its usage of the knowledge base.

A procedural coded computer model can approach this type of structure. Such a
model can access a knowledge base composed of heuristic, or numerically based
rules. The programme code, which determines how that knowledge base is accessed
and applied' can act in the role of meta-knowledge. However, conventional programming

Copyright © Genetic Computer School, 2002


Page178

languages are not adapted specifically for the construction of meta-knowledge structures
while expert systems shells are. For this reason, the scenario outlined above may tend to
arise. While it is possible that procedural and expert systems based models may reach
comparable conclusions, the procedural model may utilise a considerably larger portion
of the knowledge base than is strictly necessary to do so.

This need not be a significant shortcoming. The actual cost of performing analyses
on an ongoing basis, using what may be a less efficient model, is relatively small when
compared to the potential cost of developing a more efficient model that involves much
more advanced developmental techniques.

Abdolmohammadi [1987] draws a distinction between expert systems, which seek


data in order to make a judgement or diagnosis, and decision support systems, which
perform analyses that assist in making a decision. Expert systems can be constructed out
of heuristic decision rules obtained from experts in the problem domain. The rules, while
relevant to the problem domain, will not all be relevant to each problem within the
domain. The expert system operates by selecting those rules that are relevant (and thus
the information that it should seek from the user) according to the problem outlined to it
by the user's responses. This flexible approach inherent in expert systems, which enables
them to selectively apply their knowledge base to a problem, renders them particularly
suitable for application to unstructured tasks.

Since a procedural model embodies a less advanced meta-knowledge structure, its


inherently less flexible approach must make a higher level of assumptions about the form
of the problem that is to be solved. Such a model applies an algorithm, which aims to
produce an optimal solution to that particular problem, and only to that particular
problem. To the extent that real world problems diverge from the assumptions of the
system's developers, so the solution may be less than optimal (although in practice users
of the system often may not recognise this). The more advanced meta-knowledge
structure of expert systems models has the potential to reduce the level of assumptions
that are made about the problem under analysis. It is our contention that the auditing

Copyright © Genetic Computer School, 2002


Page179

problem we examine here is sufficiently structured to enable it to be addressed by the less


flexible meta-knowledge structure of a procedural based model.

If problems exist which are capable of being addressed adequately by procedural


techniques, as well as by expert systems, procedural techniques have the potential to offer
a solution that is developmentally less complex and therefore a more efficient approach
to the problem.

Waterman (1986) presents three sets of criteria for identifying those situations in
which the development of expert systems is appropriate. These are summarised below.

13.9 Criteria for Expert Systems development

Necessary requirements for expert systems development:

♦ Task does not require common sense.

♦ Task requires only cognitive skills.

♦ Experts can articulate their methods.

♦ Genuine experts exist

♦ Experts agree on solutions.

♦ Task is not too difficult.

♦ Task is poorly understood.

Justification for expert systems development:

♦ Task solution has a high payoff.

♦ Human expertise being lost.

♦ Human expertise scarce.

♦ Expertise needed in many locations.

Copyright © Genetic Computer School, 2002


Page180

♦ Expertise needed in a hostile environment.

Characteristics that make the use of expert systems appropriate:

♦ Task requires symbol manipulation.

♦ Task requires heuristic solutions

♦ Task is not too easy.

♦ Task has practical value.

♦ Task is of manageable size.

The first of these, described as "necessary conditions" for the development of


expert systems, acre presumably met by many problems within the auditing domain
since satisfactory auditing based expert systems have been developed. Since this first
set of criteria is presented as a set of necessary conditions that must exist before
expert systems development can take place, the author presumably intends that these
criteria must be met.

Similarly, it would appear that Waterman's third class of criteria, the appropriacy
of expert systems, must be met completely. In the absence of these criteria, the
development of expert systems will be possible, however, the violation of these
conditions would present a situation where the development of an expert system
would not be justified, since the problem concerned could clearly be addressed by
simpler modelling techniques, or was not worthy of modelling at all. Once again, it
must be assumed that the presence of pre-existing systems denotes that problems from
within the auditing domain meet these criteria.

On the other hand, Waterman's second category, which seeks to codify those
situations where the use of expert systems is justified, as opposed to possible, would
not appear to need to be met completely in each case. Satisfaction of any one of these
criteria would appear to provide a justification for the use of expert systems to address
that particular problem, provided no simpler technique were available. It seems that at

Copyright © Genetic Computer School, 2002


Page181

least two of these criteria, being the (relative) scarcity of human experts, and high
solution payoffs, are met by problems from within the auditing domain.

A further condition should be added to Waterman's second category,


"Justification for Expert Systems Development". That condition is that expert systems
should provide the most efficient technique available to address the problem being
examined. Further, we suggest that this would not be the case where the decision
problem is sufficiently structured that a simpler degree of meta-knowledge, as
embodied in procedural code is sufficient to address it, and that it will not be the case
where a problem in the real world is capable of being characterized by an extensive
set of starting assumptions.

A further condition should be added to Waterman's second category,


"Justification for Expert Systems Development". That condition is that expert systems
should provide the most efficient technique available to address the problem being
examined. Further, we suggest that this would not be the case where the decision
problem is sufficiently structured that a simpler degree of meta-knowledge, as
embodied in procedural code is sufficient to address it, and that it will not be the case
where a problem in the real world is capable of being characterized by an extensive
set of starting assumptions.

Copyright © Genetic Computer School, 2002


Page182

CHAPTER 14: DISTRIBUTED INFORMATION SYSTEMS

14.1 Introduction

The litikage of minicomputers in each District (State) office and the Bureau
mainframe computer foTrns the, basis for a Distributed Information System (Posson, 1985).
The Distributed Information System (DIS) provides a high level of local computing
and data-processing capability. Computerized files are transferred between sites
(nodes) within the network and work is done at distant locations. The DIS provides
interactive and batch processing in support of Water Resources Divisions's State and
National water data files, and provides for the diverse computational needs of the
Division. These needs include data management, hydrologic modeling and statistics,
and administrative programs.

The DIS computers are connected to each other via a TCP/IP (Transmission
Control Protocol/Internet Protocol) over an ethernet Local Area Network (LAN). The
LANs are connected together using a frame-relay Wide Area Network (WAN) called
DOINET. The WAN uses Stratacom switches and Cisco routers for the network
service backbone.

As organizations and users increasingly rely upon networked applications for


accessing information and making critical business decisions, securing distributed
applications is becoming extremely significant. With the phenomenal growth in
networks such as the Internet, information acid security issues are becoming the focus
of intense concern on the part of users and businesses.

14.2 Automated Data Processing System

The Automated Data Processing System (ADAPS) consists of a collection of


computer programs (software) and data files that form a system of standardized water
data processing procedures. In each District, there is a designated administrator or
manager who is responsible for installation and maintenance of system programs and

Copyright © Genetic Computer School, 2002


Page183

hydrologic data files. In addition, the administrator performs other tasks that are
necessary for efficient operation of the system.

Most of the data needed by ADAPS programs are supplied interactively:


However, some programs accept input from system files. Therefore, the programs
require the pathname (data set name) of the input file. Occasionally the user is
prompted to supply a pathname (either for input or output), but usually the name need
not be known. Most pathnames are stored within a program or in a file containing
pathnames, thereby relieving the user of having to keep track of the pathnames.

A security system is used in ADAPS to restrict access to the data files and to
limit the ability of some users to perform certain operations in the system. The
security system is a multilevel system. User classes of System Administrator (SYST),
Data Base Administrator (ADBA), User (USER), and Cooperator (COOP) have been
established. In addition, Ingres Access Control Lists' (Seybold, 1985, p. 3-1 to 3-28)
are used. Security measures are implemented by the local administrator or manager in
consultation with District supervisory personnel.

Water data stored in ADAPS, results from processing of data collected or


recorded at field installations operated by each District. Data collection and
processing is discussed in the ADAPS User's Manual. A vast majority of the field data
is recorded in digital form on punched paper tapes or recorded by a data collection
platform (DCP) and sent to the computer via satellite telemetry. The telemetry input
subsystem is discussed later in this manual. The field data is processed by District
personnel following established procedures. The data are reviewed for validity and
correctness, and subsequently are published in State basic data reports.

14.3 Programs

The programs (software) for ADAPS are developed by personnel of the Water
Resources Division. The master copy resides at Headquarters in Reston Virginia, and is
electronically transferred over the network or distributed by magnetic tape from
Headquarters to each of the District offices. These offices are located nationwide and

Copyright © Genetic Computer School, 2002


Page184

include Alaska, Hawaii, Guam, and Puerto Rico. Some Districts have more than one
office, and a copy of the software is located at the subdistrict offices, if they have a
minicomputer.

Most ADAPS programs are written using the Fortran77 programming language, and
several different categories of software make up the water data-processing system. The
major categories are:

• System utility programs.

• General-purpose programs.

• Command Procedure Language (CPL) programs.

• General application programs.

• General graphics programs.

• Vendor-supplied programs (subroutines).

The programs m each of the above categories is used for a specific function or
purpose. For example, the utility programs are used to initialize, create, update, and
maintain the numerous s u p p o r t , processing, and data (time-series) files that make up
the District data bank. The general-purpose programs are used to process many different
types of water data along with the subsidiary calculations and computations that go into
computing and producing a water data record. Most programs use insert and common
blocks to share and communicate data between programs, to provide software flexibility,
and to ease software maintenance. The operating system software is called
UNIX(Seybold, 1985, p. 1-2). The Command Procedure Language (CPL) is a PRIMOS
command level language that provides a programming capability (Landy, 1982).
Operating system commands or directives are passed to PRIMOS for execution after they
are stored in a CPL file (suffix .CPL). The application programs are used primarily to
compute statistical information about the hydrologic data. The graphics programs are
used for preliminary viewing of the data, for comparison purposes, and for report

Copyright © Genetic Computer School, 2002


Page185

purposes. The system contains both user-written and vendor-supplied programs such as
DISSPLA (Integrated Software Systems Corporation, 1984).

The Automated Data Processing System (ADAPS) programs are used to compute
water-data records on an electronic computer. The machine operations generally parallel
the manual operations. The sequence of processing the data is well established; however,
the particulars of each step by electronic means may change in response to continued
improvement in storage and access procedures, new or expanded needs, and search for
additional efficiencies. The general functions of the programs are to provide input and
output to and from the computer in a logical sequence; this sequence includes the
computational steps necessary to efficiently compute water data records. Once the
records are computed, they must be saved and archived for general use. Therefore,
programs argil available to provide for a broad scope of functionality including
initialization, maintenance, security, backup, recovery, restart, and other overall data-
processing requirements.

14.4 Fibs

The ADAPS files are the repositories that contain the information and data
necessary to use ADAPS for computing and processing water data records as previously
mentioned. The files consist of program source code and associated executable code,
CPL files, and operation files.

Some of these operational files are shared by the various NWIS systems and some
are specific to ADAPS. The shared and ADAPS-specific files are briefly described in
Section 3 of this manual.

Most files used and/or maintained by ADAPS are structured as MIDAS files and
managed by a utility and user-written software. The MIDAS files allow records to be
retrieved rapidly and efficiently on the basis of selected data elements defined as key
indexes (elements). MIDASPLUS utilities, user-written programs, and scripts are used to
create the file templates, create input files, and subsequently populate the files. Other
utilities are used to dump the files, delete files, cleanup files, and monitor files.

Copyright © Genetic Computer School, 2002


Page186

Some information regarding the file creation was provided in informal instructions
(Trapanese, S.M., U.S. Geological Survey, written common., 1987) to install ADAPS on
the local minicomputer and to convert existing minicomputer and mainframe data files
for use in ADAPS. Additional information concerning initialization (creation) of files is
presented later in this manual. ADAPS menu (program) options are selected to perform
the various functions necessary to create new MIDASPLUS files.

Information is also presented later in the manual to perform incremental backups of


the unit-values data files, and to restore the unit-values files using the incremental
backups.

14.5 Need for File Maintenance

In virtually every data-processing system, the possibility exists that errors may
occur that accidently alter or destroy data stored on disks in data files. This may occur
because of unexpected hardware failures, natural disasters such as power outages, etc., or
through improper processing of the data.

It is essential, therefore, to provide a means to ensure that any lost data can be
recovered. The most common method used is backup files. A backup file is merely a
copy of a file stored on magnetic tape or disk. If the file is destroyed or becomes
unusable, the backup can be used to recreate or restore the file. In transaction-oriented
(online) systems, backups are critical because updates to a file can occur at any time.
Therefore, the system should provide for creating backups on a regular basis (sometimes
every hour or day) and saving the transactions that occur to the file after the backup file is
made (incremental backups). If the file is lost, it can be re-created from the backup file
and then the transactions which have been saved can be processed against the file to bring
it back to the status it was before it was lost.

Another aspect in a transaction-oriented system is the ability to restart the system if


it goes down. For example, if a transaction (update or modification) has been sent and has
partially updated a file when power is lost, it is necessary to both restore the file and let
the user know that the transaction was not processed. This problem may be solved by

Copyright © Genetic Computer School, 2002


Page187

using the backup files and messages sent to terminals asking the users to re-send those
transactions not processed.

14.6 TalkMine: Knowledge Mining in Distributed Information Systems

TalkMine, an algorithm developed as a model of human cognitive categorization,


improves human-human and human-machine interaction by establishing a powerful tool
for collaborative knowledge organization in large distributed information systems (DIS).
The adaptive nature of TalkMine entails an open-ended human-machine symbiosis, which
can be used in the automatic, adaptive, organization of knowledge in DIS such as the
Internet, facilitating the rapid dissemination of relevant information, the discovery of new
knowledge, as well as the ability to identify fraudulent behavior in DIS. TalkMine is
based on several theories of uncertainty and belief, as well as connectionist ideas.

Even though knowledge is ultimately grounded in the brains of the users of DIS,
we can devise mechanisms that infer how tokens of knowledge in DIS relate to one
another, from the way these tokens are used by a community of users. Tokens of
knowledge are all the different types of information stored in the nodes of a network, e.g.
text files, databases, links between documents, etc. The meaning of these tokens is
ultimately established by the users of DIS, both individuals and societies. Whether
artificial information networks can ever produce autonomous creation of meaning is an
open question. What they can do exceptionally well, however, is to monitor how their
users manipulate and relate the tokens of knowledge stored. TalkMine does not pursue a
passive analysis of relationships of knowledge tokens, but rather an active categorization of
these tokens by changing the structure of a DIS to adapt it to the changing interests of its
users. We refer to this active organization of distributed information as knowledge mining.

The kind of knowledge management proposed with TalkMine will result in the
evolution of a "Second Generation Internet" or DIS. Theoretically, TalkMine is based on
the idea that there cannot be creation and open-ended evolution of new knowledge in
artificial systems until a fundamental coupling between structure and semantics is
established, that is, until some artificial embodiment is implemented. TalkMine empowers
DIS with such a coupling as a result of the continual feedback between the level of

Copyright © Genetic Computer School, 2002


Page188

information utilization by users (semantics and pragmatics), and the relations and links
between tokens of knowledge in networks which define the structure of DIS. The ability
to gather and combine information from different neighborhoods (e.g different networked
databases or WWW sub-regions) of DIS as a result of the continual integration of user
(interactive) queries, is shown to provide this ability for open-ended construction of
knowledge categories.

In practice, the kind of automatic, adaptive, self-organization of knowledge


implemented by TalkMine dramatically changes the management of information in DIS
from a centralized to a distributed new vision. Instead of the current reliance on central
databases (in Intranets) and Search Engines (on the Internet), any portion of a DIS
managed by TalkMine is capable of pointing users to relevant information in its
neighborhood. In other words, the TalkMine managed DIS is itself a distributed database
or search-engine which is continually adapting to and learning from its users and their
patterns of information usage. Each node of the network learns to relate its information to
other nodes from repeated interaction with users. With this permanent feedback loop
between user and network, the structure of the DIS reflects the knowledge of its
community of users. DIS with TalkMine function both as collaborative information
networks and as ever-present, distributed, search engines.

Copyright © Genetic Computer School, 2002


Page189

MANAGEMENT

INFORMATION

SYSTEMS

First Edition

Work Book

Copyright © Genetic Computer School, Singapore.

All rights reserved. No part of this publication may be reproduced, stored in a


retrieval system, or transmitted in any form or by any means, without the permission of
Genetic Computer School.

First Edition 2002

Genetic Computer School. 257 Selegie Road, #03-287B, Selegie Complex,


Singapore 188350.

Copyright © Genetic Computer School, 2002


Page190

Chapter 1

Question 1

a) Give an example of a professional job, and describe one or more information


systems that would provide each of the five types of benefits listed below:

i. Support for Strategic Planning

ii. Support for Management Control

iii. Support for Operation Control

iv. Improved Product Quality

v. Improved Product Delivery

List the five components of an information system and describe four roles
people play in information environment.

Question 2

What are the main reasons behind information management?

Question 3

What are the main characteristics of information?

Question 4

Explain in detail what is an information system. Give examples.

Question 5

a) Define CBIS. Explain why information has to be managed.

Copyright © Genetic Computer School, 2002


Page191

b) Information is a valuable asset in most organisations. So managers have


to be clearly focused on what information is important to them and what in not.
Explain with proper example.

c) What are the steps to be taken by managers to kept proper way valuable
information for in most organisations?

d) Information is one of the most important in very sector, either commercial


or noncommercial. Explain how and which ways it will be provide to keep proper
information.

Question 6

Support for Strategic Planning Support for Management Control Support for
Operational Control Improved Product Quality Improved Product Delivery

What is the main purpose of computer-based information systems?

Copyright © Genetic Computer School, 2002


Page192

Chapter 2

Question 1

Write short notes on the following:

a) Transaction processing

b) Decision support system

c) Office automation system

d) Management Information

e) Systems Exception Reporting

f) Data Warehouses

g) On-line Analytical Processing

Question 2

Describe the five categories into which the applications of Management


Information Systems fall.

Question 3

Describe the applications into which Management Information Systems fall.

Question 4

What is business network design? What are the four types of network redesign?

Question 5

What are the main uses of Transaction Processing Systems?

Copyright © Genetic Computer School, 2002


Page193

Question 6

Explain how on-line analytical processing tools assist in maintaining data archives?

Question 7

a) Discuss the features of Transaction Processing Systems, which are important in


creating systems and solutions.

b) Explain the difference(s) in building Executive Support Systems and


Traditional Management Information Systems?

Copyright © Genetic Computer School, 2002


Page194

Chapter 3

Question l

Distinguish between homogeneous and heterogeneous workgroups.

Question 2

Describe all the phases of the system development diagrams.

Question 3

Explain the terms

a) Analytical support

b) Communication support

c) Tracking and monitoring support in the context of personal information systems.

Question 4

What do you understand by the term organisation? Outline the three that
organisations share in common.

Question 5

Differentiate between personal and workgroup information systems.

Question 6

With the help of a diagram, illustrate the system development process.

Question 7

What is the system development life cycle? Briefly explain the different stages of
the development life cycle.

Copyright © Genetic Computer School, 2002


Page195

Question 8

Write short notes on the following:

a) Alternative Quality Process

b) System

c) Functional Specification

Copyright © Genetic Computer School, 2002


Page196

Chapter 4

Question 1

With respect to the evolution of IT applications, explain what do you


understand by the automation of work.

Question 2

What are the possible benefits arising from information management?


Illustrate your answer with examples.

Question 3

Define the term business transformation. What is its impact on the enterprise?

Question 4

Briefly describe the stages through which IT has evolved.

Question 5

What has been the impact of the emergence of Internet and virtual banking on
the financial industry?

Question 6

Explain why management thinking has failed to understand the implications


of the evolving role of IT in business system.

Question 7

Current management practice fails to adequately address the impact and


resource implications of four critical dimensions of complexity. Explain why.

Copyright © Genetic Computer School, 2002


Page197

Question 8

a) Organisations have had to make changes in their provision of Information


Service to support End User Computing (EUC).

i. Describe one way in which support of EUC has changed the structure of an
organisation's information service provision.

ii. Briefly describe the general changes in the nature of work undertaken by
information services personnel that you would expect with the introduction
of EUC into an organisation.

b) A Data Centre may be centralized or de-centralized. Describe the issues


you would consider when deciding between a centralized Data Centre provision
and a decentralized Data Centre provision.

Copyright © Genetic Computer School, 2002


Page198

Chapter 5

Question 1

Differentiate between the terms management and leadership. What are the main
functions of management?

Question 2

The main managerial roles consist of interpersonal, informational and decisional


roles respectively. Under each of the above categories of roles define the roles that a
manager plays thus explaining the function of each sub-role in turn.

Question 3

With the help of a diagram explain the managerial structure. Explain the role of
the manager at each level.

Question 4

How is the decision-making process handled at each managerial level.

Question 5

Different employees have different needs. The most effective managers know
this and have learned how to adapt their style based on the individual's needs. In this
respect, describe the following terms:

a) Surviving perspective

b) Learning perspective

c) Competing perspective

d) Relating perspective

e) Teaching perspective

Copyright © Genetic Computer School, 2002


Page199

Question 6

a) What four common functions do all managers perform? Briefly describe them.

b) What are the similarities and differences between managers at different


organisational levels?

Question 7

a) Define the team of `data and information'. To having good information what
are the steps to be taken, list and explain briefly.

b) List and describe the Characteristics of good information.

Copyright © Genetic Computer School, 2002


Page200

Chapter 6

Question 1

Write short notes on the following:

a) Knowledge representation

b) Artificial Intelligence

c) Knowledge Engineering

d) Knowledge Base

e) Expert Systems

Question 2

a) List and briefly describe the disciplines that constitute the subject of artificial
intelligence (AI).

b) List the three ways in which knowledge is represented in expert systems.

c) Do most expert systems in use today replicate the abilities of a human expert?
What do they do?

d) Explain the problems with the term artificial intelligence.

Question 3

What are the main components of human intelligence?

Question 4

Discuss whether Artificial Intelligence is possible.

Copyright © Genetic Computer School, 2002


Page201

Question 5

Define the term meta knowledge and what is its importance in the field of
artificial intelligence.

Question 6

Knowledge must be represented efficiently, and in a meaningful way. Discuss


this statement.

Question 7

What are the components of expert reasoning?

Copyright © Genetic Computer School, 2002


Page202

Chapter 7
Question 1
What does framework refer to?
Question 2
Describe the term strategic planning?
Question 3
What are the components of an IS strategy?
Question 4
What are the contents of the IS strategic plan?
Question 5
Write short notes on the following: -
a) Product differentiation
b) Competitive scope
c) Differentiation bases
Question 6
What does the term ‘hybrid manager' refer to?
Question 7
Describe the strategic planning process.
Question 8
Discuss how has the role of Information systems changed over the years?
Question 9
In Porter's view, the performance of individual corporations is determined by the
extent to which they cope with, and manipulate, the five key 'forces' which make up the
industry structure. What are those five key forces?
Question 10
Discuss how differentiation can help enterprises achieve above average
performance.

Copyright © Genetic Computer School, 2002


Page203

Chapter 8
Question 1
What are the criteria required to enhance the effectiveness of workgroups?
Question 2
What are the main drawbacks of workflow market structure?
Question 3
What are the key functions that the workflow engine must support to completely
support application requirements?
Question 4
What are the advantages that Internet provides over the workflow computing
infrastructure?
Question 5
Explain how a workgroup is connected to horizontal and vertical organizations.
Give an example of a workgroup, one horizontal and one vertical organization
connected to that workgroup, and one information flow for each connection.
Question 6
What are the advantages of workgroup computing? Question 7
What are the main limitations of transaction-based workflow systems?
Question 8
Distinguish between transaction-based, collaborative and adhoc workflow systems?
Question 9
Explain with the aid of examples what is rapid application generation.
Question 10
Explain the concept of environmental independence with respect to workflow systems.

Copyright © Genetic Computer School, 2002


Page204

Chapter 9

Question 1

How does a workgroup support system relate to a groupware?

Question 2

Distinguish between asynchronous groupware and realtime groupware. Illustrate


your answer with examples of each.

Question 3

What are the functions that groupware supports and the give some examples of
software that facilitate that support.

Question 4

Write short notes on the following: -

a) Group scheduling software

b) Electronic meeting software

c) Videoconferencing software

d) Whiteboard software

e) Workflow Automatic Software.

Illustrate your answers with examples.

Question 5

Explain the term groupware and why is it being widely used?

Question 6

What does the term CSCW (Computer-Supported Cooperative Work) refer to?

Copyright © Genetic Computer School, 2002


Page205

Question 7

Why is user testing considered more difficult in groupware systems than in


multi-user systems?

Question 8

What are the applications of groupware?

Question 9

What are the main factors that need to be taken into consideration while
designing a groupware?

Copyright © Genetic Computer School, 2002


Page206

Chapter 10
Question 1
Computation devices encode data using basically two types of codes: external and
internal code. Distinguish between these two types of codes.
Question 2
Write short notes on the following:
a) Data processing
b) Database
c) Database Management System
Question 3
In what ways is personal strategic planning beneficial?
Question 4
Why do professionals resist planning?
Question 5
Describe four categories of information systems personnel. What are the functions
of each? Which of these categories are combined in a personal information system?
Question 6
Distinguish between data and information.
Question 7
a) What tasks are involved in designing the people component of a personal
information system?
b) What is parallel installation?
c) What is pilot installation?
Question 8
Summarize how the systems development effort changes when developing
information systems at the personal and workgroup levels.

Copyright © Genetic Computer School, 2002


Page207

Chapter 11

Question 1

What is an Executive Information System? What are the features of such an


Information System?

Question 2

Describe the components of an executive Information System?

Question 3

Describe the four decision styles to which an EIS needs to be individually


tailored to.

Question 4

Explain the difference(s) in building Executive Support Systems and


Traditional Management Information Systems?

Question 5

How useful are Executive Support Systems?

Question 6

What are the main issues that an Executive Support System need to consider?

Question 7

Explain the evolution of Executive Information Systems.

Question 8

An EIS itself does not have a clear-cut goal as do most conventional


computer based information systems. Discuss this statement.

Question 9

What is the role that psychological profiling plays in Executive Information


Systems?

Copyright © Genetic Computer School, 2002


Page208

Chapter 12

Question 1

List some of the features of a decision support system.

Question 2

List some Group Decision Support Tools.

Question 3

Discuss the advantages and disadvantages of Group Decision Support Systems.

Question 4

a) Explain the basic structure and process of the model management


component of a DSS program.

b) Explain the basic structure and process of the data management component
of a DSS program.

c) Explain the relationship of the data interface program in the DSS and
extraction programs on an organizational computer.

Question 5

How can a DSS be developed in-house using commercial program components?

Question 6

a) What is a decision support system (DSS)? How does it differ from a


management information system (MIS)?

b) In what ways is building decision support systems different from building


traditional MIS systems?

Copyright © Genetic Computer School, 2002


Page209

c) Explain the two ways in which the term management information system is
used. Give the broad definition of MIS. Give the narrow definition of MIS.

d) Define the following terms:

i) Management information system and Decision support system.

Question 7

What is the main purpose of Decision Support Systems?

uestion 8

Describe the three basic components of a Decision Support System.

Copyright © Genetic Computer School, 2002


Page210

Chapter 13

Question 1

What are the components of expert systems?

Question 2

Distinguish between forward chaining and backward chaining.

Question 3

What are the main guidelines that need to be considered to determine whether a
problem is suitable for an expert solution?

Question 4

What are the five stages in the development of an expert system? In which two
ways does this process differ from the process used to develop other personal,
workgroup, and organisational information systems?

Question 5

a) List and briefly describe the disciplines that constitute the subject of artificial
intelligence (AI).

b) List the three ways in which knowledge is represented in expert systems.

c) Do most expert systems in use today replicate the abilities of a human expert?
What do they do?

d) Explain the problems in defining the term Artificial Intelligence.

Question 6

What does the term deductive inference refer to? What are its advantages?

Copyright © Genetic Computer School, 2002


Page211

Question 7

What are the limitations of expert systems?

Question 8

Distinguish between rule-based systems and expert systems.

Question 9

What are the necessary criteria for expert systems development?

Copyright © Genetic Computer School, 2002


Page212

Chapter 14

Question 1

Write short notes on the following:

a. Automatic Data Processing System

b. Distributed Information System

c. Knowledge Mining

Question 2

How does TalkMine influence the management of information in Distributed


Information Systems?

Question 3

What are the major categories of ADAPS programs? Outline the purpose of each
type of program.

Question 4

Why do files need to be maintained? How are they maintained?

Question 5

Why is a transaction-oriented system considered more secure than a data


processing system?

Question 6

How does Talkmine ensure the evolution of a distributed Information System?

Question 7

How is security maintained in Automatic Data Processing Systems?

Question 8

What are the main objectives of Distributed Information Systems?

Copyright © Genetic Computer School, 2002


Page213

Copyright © Genetic Computer School, 2002


Page214

REFERENCES
- Management Information Systems ActiveBook, 7/e
Kenneth C. Laudon, New York University
Jane P. Laudon, Azimuth Information Systems, Prentice Hall, 2002
- Essentials of Management Information Systems, 4/e
Jane P. Laudon, Azimuth Information Systems
Kenneth C. Laudon, New York University, Prentice Hall, 2001
- Developing Management Skills, 5/e
David A. Whetten, Brigham Young University
Kim S. Cameron, University of Michigan, Prentice Hall, 2002
- Management, 7/e
Stephen P. Robbins, San Diego State University
Mary Coulter, Southwest Missouri State University, Prentice Hall, 2002
- Essence of Information Systems, 2/e
Chris Edwards, Cranfield School of Mgt., Cranfield, UK John Ward, Cranfield
School of Mgt., Cranfield, UK' Andy Byetheway, Cranfield School of Mgt., Cranfield,
UK, Prentice Hall PTR,1998
- Managing Technological Change: Strategies For College And University Leaders
By: Bates, Anthony W.; Bates, Tony; Other Bates, Published: November 1999
Jossey-Bass
- Information Systems Analysis and Modeling: An Information
MacRodynamics Approach (Kluwer International Series in Engineering and
Computer Science, 2"d Edition
by Vladimir S. Lerner
Hardcover (October 1999),
Kluwer Academic Publishers; ISBN: 0792386833
- Information Modeling in the New Millennium
by Matti Rossi (Editor)
Idea Group Publishing; ISBN: 1878289772, Hardcover - 528 pages (April 2001)
- Workplace Studies : Recovering Work Practice and Informing System
Design by Paul Luff (Editor), Jon Hindmarsh, Christian Heath
Cambridge University Press, Hardcover (August 2000)

Copyright © Genetic Computer School, 2002


Page215

Copyright © Genetic Computer School, 2002

You might also like