|
|
|
|
Lesson#26
|
Evaluating and Institutionalizing Organization Development Interventions-1
|
|
|
|
Evaluating and Institutionalizing Organization Development Interventions
Measurement:
Providing useful implementation and evaluation feedback involves two
activities: selecting the appropriate variables and designing good measures.
Selecting Variables:
Ideally, the variables measured in
OD evaluation should derive from the theory or conceptual model underlying
the intervention. The model should incorporate the key features of the
intervention as well as its expected results. The general diagnostic models
described earlier meet these criteria. For example, the joblevel diagnostic
model proposes several major features of work: task variety, feedback, and
autonomy. The theory argues that high levels of these elements can be
expected to result in high levels of work quality and satisfaction. In
addition, as we shall see, the strength of this relationship varies with the
degree of employee growth need: the higher the need, the more that job
enrichment produces positive results. The job-level diagnostic model
suggests a number of measurement variables for implementation and evaluation
feedback. Whether the intervention is being implemented could be assessed by
determining how many job descriptions have been rewritten to include more
responsibility or how many organization members have received cross-training
in other job skills. Evaluation of the immediate and long- term impact of
job enrichment would include measures of employee performance and
satisfaction over time. Again, these measures would likely be included in
the initial diagnosis, when the company’s problems or areas for improvement
are discovered. Measuring both intervention and outcome variables is
necessary for implementation and evaluation feedback. Unfortunately, there
has been a tendency in OD to measure only outcome variables while neglecting
intervention variables altogether.
It generally is assumed that the
intervention has been implemented and attention, therefore, is directed to
its impact on such organizational outcomes as performance, absenteeism, and
satisfaction. As argued earlier, implementing OD interventions generally
take considerable time and learning.
It must be empirically determined that
the intervention has been implemented; it cannot simply be assumed.
Implementation feedback serves this purposes guiding the implementation
process and helping to interpret outcome data Outcome measures are ambiguous
without knowledge of how well the intervention has been implemented. For
example, a negligible change in measures of performance and satisfaction
could mean that the wrong intervention has been chosen, that the correct
intervention has not been implemented effectively, or that the wrong
variables have been measured. Measurement of the intervention variables
helps determine the correct interpretation of outcome measures. As suggested
above, the choice of intervention variables to measure should derive from
the conceptual framework underlying the OD intervention. OD research and
theory increasingly have come to identify specific organizational changes
needed to implement particular interventions. These variables should guide
not only implementation of the intervention but also choices about what
change variables to measure for evaluative purposes. The choice of what
outcome variables to measure also should be dictated by intervention theory,
which specifies the kinds of results that can be expected from particular
change programs. Again, the material in this book and elsewhere identifies
numerous outcome measures, such as job satisfaction, intrinsic motivation,
organizational commitment, absenteeism, turnover, and productivity.
Historically, OD assessment has focused on attitudinal outcomes, such as job
satisfaction, while neglecting hard measures, such as performance.
Increasingly, however, managers and researchers are calling for development
of behavioral measures of OD outcomes. Managers are interested primarily in
applying OD to change work-related behaviors that involve joining,
remaining, and producing at work, and are assessing OD more frequently in
terms of such bottom-line results.
Designing Good Measures:
Each of the measurement methods
described earlier has advantages and disadvantages. Many of these
characteristics are linked to the extent to which a measurement is
operationally defined, reliable, and valid. These assessment characteristics
are discussed below. 1.
Operational definition
. A good measure is operationally defined; that is, it specifies the
empirical data needed how they will be collected and, most important, how
they will be converted from data to information. For example, Macy and
Mirvis developed operational definitions for the behavioral outcomes (see
Table 9). They consist of specific computational rules that can be used to
construct measures for each of the behaviors. Most of the behaviors are
reported as rates adjusted for the number of employees in the organization
and for the possible incidents of behavior. These adjustments make it
possible to compare the measures across different situations and time
periods. These operational definitions should have wide
applicability across both industrial and service organizations, although
some modifications, deletions, and additions may be necessary for a
particular application. Operational definitions are extremely important in
measurement because they provide precise guidelines about what
characteristics of the situation are to be observed and how they are to he
used. They tell OD practitioners and the client system exactly how
diagnostic, intervention, and outcome variables will be measured. 2.
Reliability
. Reliability concerns the extent to which a
measure represents the “true” value of a variable; that is, how accurately
the operational definition translates data into information. For example,
there is little doubt about the accuracy of the number of cars leaving an
assembly line as a measure of plant productivity; although it is possible to
miscount, there can be a high degree of confidence in the measurement. On
the other hand, when people are asked to rate their level of job
satisfaction on a scale of 1 to 5, there is considerable room for variation
in their response. They may just have had an argument with their supervisor,
suffered an accident on the job, been rewarded for high levels of
productivity, or been given new responsibilities. Each of these events can
sway the response to the question on any given day. The individuals’ “true”
satisfaction score is difficult to discern from this one question and the
measure lacks reliability. OD practitioners can improve the reliability of
their measures in four ways. First, rigorously and operationally define the
chosen variables. Clearly specified operational definitions contribute to
reliability by explicitly describing how collected data will be converted
into information about a variable. An explicit description helps to allay
the client’s concerns about how the information was collected and coded.
Second, use multiple methods to measure a particular variable. The use of
questionnaires, interviews, observation, and unobtrusive measures can
improve reliability and result in more comprehensive understanding of the
organization. Because each method contains inherent biases, several
different methods can be used to triangulate on dimensions of organizational
problems. If the independent measures converge or show consistent results,
the dimensions or problems likely have been diagnosed accurately.’ Third,
use multiple items to measure the same variable on a questionnaire. For
example, in Job Diagnostic Survey for measuring job characteristics, the
intervention variable “autonomy” has the following operational definition:
the average of respondents’ answers to the following three questions
(measured on a seven— point scale): 1. The job permits me to decide on my
own how to go about doing the work. 2. The job denies me any chance to use
my personal initiative or judgment in carrying out the work. (Reverse
scored) 3. The job gives me considerable opportunity for independence and
freedom in how I do the work.
Table 9: Behavioral Outcomes for Measuring OD Interventions: Measures and
Computational Formulas Behavioral Outcomes for measuring OD Interventions:
Measures and Computational Formulae Behavioral Measure Computational Formula
Absenteeism rate (monthly) Σ Absence days Average workforce size x working
days Turnover rate (monthly Σ Tardiness incidents Average workforce size x
working days Internal stability rate (monthly) Σ Turnover incidents Average
workforce size Strike rate (yearly) Σ Internal movement incidents Average
workforce size Accident rate (yearly) Σ Striking Workers x Strike days
Average workforce size x working days Grievance rate (yearly) Σ of
Accidents, illnesses Total yearly hours worked
X 200,000
Σ Grievance incidents Average workforce size Σ Aggrieved individuals Average
workforce size x working days Productivity Total Output of goods or services
(units or $) Direct and/or indirect labor (hours or $) Below standard Actual
versus engineered standard Below budget Actual versus budgeted standard
Variance Actual versus budgeted variance Per employee Output/average
workforce size Quality: Total Scrap + customer returns + Rework – Recoveries
($, units or hours) Below standard Actual versus engineered standard Below
budget Actual versus budgeted standard Variance Actual versus budgeted
variance Per employee Output/average workforce size Downtime Labor ($) +
Repair costs or dollar value of replaced equipment ($) Inventory, supply and
material usage Variance (actual versus standard utilization) ($) By asking
more than one question about “autonomy,” time survey increases the accuracy
of its measurement of this variable. Statistical analyses (called
psychometric tests) are readily available for assessing the reliability of
perceptual measures, and OD practitioners should apply these methods or seek
assistance from those who can apply them.’’ Similarly, there are methods for
analyzing the content of interview and observational data, and OD evaluators
can use these methods to categorize such information so that it can be
understood and replicated. Fourth, use standardized instruments. A growing
number of standardized questionnaires are available for measuring OD
intervention and outcome variables. 3.
Validity
. Validity concerns the extent to which, a measure actually reflects the
variable it is intended to reflect. For example, the number of cars leaving
an assembly line might be a reliable measure of plant productivity but it
may not be a valid measure. The umber of cars is only one aspect of
productivity; they may have been produced at an unacceptably high cost.
Because the number of cars does not account for cost, it is not a completely
valid measure of plant productivity. OD practitioners can increase the
validity of their measures in several ways. First, ask colleagues and
clients if a proposed measure actually represents a particular variable.
This is called face validity or content validity. If experts and clients
agree that the measure reflects the variable of interest, then there is
increased confidence in the measure’s validity. Second, use multiple
measures of the same variable, as described in the section about
reliability, to make preliminary assessments of the measure’s criterion or
convergent validity. That is, if several different measures of the same
variable correlate highly with each other, especially if one or more of the
other measures have been validated in prior research, then there is
increased confidence in the measure’s validity. A special case of criterion
validity, called discriminant validity, exists when the proposed measure
does not correlate with measures that it is not supposed to correlate with.
For example, there is no good reason for daily measures of assembly—line
productivity to correlate with daily air temperature. The lack of a
correlation would be one indicator that the number of cars is measuring
productivity and not some other variable. Finally, predictive validity is
demonstrated when the variable of interest accurately forecasts another
variable over time. For example, a measure of team cohesion can be said to
be valid if it accurately predicts improvements in team performance in the
future. It is difficult, however, to establish the validity of a measure
until it has been used. To address this concern, OD practitioners should
make heavy use of content validity processes and use measures that already
have been validated. For example, presenting proposed measures to colleagues
and clients for evaluation prior to measurement has several positive
effects: it builds ownership and commitment to the data-collection process
and improves the likelihood that the client system will find the data
meaningful. Using measures that have been validated through prior research
improves confidence in the results and provides a standard that can be used
to validate any new measures used in collecting the data.
Plant: Individual:
Research Design:
In addition to measurement, OD practitioners must make choices about how to
design the evaluation to achieve valid results. The key issue is how to
design the assessment to show whether the intervention did in fact produce
the observed results. This is called internal validity. The secondary
question of whether the intervention would work similarly in other
situations is referred to as external validity. External validity is
irrelevant without first establishing an intervention’s primary
effectiveness, so internal validity is the essential minimum requirement for
assessing OD interventions. Unless managers can have confidence that the
outcomes are the result of the intervention, they have no rational basis for
making decisions about accountability and resource allocation. Assessing the
internal validity of an intervention is, in effect, testing a
hypothesis—namely, that specific organizational changes lead to certain
outcomes. Moreover, testing the validity of an intervention hypothesis means
that alternative hypotheses or explanations of the results must be rejected.
That is, to claim that an intervention is successful, it is necessary to
demonstrate that other explanations— in the form of rival hypotheses—do not
account for the observed results. For example, if a job enrichment program
appears to increase employee performance, such other possible explanations
as new technology, improved raw materials, or new employees must be
eliminated. Accounting for rival explanations is not a precise, controlled,
experimental process such as might be found in a research laboratory. OD
interventions often have a number of features that make determining whether
they produced observed results difficult. They are complex and often involve
several interrelated changes that obscure whether individual features or
combinations of features are accounting for the results. Many OD
interventions are long-term projects and take considerable time to produce
desired outcomes. The longer the time period of the change program, the
greater are the chances that other factors, such as technology improvements,
will emerge to affect the results. Finally, OD interventions almost always
are applied to existing work units rather than to randomized groups of
organization members. Ruling out alternative explanations associated with
randomly selected intervention and comparison groups is, therefore,
difficult. Given the problems inherent in assessing OD interventions,
practitioners have turned to quasiexperimental research designs. These
designs are not as rigorous and controlled as are randomized experimental
designs, but they allow evaluators to rule out many rival explanations for
OD results other than the intervention itself, Although several
quasi-experimental designs are available, those with the following three
features are particularly powerful for assessing changes: 1.
Longitudinal measurement
. This involves measuring results
repeatedly over relatively long time periods. Ideally, the data collection
should start before the change program is implemented and continue for a
period considered reasonable for producing expected results. 2.
Comparison unit
. It is always desirable to compare
results in the intervention situation with those in another situation where
no such change has taken place. Although it is never possible to get a
matching group identical to tile intervention group, most organizations
include a number of similar work units that can be used for comparison
purposes. 3.
Statistical analysis
.
Whenever possible, statistical methods should be used to rule out the
possibility that the results are caused by random error or chance. Various
statistical techniques are applicable to quasiexperimental designs, and OD
practitioners should apply these methods or seek help from those who can
apply them.
Table 10: Quasi Experimental Research Design
Quasi- Experimental Research Design Monthly Absenteeism (%) SEP. OCT. NOV.
DEC. JAN FEB MAR APR Intervention group 2.1 5.3 5.0 5.1 Start of
intervention 4.6 4.0 3.9 3.5 Comparison group 2.5 2.6 2.4 2.5 2.6 2.4 2.5
2.5 Table 10 provides an example of a quasi-experimental design having these
three features. The intervention is intended to reduce employee absenteeism.
Measures of absenteeism are taken from company monthly records for both the
intervention and comparison groups. The two groups are similar yet
geographically separate subsidiaries of a multi-plant company. Table 10
shows each plant’s monthly absenteeism rate for four consecutive months both
before and after the start of the intervention. The plant receiving the
intervention shows a marked decrease in absenteeism in the months following
the intervention, whereas the control plant shows comparable levels of
absenteeism in both time periods. Statistical analyses of these data suggest
that the abrupt downward shift in absenteeism following the intervention was
not attributable to chance variation. This research design and the data
provide relatively strong evidence that the intervention was successful.
Quasi-experimental research designs using longitudinal data, comparison
groups, and statistical analysis permit reasonable assessments of
intervention effectiveness. Repeated measures often can be collected from
company records without directly involving members of the experimental and
comparison groups. These unobtrusive measures are especially useful in OD
assessment because they do not interact with the intervention and affect the
results. More obtrusive measures, such as questionnaires and interviews, are
reactive and can sensitize people to the intervention. When this happens, it
is difficult to know whether the observed findings are the result of the
intervention, the measuring methods, or some combination of both. Multiple
measures of intervention and outcome variables should be applied to minimize
measurement and intervention interactions. For example, obtrusive measures
such as questionnaires could be used sparingly, perhaps once before and once
after the intervention. Unobtrusive measures, such as the behavioral
outcomes shown in Table 9, could be used repeatedly, thus providing a more
extensive time series than the questionnaires. When used together the two
kinds of measures should produce accurate and non-reactive evaluations of
the intervention. The use of multiple measures also is important in
assessing perceptual changes resulting from intervention. Considerable
research has identified three types of change alpha, beta, and gamma—that
occur when using self-report, perceptual measures.
Alpha Change
concerns a difference that occurs along some
relatively stable dimension of reality. This change is typically a
comparative measure before and after an intervention. For example,
comparative measures of perceived employee discretion might show an increase
after a job enrichment program. If this increase represents alpha change, it
can be assumed that the job enrichment program actually increased employee
perceptions of discretion. If comparative measures of trust among team
members showed an increase after a team-building intervention, then we might
conclude that our OD intervention had made a difference.
Beta Change:
Suppose, however, that a decrease in trust
occurred – or no change at all. One study has shown that, although no
decrease in trust occurred, neither did a measurable increase occur as a
consequence of team-building intervention. Change may have occurred,
however. The difference may be what is called a beta change. As a result of
team-building intervention, team members may view trust very differently.
Their basis for judging the nature of trust changed, rather than their
perception of a simple increase or decrease in trust along some stable
continuum. This difference is called beta change. For example,
before-and-after measures of perceived employee discretion can decrease
after a job enrichment program. If beta change is involved; it can explain
this apparent failure of the intervention to increase discretion. The first
measure of discretion may accurately reflect the individual’s belief about
the ability to move around and talk to fellow workers in the immediate work
area. During implementation of the job enrichment intervention, however, the
employee may learn that the ability to move around is not limited to the
immediate work area. At a second measurement of discretion, the employee,
using this new and recalibrated understanding, may rate the current level of
discretion as lower than before.
Gamma change
involves fundamentally redefining the
measure as a result of an OD intervention. In essence, the framework within
which a phenomenon is viewed changes. A major change in the perspective or
frame of reference occurs. Staying with the example, after the intervention
team members might conclude that trust was not a relevant variable in their
team building experience. They might believe that the gain in their clarity
and responsibility was the relevant factor and their improvement as a team
had nothing to do with trust. For example, the presence of gamma change
would make it difficult to compare measures of employee discretion taken
before and after a job enrichment program. The measure taken after the
intervention might use the same words, but they represent an entirely
different concept. The term “discretion” may originally refer to the ability
to move about the department and interact with other workers. After the
intervention, discretion might be defined in terms of the ability to make
decisions about work rules, work schedules, and productivity levels. In sum,
the job enrichment intervention changed the way discretion is perceived and
how it is evaluated. These three types of change apply to perceptual
measures. When other than alpha changes occur, interpreting measurement
changes becomes the more difficult. Potent OD interventions may produce both
beta and gamma changes, which severely complicates interpretations of
findings reporting change or no change. Further, the distinctions among the
three different types of change suggest that the heavy reliance on
questionnaires, so often cited in the literature, should be balanced by
using other measures, such as
interviews and unobtrusive records. Analytical methods have been developed
to assess the three kinds of change, anti OD practitioners should gain
familiarity with these recent techniques.
Case: The Farm Bank
The Farm Bank is one of the state’s oldest and most solid banking
institutions. Located in a regional marketing center, the bank has been
active in all phases of banking, specializing in farm loans. The bank’s
president, Frank Swain, 62, has been with the bank for many years and is
prominent in local circles. The bank is organized into six departments (as
shown in Figure below). A senior vice president heads each department. All
six of them have been with the bank for years, and in general they reflect a
stable and conservative outlook.
The Management Information System
Two years ago, President Swain felt that the bank needed to “modernize its
operations. With the approval of the board of directors, he decided to
design and install a comprehensive management information system (MIS). The
primary goal was to improve internal operations by supplying necessary
information on a more expedited basis, thereby decreasing the time necessary
to service customers. The system was also to be designed to provide economic
operating data for top management planning and decision-making. To head this
department he selected Al Hassier, 58, a solid operations manager who had
some knowledge and experience in the computer department. After the system
was designed and installed, Al hired a young woman as his assistant. Valerie
Wyatt was a young MBA with a strong systems analysis background. In addition
to bring the only woman and considerably younger than any of the other
managers at this level, Wyatt was the only MBA. In the time since the system
was installed, the MIS has printed thousands of pages of operating
information, including reports to all the vice presidents, all the branch
managers, and the president. The reports include weekly, monthly, and
quarterly summaries and include cost of productions, projected labor costs,
overhead costs, and projected earnings figures for each segment of the
bank’s operations. The MIS Survey Swain was pleased with the system but
noticed little improvement in management operations. In fact, most of the
older vice presidents were making decisions and function pretty much as they
did before the MIS was installed. Swain decided to have Wyatt conduct a
survey of the users to try to evaluate the impact and benefits of the new
system. Wyatt was glad to undertake the survey, because she had long felt
the system was too elaborate for the bank’s needs. She sent out a
questionnaire to all department heads, branch managers, and so on, inquiring
into their uses of the system. As she began to assemble the survey data, a
pattern began to emerge. In general, most of the managers were strongly in
favor of the system but felt that it should be modified. As Wyatt analyzed
the responses, several trends and important points came out: (1) 93 percent
reported that they did not regularly use the reports because the information
was not in a useful form, (2) 76 percent reported that the printouts were
hard to interpret, (3) 72 percent stated that they received more data than
they wanted, (4) 57 percent reported finding some errors and inaccuracies,
and (5) 87 percent stated that they still kept manual records because they
did not fully trust the MIS.
The Meeting
Valerie Wyatt finished her report, excitedly rushed into Al Hassler’s
office, and handed it to him. Hassler slowly scanned the report and then
said, “You’ve done a good job here, Val. But now that we have the system
operating, I don’t think we should upset the apple cart, do you? Let’s just
keep this to ourselves for the time being, and perhaps we can correct most
of these problems. I’m sure Frank wouldn’t want to hear this kind of stuff.
This system is his baby, so maybe we shouldn’t rock the boat with this
report.” Valeries returned to her office feeling uncomfortable. She wondered
what to do.
Case Analysis Form Name: ____________________________________________ I.
Problems
A. Macro 1. ____________________________________________________ 2.
____________________________________________________ B. Micro 1.
_____________________________________________________ 2.
_____________________________________________________
II. Causes
1. _____________________________________________________ 2.
_____________________________________________________
3. _____________________________________________________
III. Systems affected
1. Structural ____________________________________________ 2. Psychosocial
__________________________________________. 3. Technical
______________________________________________ 4. Managerial
_____________________________________________ 5. Goals and values
__________________________________________
IV. Alternatives
1. _________________________________________________________ 2.
_________________________________________________________ 3.
________________________________________________________
V. Recommendations
1. _________________________________________________________ 2.
__________________________________________________________ 3.
__________________________________________________________
Case Solution: The Farm Bank I. Problems A. Macro
1. Client system unprepared for change. 2. Client system unfamiliar with and
unprepared for MIS.
B. Micro
1. Top-down approach (Swain’s) excluded staff from decision and preparation
for MIS. 2. Survey should have preceded, not followed, MIS. 3. Hassler not
assertive enough to fulfill Swain’s goals by keeping Swain informed. 4.
Particulars in MIS need to be changed (limit info after determining needs,
change format, etc.). 5. Valarie Wyatt has been charged by Swain to make
survey but her boss, Hassler, has told her not to give the report to Swain.
II. Causes
1. Conservative nature of firm (and age of staff). 2. Lack of education
regarding MIS. 3. Lack of planning regarding functions MIS would perform for
managers and firm. 4. Hassler more interested in personal security than in
fulfilling purpose for which he was hired.
III. Systems affected
1. Structural - Chain of command prohibited Wyatt from improving MIS through
using results of report. 2. Technical - MIS needs new form and new
limitations. These are not being carried out. 3. Behavioral – Wyatt’s
“fulfillment” and satisfaction of job well done are restricted. Other
staff’s expectations brought on by survey are frustrated by lack of
follow-through. Swain hopes are not fulfilled. Hassler knows, somewhere, he
is not fulfilling his role. Managerial decisions companywide are not being
made in the best possible way, since information is not being managed in the
most effective way possible. 4. Managerial – Hassler is uncomfortable about
taking things up the chain. Possibly the president, Frank Swain, has
intimidated subordinates in the past. Or Hassler does not want to rock the
boat, has a “full plate”, or maybe is lazy. It is difficult to access
motives of managers. 5. Goals and values – Excellence and organization
improvement does not seem to be valued by most managers except possibly
Wyatt.
IV. Alternatives
1. Wyatt could convince Hassler it’s in his best interest to show Swain
results of survey. 2. Wyatt could go along with Hassler’s inaction. 3. Wyatt
could go around Hassler and tell Swain.
V. Recommendations
Wyatt needs to submit the report to Swain since this is the person who
assigned her to do the survey. She needs to explain tactfully to Hassler the
importance of her giving Swain the report. Once the report is sent to Swain,
The Farm Bank needs to embark on a strategy of solving the problems
identified in the survey. The approach should be an integrated one involving
the people who use the MIS with them identifying specific problems and the
steps to correct the problems. Hassler needs to be involved in making the
changes as well as Wyatt.
|
|
|
|