+ All Categories
Home > Documents > Measuring Organizational Effectiveness in Institutions of Higher Education

Measuring Organizational Effectiveness in Institutions of Higher Education

Date post: 08-Jan-2017
Category:
Upload: kim-cameron
View: 213 times
Download: 1 times
Share this document with a friend
30
Measuring Organizational Effectiveness in Institutions of Higher Education Author(s): Kim Cameron Source: Administrative Science Quarterly, Vol. 23, No. 4 (Dec., 1978), pp. 604-632 Published by: Sage Publications, Inc. on behalf of the Johnson Graduate School of Management, Cornell University Stable URL: http://www.jstor.org/stable/2392582 . Accessed: 10/06/2014 00:14 Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at . http://www.jstor.org/page/info/about/policies/terms.jsp . JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship. For more information about JSTOR, please contact [email protected]. . Sage Publications, Inc. and Johnson Graduate School of Management, Cornell University are collaborating with JSTOR to digitize, preserve and extend access to Administrative Science Quarterly. http://www.jstor.org This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AM All use subject to JSTOR Terms and Conditions
Transcript
Page 1: Measuring Organizational Effectiveness in Institutions of Higher Education

Measuring Organizational Effectiveness in Institutions of Higher EducationAuthor(s): Kim CameronSource: Administrative Science Quarterly, Vol. 23, No. 4 (Dec., 1978), pp. 604-632Published by: Sage Publications, Inc. on behalf of the Johnson Graduate School of Management,Cornell UniversityStable URL: http://www.jstor.org/stable/2392582 .

Accessed: 10/06/2014 00:14

Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at .http://www.jstor.org/page/info/about/policies/terms.jsp

.JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range ofcontent in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new formsof scholarship. For more information about JSTOR, please contact [email protected].

.

Sage Publications, Inc. and Johnson Graduate School of Management, Cornell University are collaboratingwith JSTOR to digitize, preserve and extend access to Administrative Science Quarterly.

http://www.jstor.org

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 2: Measuring Organizational Effectiveness in Institutions of Higher Education

Measuring Organiza- tional Effectiveness in Institutions of Higher Education

Kim Cameron

? 1978 by Cornell University. 0001 -8392/78/2304-0621$00.75

Special thanks is given to Richard Hackman, Bob Miles, John Kimberly, and Larry Cummings as well as to the ASQ reviewers for their helpful comments and suggestions on earlier versions of this paper. Financial support from the Richard D. Irwin Foundation is also gratefully ac- knowledged. 1

The following are some of the criticisms which have been advanced concerning the goal approach to effectiveness: (1) There is a focus on official or manage- ment goals to the exclusion of the or- ganizational member, organizational con- stituency, and societal goals (Blau and Scott, 1961; Scriven, 1967). (2) There is neglect of implicit, latent, or informal procedures and goals (Merton, 1957). (3) There is neglect of the multiple and con- tradictory nature of organizational goals (Rice, 1963). (4) Environmental influences on the organization and its goals are ig- nored (Lawrence and Lorsch, 1969). (5) Organizational goals are retrospective and serve to justify organizational action, not to direct it (Weich, 1969). (6) Organiza- tional goals change as contextual factors and organizational behavior change (Warner, 1967; Pfiffner, 1977).

December 1978, volume 23

This study examines the concept of organizational effec- tiveness in institutions of higher education. Some obsta- cles to the assessment of organizational effectiveness in higher education are discussed, namely criteria problems and the unique organizational attributes of colleges and universities, and criteria choices addressing these issues are outlined. Criteria were generated from dominant coa- lition members in six institutions, and nine dimensions of organizational effectiveness were derived. Reliability and validity of the dimensions were tested, and evidence was found for certain patterns of effectiveness across the nine dimensions.

For the past 50 years, organizational researchers have been concerned with the "effectiveness" of organizations, yet confusion persists regarding what organizational effective- ness is. It has rarely been possible to compare studies of effectiveness, since few have used common criteria for in- dicating effectiveness (Campbell, 1973; Steers, 1975), and effectiveness has been a label pinned on a wide variety of organizational phenomena from a wide variety of perspec- tives. Difficulty in empirically assessing organizational effec- tiveness has arisen because no one ultimate criterion of ef- fectiveness exists. Instead, organizations may pursue multi- ple and often contradictory goals (Warner, 1967; Perrow, 1970; Hall, 1972, 1978; Dubin, 1976), relevant effectiveness criteria may change over the life cycle of an organization (Yuchtman and Seashore, 1967; Kimberly, 1976; Miles and Cameron, 1 977), different constituencies may have particular importance at one time or with regard to certain organiza- tional aspects and not others (Friedlander and Pickle, 1968, Scott, 1977; Barney, 1978), criteria at one organizational level may not be the same as those at another organiza- tional level (Price, 1972; Weick, 1977), and the relationships among various effectiveness dimensions may be difficult to discover (Seashore, Indik, and Georgopolous, 1960; Mahoney and Weitzel, 1969; Kirchhoff, 1975). In short, or- ganizational effectiveness may be typified as being mutable (composed of different criteria at different life stages), com- prehensive (including a multiplicity of dimensions), divergent (relating to different constituencies), transpositive (altering relevant criteria when different levels of analysis are used), and complex (having nonparsimonious relationships among dimensions).

A number of excellent papers have recently been published which outline many of the inadequacies and complexities of organizational effectiveness research, especially Goodman and Pennings (1977), and which also provide helpful sugges- tions for improving research methodology. Fewer empirical studies have been reported, however, which explicitly ad- dress those issues. The purpose of this paper is to present the results of an empirical study that attempts to deal di- rectly with several of the important problems currently plaguing organizational effectiveness research.

PROBLEMS IN ASSESSING ORGANIZATIONAL EFFECTIVE N ESS

Criteria problems are the major obstacles to the empirical assessment of organizational effectiveness, and they are of

604/Administrative Science Quarterly

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 3: Measuring Organizational Effectiveness in Institutions of Higher Education

2

Criticisms of the system-resource ap- proach include the following: (1) Effi- ciency and effectiveness are not sepa- rated under this approach (Price, 1972). (2) Focusing only on inputs may have damaging effects on outputs (Scott, 1977). (3) This approach assumes that the only valuable aspects of organizations are those which aid further input acquisi- tion (Scott, 1977). (4) Only the organiza- tional directors' viewpoint is taken (Scott, 1977). (5) It is really the same as the goal model since increasing inputs is an organizational operative goal (Kirchhoff, 1977). (6) This approach is in- appropriate when considering nonprofit organizations (Molnar and Rogers, 1976). 3

Criticisms of the process model include the difficulty of monitoring organizational processes (Dornbusch and Scott, 1975), the expense of gathering data on pro- cesses (Scott, 1977), the focus on means to the neglect of ends (Campbell, 1977), and the inaccuracy of most process data. "Almost every individual instance of [process data] reporting has something wrong with it" (Haberstroh, 1965:182).

Organizational Effectiveness

two general kinds. The first relates to the selection of the type of criteria indicating effectiveness, and the second re- lates to the sources or originators of the criteria. Problems of criteria type generally focus on (1) the aspect of the or- ganization being considered, e.g., goal accomplishment, re- source acquisition, internal processes, (2) the universality or specificity of criteria, (3) the normative or descriptive charac- ter of criteria, and (4) the static or dynamic quality of criteria.

Organizational Aspects

Outputs and goal accomplishment are probably the most widely used criteria of effectiveness (Georgopolous and Tannenbaum, 1957; Etzioni, 1964; Price, 1972; Hall, 1978). Not only were the earliest approaches to effectiveness guided by a rationalistic goal model, but recent writers (Price, 1968; Campbell, 1977; Scott, 1977) have continued to ad- vocate accomplishment of goals as the defining characteris- tic of organizational effectiveness.

Others, however, have pointed out problems with specifying goal accomplishment as the criterion for effectiveness' (Merton, 1957; Blau and Scott, 1962; Rice, 1963; Scriven, 1967; Warner, 1967; Pfeffer, 1977). Consequently, alterna- tives to the goal approach have been proposed.

One alternative to the goal model - the system resource model or the natural systems approach - was introduced by Yuchtman and Seashore (1967). This approach focuses on the interaction of the organization with its environment, and defines organizational effectiveness as the ability of the organization to exploit its environment in the acquisition of scarce and valued resources. Organizational inputs and ac- quisition of resources replace goals as the primary criteria of effectiveness.2

Another approach relies on internal organizational processes as the defining characteristics of effectiveness. Steers (1977: 7), for example, stated, "One solution that at least minimizes many of the obstacles to addressing effective- ness is to view effectiveness in terms of a process instead of an end state." Similarly, Pfeffer (1977) suggested that to study organizational effectiveness, it was necessary to consider the process by which organizations articulate preferences, perceive demands, and make decisions. Or- ganizational development approaches (Beckhard, 1969), or- ganizational health models (Bennis, 1966) or Likert's (1967) system 4" are variations on the process model in that each

uses internal organizational activities or practices as the dom- inant criteria of effectiveness.3

Universality of Criteria

Georgopolous and Tannenbaum (1957), Caplow (1964), Fried- lander and Pickle (1968), Mott (1972), and Duncan (1973) are among those who suggest that effective organizations are typified largely by the same criteria (e.g., adaptivity, flexibil- ity, sense of identity, absence of strain, capacity for reality testing capacity) and that research on effectiveness should include the appropriate universal indicators. Others point out that organizations have different characteristics, goals, and constituencies, and that each organization (or each type of organization) requires a unique set of effectiveness criteria

6O5IASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 4: Measuring Organizational Effectiveness in Institutions of Higher Education

4

Miles and Cameron (1977) in their study of the U.S. tobacco industry, for exam- ple, found that one firm, R. J. Reynolds, was most effective if static criteria were used, whereas another firm, Philip Mor- ris, was most effective when dynamic criteria were considered.

(Rice, 1961; Hall, 1972; Scott, 1977). The researcher, in other words, must choose a level of specificity for criteria.

Normative/Descriptive Criteria

A related problem refers to the extent to which the re- search selects derived or prescribed criteria (Price, 1972). McGregor (1960), Argyris (1962), Bennis (1966), Likert (1967), and others have all indicated what qualities effective organizations should possess, and they approach the prob- lem of effectiveness deductively by stating that the organi- zation must meet these standards to be effective. Other writers have used a descriptive approach in which organiza- tional characteristics or criteria are described (inductively de- rived) and a priori evaluative standards are avoided (Mahoney etal., 1967, 1969, 1974; Price, 1972 ; Webb, 1974; Steers, 1977). Thompson (1967) has suggested that the difference may be typified as goals for the organization versus goals of the organization.

Dynamic/Static Nature of Criteria

A fourth problem refers to static versus dynamic variables. Most studies of organizational effectiveness include static views of inputs, processes, or outcomes (Mahoney, 1967; Seashore and Yuchtman, 1967; Negandhi and Reimann, 1973; Hall, 1978) although a few use criteria indicating changes over time (Webb, 1974; Pennings, 1975, 1976). Even when change criteria are included, however, the ap- proach is generally analogous to a blurred snapshot in which indications of movement can be detected than to a motion picture in which the criteria changes can be tracked as they occur. Research conducted by Kimberly (1976) and by Miles and Cameron (1977) are among the few examples of studies in which longitudinal data on effectiveness have been gathered and monitored over time.4

Sources of Criteria

Organizational effectiveness criteria are also likely to differ depending on whose viewpoint is taken, that is, on their sources. For example, the appropriate organizational consti- tuency, the level of analysis specified by the criteria, and the use of organizational records versus perceptual reports are all choices facing the researcher.

Constituencies. Effectiveness criteria always represent someone's values and biases, but there are conflicting opin- ions about who should determine effectiveness criteria and who should provide data for their measurement. Some in- vestigators advocate relying on major decision makers and directors, or the organization's dominant coalition, to gener- ate the criteria and to supply effectiveness information (Yuchtman and Seashore, 1967; Gross 1968; Price, 1968; Pennings and Goodman, 1977). Others suggested that these top administrators or managers have narrow and biased per- ceptions, so that a broad range of constituencies should be tapped (Pfiffner and Sherwood, 1960; Steers, 1975; Katz and Kahn, 1978). Still another group (Bass, 1952; Friedlander and Pickle, 1968; Reinhardt, 1973; Scott, 1977) points out that constituencies outside the organization are relevant for generating criteria inasmuch as derived goals (Perrow, 1961), "macroquality" criteria (Reinhardt, 1973), or information con-

606/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 5: Measuring Organizational Effectiveness in Institutions of Higher Education

Organizational Effectiveness

cerning the organization's contribution to the supersystem (Katz and Kahn, 1978) are obtained from that group. Came- ron (1 978a) and Miles (1 979) point out that various strategic constituencies exist for every organization, and that ratings from different constituencies may be more or less appropri- ate depending on the purpose of the evaluation and the domain of effectiveness.

Seashore (1976) and Scott (1977) both suggest that effec- tiveness criteria differ among separate constituencies be- cause each constituency perpetuates criteria in its own self interest. Friedlander and Pickle (1 968) and Molnar and Ro- gers (1976) found empirical evidence supporting this view.

Level of analysis. Bidwell and Kasarda (1975), Hirsch (1975), and Katz and Kahn (1978) are among those who advocate relying on the supersystem or the external organizational set to determine effectiveness criteria (they define effective- ness as the ability of the organization to adapt to, manipu- late, or fulfill expectations of the external environment); whereas writers such as Webb (1974), Scott (1977), Steers (1977), and Weick (1977) suggest that criteria should relate to the organization as a unit (they see effectiveness related to the goals, processes, or characteristics of the organization itself). Pennings and Goodman (1977) propose an approach to effectiveness which focuses on organization subunits (or- ganizational effectiveness is associated with the contribu- tions of and the coordination among subunits), and Kaufman (1960), Argyris (1962), Lawler, Hall, and Oldham (1974) and others, focus on individual performance as criteria of organi- zation effectiveness (organizational effectiveness is assumed to be indicated by individual behaviors and/or satisfaction).

Organizational records versus perceptual criteria. A third source of criteria concerns the use of organizational records instead of personal perceptions. Records are sources in which information concerning effectiveness criteria may be obtained with no direct involvement by organizational mem- bers (e.g., archival records such as organizational histories, changes in personnel, stock price changes) whereas per- sonal perceptions are criteria collected directly from organi- zational members (generally through questionnaires, inter- views, or direct observation). Campbell (1977) labeled criteria obtained from organizational records "objective criteria" and asserted that such measures are inappropriate and Ipreor- dained to fail in the end." Effectiveness criteria, according to him, should always be subjective. On the other hand, Seashore and Yuchtman (1 967) relied totally on organiza- tional records and argued that these were the most appro- priate sources. Economists have generally relied on objective sources for criteria, whereas industrial and organizational psychologists have more often used perceptions. Studies such as those done by Pennings (1975, 1976) have included both objective and perceptual indicators.

Figure 1 compares the types and sources of effectiveness criteria which were selected in 20 recent empirical studies of organizational effectiveness. Empirical studies have been plotted in the figure based on the sources used to assess criteria and the types of criteria included in the investigation. The figure points out the variety of criteria choices made by

6O7IASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 6: Measuring Organizational Effectiveness in Institutions of Higher Education

researchers, since only 9 of 43 cells contain overlapping choices. Most empirical investigations, in other words, have used sources and types of effectiveness criteria which are not comparable with other empirical investigations. Further- more, the large number of blank cells in the figure illustrates the difficulty of providing a complete picture of organiza- tional effectiveness in any one study as well as the lack of information on a large number of possible criteria types. Or- ganizational effectiveness criteria on one level of analysis, for example, may be different from criteria on other levels. Not only do the pragmatics of research constrain the types and sources of criteria that can be considered, but some choices of criteria may be more appropriate in one type of organization than in another (Molnar and Rogers, 1976).

In institutions of higher education, for example, unique or- ganizational characteristics have presented special problems

U-nssal Oga.zatio. Specific

Normate Descnptke Normative Descnptie

Dutcomes I-puts P UsOutcomes "puts P UeROtomes I-puts PUcss Outcomes "puts PU (Goa I Acc.) (Resources) (GUoa I AccR) psu cs (Goa rActc ) (Resources) (G oa I Acc.) Rsnutrs

P

Stic Dy Sta Sta Dya- Static Dyna- Stic - Stic tatic Dyna- Dyna- Dytis- DySa- Dytis- Dy ati c Dyis- Dtti yc Sats c Dyt miiStti Dyti- D t ati Dyic U U Static myc 5ta0 mic Static m~ Statc ic Stotic

Georgopolous & Pennings, Taonenbaum, 1957 17,17

MahoneRiyeta/ W 1974 1967, 1969Wb,17

L-

O E ,j ,. i < 4 W l l H Usborn

& Gorgopolous Hunt, 1974 & Mann, 1962

| - E-Ghorpade, 1968- Rushing,1974 Hall,1978

U1 U Mott, 1972 >0

Friedlander& Reiman 19& T STD Pickle, 1968 Stewart, 76

ff Mona I & :

Rogers, 1976

ip p X ~~~Yuh~an, 1967 uh ~ [

O :U Ghorpade, 1968

<

Molnor & I I I I I I I I I I Bidwel I

Rogers, 1976 ~ ~ ~ ~ ~ ~ ~ ~ ~ ~~~~~Ksad,17

Hirsc 1975THISSTUD

Figure 1. Selections of sources and types of criteria for 21 emperical studies of organizational effectiveness.

608/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 7: Measuring Organizational Effectiveness in Institutions of Higher Education

Organizational Effectiveness

for researchers in selecting and assessing criteria for organi- zational effectiveness. Choices regarding the types and sources of criteria illustrated in Figure 1 have been particu- larly difficult to make in studying these organizations, so that the characteristics of the institutions as well as problems associated with the concept of organizational effectiveness have served as obstacles to empirical assessment of effec- tiveness in colleges and universities. In fact, almost no studies have been conducted to measure organizational ef- fectiveness in institutions of higher education.

Although some instruments, such as the Educational Testing Service's Institutional Functioning Inventory (1970), Pace's College and University Environment Scales (1969), or WICHE's Management Information System materials have been widely distributed and used, none of these instruments purported to assess criteria of organizational effectiveness. Several researchers have conducted studies of quality of graduate programs (Cartter 1966, 1977; Blau and Margulis, 1973), while others have investigated objective correlates of those quality ratings (Beyer and Snipper, 1974). Still other researchers have focused on individual variables such as student achievement, teaching processes, and learning cli- mates (Astin, 1968, 1971, 1977; Feldman and Newcomb, 1969; Bowen, 1977), but colleges and universities as organi- zations were not the primary focus in these studies. Clark (1970) and Blau (1973) reported two important empirical studies of colleges as organizations, but neither was inter- ested in assessments of effectiveness per se.

Problems in Assessing Effectiveness in Higher Education Some formidable problems stand as obstacles to the selec- tion and assessment of criteria of effectiveness in institu- tions of higher education. First, it is difficult to specify con- crete, measurable goals and outcomes. Some researchers have lamented the "complexity, diffuseness, ambiguity, and changeability" and typify educational goals and outcomes (National Institute of Education, 1975), and some have suggested that without meaningful and measurable objec- tives, it is impossible to assess the effectiveness of higher education (Warner and Havens, 1968; Chickering, 1971; Hayman and Stenner, 1971). Barro (1973), for example, stated that because information on effectiveness is not usu- ally collected by colleges and universities, prospects for the evaluation of effectiveness "do not seem very good," and Hutchins (1977: 5) asserted:

The only way you can criticize a university, the only way you can appraise it, the only way you can determine whether it's good or bad or medium or indifferent, is to know what it's about, what it's supposed to be, what it's supposed to be doing. If you don't know these things, you haven't any standards of criticism ... [Univer- sities] haven't any very clear ideas of what they're doing or why. They don't even know what they are.

Second, the evaluation of institutional effectiveness engen- ders skepticism and defensiveness in the academic commu- nity. Several commentators (Dressel, 1972; Barro, 1973; Bowen, 1973) hypothesized that calls for evaluations of ef- fectiveness or institutional accountability are seen as the public trying to scrutinize and control higher education, or as the existence of defects that need to be corrected. The

609/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 8: Measuring Organizational Effectiveness in Institutions of Higher Education

implication of pressures to evaluate seems to be that free- dom to experiment and innovate, to risk failure, or to estab- lish unique quality standards is no longer the prerogative of the institution and that evaluations restrict academic freedom. Individual institutions, furthermore, tend to view themselves as having unique characteristics and goals, and as not being comparable to other institutions. Dressel's (1971: 6, 7) re- port of an administrator's position on evaluation is illustrative of the approach taken by many administrators in higher education: This evaluation will be a waste of time, for either it will demon- strate that the program is excellent or that it is defective in some sense. In the first case it is a waste of time because we already know that it's a good program, and in the second, it's a waste of time because we would not believe any evidence of weakness.

Third, the financial concerns of colleges and universities have led to research on efficiency rather than on effective- ness. Meeth (1974) suggested that the central concern of higher education in the 1970s has been how to provide quality education for less money by focusing on efficiency. Efficiency has generally been defined as the ratio of costs to some output, or as the amount of energy lost in the production of organizational output (Katz and Kahn, 1978). In higher education, efficiency has most often been measured by indicators such as costs per student, student-faculty ratios, costs per faculty member, costs per square foot, etc. (Bowen and Douglas, 1971; O'Neill, 1971; Mood eta!., 1972; Meeth, 1974; Hartmark, 1975). These criteria of effi- ciency, while being well used, are not sufficient for under- standing institutional success inasmuch as educational in- stitutions must not only demonstrate efficiency, i.e., using resources with little waste, but they must also be able to demonstrate the effective use of resources as well. Fincher (1972) pointed out that efficiency and effectiveness could not be assessed by the same criteria, and more emphasis was needed on criteria of effectiveness. Finally, even the applicability of the concept of organizational effectiveness to colleges and universities has been ques- tioned, as by writers who have applied the terms "organized anarchy" or "loosely coupled system" to colleges and uni- versities (Cohen and March, 1974; Weick, 1976). March and Olson (1976:176), for example, have suggested that organi- zations in higher education are "complex 'garbage cans' into which a striking variety of problems, solutions, and partici- pants may be dumped." Any attempt to make statements about the effectiveness of such organizations, therefore, is seen as tenuous, since the rules, goals, and choices operat- ing within these organizations are ambiguous, changing, and often not recognized.

It has been found (Cameron, 1978b), however, that institu- tions of higher education vary on a continuum from loose coupling, i.e., organized anarchies, to tight coupling i.e., structured bureaucracies. Some colleges for example, main- tain a relatively homogeneous structure and operation with many effectiveness criteria being relevant for the subsys- tems within the institution. In others, common criteria are difficult to find since subsystems are mostly autonomous.

61 O/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 9: Measuring Organizational Effectiveness in Institutions of Higher Education

5

Several sources of organizational effec- tiveness criteria proved to be of particular value, among which were Price (1968), Pace (1969), the Institutional Functioning Inventory (1970), Mott (1972), Blau (1973), Campbell (1973, 1974), Balderston (1974), Micek and Wallhaus (1974), Hartmark (1 975), the Michigan Survey Research Center Assessment Package (1975), National Institute of Education Reports (1975), and Steers (1975). 6

Student representatives were not in- cluded in the study's dominant coalition because (1) students are not generally in a position to directly influence the direc- tion and functioning of the institution; (2) they generally have more limited infor- mation about the overall institution than do other dominant coalition members; (3) they have been found in other studies not to differ significantly in their percep- tions of the institution from faculty members or administrators (Educational Testing Service, 1970); and most impor- tantly, (4) constraints on time and money prohibited a representative sample from being gathered from relevant student groups on various campuses.

Organizational Effectiveness

The problem of studying organizational effectiveness in or- ganizations which vary on the loosely coupled to tightly coupled continuum lies in identifying a core group of effec- tiveness criteria that are relevant to organizational members, applicable across subunits, and comparable across institu- tions. The criteria choices made in this study were oriented toward identifying such criteria.

CRITERIA CHOICES

Selections of Criteria

The problem of ambiguity and diffuseness of goals in col- leges and universities was addressed by focusing on organi- zational characteristics rather than on goals, since it seemed unlikely that goals or outcomes were made operational in most institutions. Both objective and perceptual criteria were obtained from some institutions of higher education, and anonymity for both institutions and individuals was guaran- teed in an attempt to reduce defensiveness and reporting bias. The study focused on the organizational level, since it has been the most neglected in research on higher educa- tion, and because it would allow for comparisons among institutions. Criteria specifically related to institutions of higher education were used instead of universal criteria ap- plicable to all types of organizations. The generality of criteria often resulting from a universalistic approach and the unique organizational features of colleges and universities made this choice seem reasonable. Since there is no prece- dent for criteria of effectiveness in institutions of higher education, this study used an inductive approach in generat- ing them rather than prescribing a priori standards. And, al- though indications of organizational change over time were sought as criteria, the study was not longitudinal, and the effectiveness indicators are best typified as static rather than dynamic. Figure 1 points out where this study falls in relation to other empirical investigations of effectiveness.

Many of the criteria used to assess organizational effective- ness were initially generated from a search of the literatures Approximately 130 variables emerged from examining this literature, and they provided a framework from which inter- views were later conducted with individuals at several col- leges and universities.

Selections of Constituency

The strategic constituency chosen to be interviewed in deriv- ing the effectiveness criteria for this study was the internal dominant coalition. The internal dominant coalition refers to representatives of the major subunits or interest groups within the college or university, who influence the direction and functioning of the organization (Thompson, 1967). In the institutions in this study, this included academic, financial, general, and student affairs administrators, deans, and heads of academic departments.6 Only formal position holders or formal representatives were included in defining the domi- nant coalition. Whereas informal leaders or charismatic per- sonalities may have an influence on organizational direction, resource allocation, or functioning, it is extremely difficult to identify who those individuals are; therefore, formal position

61 1/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 10: Measuring Organizational Effectiveness in Institutions of Higher Education

holders were relied upon as being representative. The dominant coalition was selected first because several writers (Yuchtman and Seashore, 1967; Price, 1972; Pen- nings and Goodman, 1977) argued that the organization's major decision makers or the dominant coalition should be the sources of criteria for organizational effectiveness and their measurement, since they comprised the resource al- locators, the determiners of organizational policy, and the explicators of organizational goals. Thompson (1967) suggested that the dominant coalition was the most likely group to make specific both the cause and effect relation- ships within an organization and the hierarchy of outcomes to be preferred. Furthermore, as Pennings and Goodman (1977:152) noted, because members of the dominant coali- tion served as the representatives in the bargaining process within an organization, "consensus among members of the dominant coalition can be employed as a vehicle for obtain- ing effectiveness data." Van de Ven (1977) suggested, further, that solving the wrong problem with the right methods can be avoided only if users of information about organizational effectiveness are included as sources. Mem- bers of the dominant coalition are among the major users of information about organizational effectiveness. Second, members of the dominant coalition were assumed to be a knowledgeable source about each of the organiza- tional aspects under investigation at the institutional level. The mutability, comprehensiveness, divergence, transposi- tiveness, and complexity of organizational effectiveness re- quire that a limited domain of effectiveness be specified in evaluations, or that a specific operationalization of the con- cept be determined. This domain of effectiveness is defined by the aspects of the organization being studied coupled with the level of analysis used (Cameron, 1978a). In this study, the focus was limited to institutional characteristics relating to acquisition of resources, the vitality and viablility of internal processes and practices, and organizational out- comes and emphases. The dominant coalition is likely to be a more reliable source of information for these organizational aspects than other constituencies - for example, most ex- ternal constituencies.

Selections of Institutions It was assumed that in large, diverse institutions, dominant coalition members had less college-wide information than in smaller institutions because of the size and autonomy of departments and programs. Thompson (1967) argued that dominant coalition members, as representatives in the inter- nal organizational negotiations, became exposed to organiza- tion-wide information as they functioned in their roles, and he suggested that more information was available to them when the dominant coalition was smaller. The size of the institutions included was therefore limited to those with under 1 0,000 undergraduate students, and the focus of the study was the undergraduate part of the in- stitutions. These constraints eliminated from consideration large, loosely coupled universities having many semi- autonomous professional schools from the study and helped increase the likelihood that respondents would have infor- mation related to the overall organizational level.

61 2/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 11: Measuring Organizational Effectiveness in Institutions of Higher Education

Organizational Effectiveness

METHOD

Interviews were conducted with individuals associated with a variety of institutions of higher education to ensure that the effectiveness criteria had relevance for colleges and uni- versities and that the criteria could be measured. Separate date were collected in two studies. The first study repre- sented an initial attempt to assess the reliability and validity of the effectiveness criteria through questionnaires and interviews. The second study was designed primarily to ef- fect refinements and improvements in the instruments and to improve their psychometric properties.

Institutional Sample

The first study included four colleges in New England with two more schools added in the second study. Two institu- tions were public and four were private, and their under- graduate enrollments ranged from approximately 1,000 to approximately 10,000. Two institutions were primarily com- muter schools, with the others being mostly residential; four had unionized faculties, while two did not; and one of the institutions was in a rural setting, while the other five were in or near cities with a population of over 100,000.

Interviews to Derive Dimensions

Four or five top administrators at six colleges in the north- eastern United States along with about ten faculty members were interviewed. They were usually the provost or academic vice-president, the president, the financial or ad- ministrative vice-president, the dean of student affairs, an assistant to the president or a director of institutional re- search and one or two department heads on each campus. Individuals were asked to respond to questions, including the following: 1. What organizational characteristics do effective colleges possess? 2. What is it at this institution that makes a difference in terms of its effectiveness?

3. What would have to change in order to make this institution more effective?

4. Think of an institution of higher education that you judge to be effective. What is it that makes that institution effective?

5. Of the 130 or so items generated from the literature, which ones are not relevant to the effectiveness of this school?

6. Of the 130 items, which ones are not measurable or for which are data not available?

Interviews lasted from one and one half to four hours, and special emphasis was placed on criteria relating to the or- ganizational level of analysis. For example, references to in- dividuals or to specific departments or programs were avoided; instead, criteria were sought that characterized the entire institution. Therefore, the success of the president's personal leadership style or the characteristics of a unique program in one department were not generally included, whereas the institution's orientation toward participatory de- cision making involving the faculty, or the emphasis it placed on developing community-oriented programs were. Some of the effectiveness criteria resulting from the interviews did

61 3/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 12: Measuring Organizational Effectiveness in Institutions of Higher Education

relate to aggregates of individuals, e.g., student educational satisfaction, but the focus in these criteria tended to be on the entire organization rather than on one institutional sub- unit.

Certain clusters of items became apparent as the criteria emerged from the interviews, and on an a priori, intuitive basis, nine separate groupings of criteria were formed. As a rationale for this strategy of combining criteria into dimen- sions on an intuitive basis, Campbell (1977: 23) stated, "Criterion combination quite properly is based on value judgments, and there is no algorithm or higher order truth to which we can appeal." Several alternative groupings were tried but the one used here represents the only grouping that encompassed all the effectiveness criteria generated from the interviews.

These nine dimensions represented conceptually different constructs, although they were not assumed to be indepen- dent. The nine effectiveness dimensions and the criteria they encompassed were:

1. Student educational satisfaction - criteria indicated the degree of satisfaction of students with their educational experiences at the institution. 2. Student academic development - criteria indicated the extent of academic attainment, growth, and progress of students at the institution. 3. Student career development - criteria indicated the extent of occupational development of students, and the emphasis on career development and the opportunities for career development pro- vided by the institution. 4. Student personal development - criteria indicated student de- velopment in nonacademic, noncareer oriented areas, e.g., socially, emotionally, or culturally, and the emphasis on personal develop- ment and opportunities provided by the institution for personal development. 5. Faculty and administrator employment satisfaction - criteria in- dicated satisfaction of faculty members and administrators with jobs and employment at the institution.

6. Professional development and quality of the faculty - criteria indicated the extent of professional attainment and development of the faculty, and the amount of stimulation toward professional de- velopment provided by the institution. 7. Systems openness and community interaction - criteria indi- cated the emphasis placed on interaction with, adaptation to, and service in the external environment. 8. Ability to acquire resources - criteria indicated the ability of the institution to acquire resources from the external environment, such as good students and faculty, financial support, etc.

9. Organizational health - criteria indicated benevolence, vitality, and viability in the internal processes and practices at the institution.

Instruments

Two types of instruments were developed to measure the criteria in the nine dimensions. The first was a questionnaire asking respondents to describe the extent to which their college possessed certain organizational characteristics (ef- fectiveness criteria). Questionnaire items centered mostly on ratings of organizational traits (e.g., how much emphasis

6 1 4/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 13: Measuring Organizational Effectiveness in Institutions of Higher Education

Organizational Effectiveness

was given to college-community relations?) rather than on personal feelings or affect (e.g., how do you like this school?), in order to reduce the possibility of obtaining highly intercorrelated perceptions all related to the general satisfaction of respondents. Appendix A lists the question- naire items assessing the effectiveness dimensions.

The second instrument included a set of questions designed to obtain objective data from the records of each institution. Appendix A also lists these items for the eight dimensions measured. These objective data were provided by the academic vice-president or provost, the financial vice- president, the dean of students, the director of institutional research, the director of development, or other appropriate administrators at each institution. The reason for developing both objective and perceived instruments was to provide data for testing the external validity of the dimensions, since there was no way to determine the amount of bias existing in the ratings of the dominant coalition members without such a test.

A modified form of Cattell's (1966) "marker item" procedure was used to guide the additions and refinements made to the questionnaire items for the second study. This proce- dure suggests that items be chosen which have meaning central to the concept being measured, i.e., face validity, and that overlap should occur with other criteria known to be indicators of the concept. Items were added to several of the scales, consequently, in order to make certain that the central concept indicated by the title of the effectiveness dimension was being measured. These new items were similar to Cattell's marker items. Mean within-dimension cor- relations ranged from .491 to .636 for the marker items, providing evidence that the central meanings of the dimen- sions, as specified by their titles, were being tapped.

Respondent Sample

The questionnaires were mailed, under a covering letter signed by the president or academic vice-president, to ap- proximately 75 administrators and academic department heads at each of the six institutions. Anonymity for all re- spondents and institutitions was guaranteed. Reports of the results of the study were promised to each participating institution, but respondents and institutional names were kept confidential. Respondents to the questionnaire were di- vided into five job categories: general, academic, financial, and student affairs administrators, and academic department heads. About half of the respondents were faculty members and about half were administrators. Usable questionnaires returned in the first study were 191 (70 percent); 134 (72 percent) were returned in the second study. The frequencies of returns for the five respondent categories are shown in Table 1.

Analysis

At least two different strategies were possible for analyzing the data obtained from these dominant coalition members. One was to emphasize the reliability or internal consistency of measures of the central concepts in the nine effective- ness dimensions, and the other was to ensure the inclusion

61 5/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 14: Measuring Organizational Effectiveness in Institutions of Higher Education

Table 1

Response Rates for Five Categories of Respondents in Six Institutions

Responses'g

Study 1 Study 2 Job Category N % N %

Administrators General 23 82 20 77 Academic 37 70 15 68 Financial 16 70 7 54 Student affairs 34 85 32 70

Academic department head 81 62 60 77

Total 191 70 134 72

Responses across institutions ranged from 54% to 84%.

of all variables generated by the interviews regardless of their relationships to the nine central concepts. The former strategy was adopted because, first, inasmuch as reliability is a prerequisite for validity (Nunnally, 1967; Kerlinger, 1973), it was important for the internal consistency (reliability) of the criteria to be demonstrated in order that the effective- ness dimensions could be validated. Since the questionnaire items were constructed to assess the criteria comprising the dimensions, if it was found that one of the items had low internal consistency in relation to other items thought to measure the same dimension, the item was dropped since there was no way to determine whether the variance in the item was attributable to another construct being assessed (trait variance) or to method or to random error. It was thought more important to demonstrate the reliability of the measures than to focus solely on the comprehensiveness of the criteria. This is similar to the strategies used by Mahoney (1967) and by Seashore and Yuchtman (1967) in the generation of their effectiveness criteria.

Second, it had been determined that institutional data were not available for every single criterion that emerged from the interviews. Therefore, unless a large number of question- naire items turned out to be unrelated to the nine underlying dimensions, it was appropriate for reasons of meaningful- ness and parsimony to concentrate on the nine central con- cepts indicated by the dimension titles.

RESULTS

Internal Consistency and Discriminant Validity

Eight of the questionnaire items in the first study were found to have low correlations within their own effective- ness dimension as well as with items from the other eight effectiveness dimensions. These eight items, which had an average intrascale correlation below .20, included quality of written work of students, attrition of students because of too few extracurricular activities, faculty grievances, attrition of faculty because of dissatisfaction, proportion of the budget available for professional development, work effi- ciency, and pay satisfaction. Moreover, there were no high intercorrelations among the eight items themselves; con- sequently, they were not i incue in other statistical

61 6/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 15: Measuring Organizational Effectiveness in Institutions of Higher Education

Organizational Effectiveness

Table 2

Between-Dimension and Between-item Correlations for the Nine Effectiveness Dimensions

Mean Item Correlations Dimensionst

No. of items x s.d. Inside Outside 1 2 3 4 5 6 7 8 9

t1. 4 3.78 .8 .37 .23 (.70)* 3 2.84 .7 .36 .29 (.63)

2. 3 4.79 1.0 .38 .20 .37 (.65) 5 3.79 1.0 .40 .26 .56 (.77)

3. 4 3.65 .9 .27 .14 .33 .22 (.60) 5 4.27 1.0 .33 -.01 age .05 -.20 (.71)

4. 3 4.24 .9 .63 .13 age .40 .54 .23 (.66) 4 3.23 1.2 .61 .29w .56 .39 -.02 (.86)

5. 6 3.94 1.1 .40 .21 Ad .49 .36 .31 .34 (.91) 6 4.74 1.5 .57 .30* .60 .39 .05 .36 (.89)

6. 4 4.62 .9 .31 .17 .31 .32 .25 .32 .47 (.73) 5 4.48 1.1 .50 .240 .42 .43 .02 .37 .37 (.83)

7. 5 3.52 1.2 .47 .240 .41 .33 .28 .34 .50 .46 (.90) 5 3.96 1.2 .51 .27 Ad .44 .46 .13 .55 .43 .45 (.84)

8. 5 4.79 1.0 .46 .260 .57 .56 .33 .46 .54 .42 .47 (.81) 6 4.49 1.1 .50 .330 .68 .66 -.04 .59 .58 .55 .59 (.86)

9. 15 3.79 1.0 .46 .23w .48 .28 .34 .39 .59 .41 .55 .50 (.92) 17 3.91 1.2 .40 .30 .65 .57 -.10 .52 .69 .49 .56 .69 (.93)

Significant differences between inside and outside correlations at thep<.05 level. S Significant differences between inside and outside correlations at thep<.01 level. *g

Significant differences between inside and outside correlations at thep <.001 level.

t 1. Student Educational Satisfaction; 2. Student Academic Development; 3. Student Career Development; 4. Student Personal Development; 5. Faculty and Administrator Employment Satisfaction; 6. Professional Development and Quality of the Faculty; 7. System Openness and Community Interaction; 8. Ability to Acquire Resources; 9. Organizational Health.

Numbers in parentheses are reliability coefficients.

Note: The top numbers for each dimension refer to the first study, and the bottom numbers refer to the second study.

analyses of the dimensions. Coefficient alpha was applied to test the internal consistency reliability of the effectiveness dimensions and acceptable levels of reliability were found for each of them. Nunnally (1967) suggested that for explor- atory research, a reliability of between .50 and .60 was ac- ceptable, and in the first study the lowest reliability coeffi- cient among the nine effectiveness dimensions was .601, while the highest was .928. In the second study, reliability coefficients ranged from .628 to .924. The relatively high correlations of the marker items in the second study with the appropriate effectiveness dimensions also provided some evidence for the face validity of the dimensions. The internal consistency reliability for each of the dimensions is shown in Table 2.

Factor analytic procedures also largely confirmed the exis- tence of the dimensions. Oblique, varimax, and quartimax rotations were used in both studies, and the number of factors was limited to between six and twelve to try to

61 7/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 16: Measuring Organizational Effectiveness in Institutions of Higher Education

uncover any underlying dimensions. Appendices B and C contain the factors produced by an orthogonal rotation pat- tern in which an eigenvalue of 1.0 specified the number of factors. In the first study, two of the effectiveness dimen- sions loaded on the same factor and two other dimensions split into two factors. After several questionnaire items were reworded to improve their meaningfulness and clarity for respondents, the second study produced a single factor for each of the dimensions except Student Educational Satisfac- tion, which did not load on any of the factors. Furthermore, a nine-factor rotation still did not produce a factor for this dimension.

Average within-dimensions correlations for each item were compared to the mean correlations of each item with all items outside its own effectiveness dimension as one test of the discriminant validity of the items. It was found that within-dimension mean correlations were higher than the mean outside correlations for every item except one in the first study and for all items in the second study. The single item in the first study (opportunities for personal develop- ment) was eliminated from further analysis. As Table 2 indi- cates, this finding confirmed that the dimensions were composed of items with high internal consistency and that they were distinguishable one from another. Also, after the median correlation coefficient for all items within a dimen- sion was computed, correlations between the dimension and all outside items were inspected to determine overlapping among items. The purpose was to uncover the effective- ness items that correlated highly with more than one di- mension and to determine which dimensions had overlap- ping items. Several items were slightly reworded prior to the second study as a result of this analysis in order to help clarify the conceptual differences among the effectiveness dimensions for future respondents.

The Student Educational Satisfaction dimension and the Organizational Health dimension in the second study were found to contain discriminating items, but the dimensions taken as a whole were weak in discriminant validity. Table 2 demonstrates, for example, that mean within-dimension cor- relations were not significantly higher than were correlations outside the dimension for either Student Educational Satis- faction or for Organizational Health. Whereas correlations within dimensions were higher in value for each of these two dimensions, an insignificant t-test indicated a relatively high intercorrelation between these two dimensions and others.

Between-School and Between-Job Differences

Analyses of variance were performed to determine whether the effectiveness dimensions differentiated among the schools and among the respondent groups. For the scales to be employable in assessments of effectiveness, there needed to be some significant differences among the in- stitutions. If all institutions scored the same on the nine effectiveness dimensions, the instruments would be of no use in assessing relative effectiveness in institutions of higher education. Furthermore, one method of testing con- struct validity is to demonstrate differences among groups

6 1 8/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 17: Measuring Organizational Effectiveness in Institutions of Higher Education

Organizational Effectiveness

expected to score differently on a measure (Cronbach and Meehl, 1955).

The five respondent job categories were also analyzed to determine if differences among them existed. Multivariate and univariate analysis of variance procedures were used to test for significant effects.

The results, summarized in Table 3, suggest that institutional affiliations do have a significant effect on responses for combined organizational effectiveness (MANOVAp < .001), but that the job or position held is not as important. That is, in both studies, the differences are significant among the means of the institutions but not for the five job categories. In the first study, the MANOVA F-test based on Wilks' lambda for job resulted in a significance level of p < .03 while the theta (0) value, normally a more conservative test, resulted in a significance level of greater than .05. No statis- tical significance for job resulted in the second study.

Table 3

Multivariate and Univariate Analysis of Variance for the Effectiveness Dimensions Institution Job Interaction

Institution X Job

F 0 F 0 F 0

MANOVAt 4.7600 .28100 1.50 .156 1.06 .174 19.060 .28200 1.08 .127 1.37 .228

Interaction Dimension Multiple R2 Institution Job Institution X Job

F q2 F q2 F

1. Student educational .124 6.080 .09 1.71 .04 1.29 satisfaction .478 23.97 *w .39 3.38 .30 4.34

2. Student academic .185 9.73w .13 3.39 .06 1.51 development .517 44.55X .50 .75 .12 2.50

3. Student career .159 9.18 .12 2.01 .03 .91 development .609 70.34, .60 .32 .15 1.98

4. Student personal .087 2.61 * .03 2.91 .05 2.05 development .366 12.56" .60 1.24 .22 2.53

5. Faculty and administrator .082 4.11 * .07 .62 .02 1.41 employment satisfaction 4.080 18.34O .37 1.24 .22 2.53

6. Professional development .162 9.00X .14 1.81 .03 .43 and qualityof faculty .349 15.70y .34 .20 .11 1.53

7. System openness and .229 14.73" .20 1.52 .05 1.57 community interaction .282 3.54 .15 2.04 .23 2.23

8. Abilityto acquire .207 11.53 .17 2.19 .06 1.06 resources .552 52.65y .54 .51 .14 2.18

9. Organizational health .223 13.38 .18 2.31 .05 1.00 .559 51.41 'O .52 4.79000 .35 4.8900

p<.05 00

p < .01 p<.01

p <.001 t Degrees of freedom were 27 and 476 for institution, 36 and 612 for job, and 108 and 1 199 for the interaction in the first study, and 9 and 1 1 1 for institution, 36 and 417 for job, and 36 and 417 for the interaction in the second study. Note: The top numbers for each dimension refer to the first study, and the bottom numbers refer to the second study.

61 9/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 18: Measuring Organizational Effectiveness in Institutions of Higher Education

General administrators

Academic administrators 60 -

Financial administrators * _

Student affairs administrators ...........

Department heads

55--

s:~~~~~~~~~~~~~ /

V

40 i i 1 2 3 4 5 6 7 8 9

? 0 a) > OC

.2 05 .) 00. ? X- E_

had a significa Uantefc in deemnn te pretoso

M 0 W0 0 -

-~~ (j) CE C E~Q

> > > o XX B0 ~~~ ~o -

0W 0

Figure 2. Organizational effectiveness profiles for the 5 job categories.

Using univariate ANOVA procedures for each separate effec- tiveness dimension showed that the employing institution had a significant effect in determining the perceptions of the respondents for every dimension (p < .01). The amount of variance accounted for among the dimensions by this institutional factor (pq

2I) ranged from 3 percent to 20 percent in the first study and 1 5 percent to 60 percent in the second

60/S

55--

4 5

40

I I I I ~ ~ ~ ~ ~ ~ ~~~~~~~~Insiuio I

1 2 3 4 5 6 7 8 9

t5 E ~~E E t 0 5 O 0 2.~~~~~~ C

W0

CE T a .E (

> > > ~ ~ a> 0 a) =

0>0 E 2 C 0 ,~~~~~O > 0r

C ~~~~~W

Figure 3. Organizational effectiveness profiles for the 6 institutions.

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 19: Measuring Organizational Effectiveness in Institutions of Higher Education

Organizational Effectiveness

study. On the other hand, the job or position of the respon- dent had significant effects at the p < .05 level for only two effectiveness dimensions: Student Academic Development and Student Personal Development in the first study, and Student Education Satisfaction and Organizational Health in the second study. The interaction of the school and the job category was significant (p < .05) only for Student Personal Development in the first study and for four dimensions in the second study.

A profile analysis, shown in Figure 2, also confirmed the similarity of the different job categories. Mean scores for each respondent group included in the six institutions were plotted across the nine dimensions of effectiveness and tested for differences in levels (Nunnally, 1967). According to Van de Geer (1971), the MANOVA procedure had already tested for differences in parallelism. None of the respondent group pairs differed significantly (.50) in the levels of their ratings, so that it can be concluded that the dominant coali- tion members in these institutions had similar perceptions of effectiveness.

A second profile analysis plotting institutional means on the nine effectiveness dimensions revealed that the institutions not only varied significantly in their effectiveness profiles (significant differences exist among at least two of the in- stitutions on every dimension), but that certain patterns of organizational effectiveness could be distinguished. Institu- tion 5, for example, showed high effectiveness on all the dimensions except Student Career Development and System Openness and Community Interaction. This may indicate a tendency away from occupational and community involve- ment - an external emphasis - by this institution. Institu- ness was achieved in the career and community oriented dimensions with low effectiveness scores on other dimen- sions. This occupational and community-oriented success may be somewhat surprising to the institution since the catalogues of each of all six institutions claimed a liberal arts undergraduate emphasis.

Institution 1 showed relatively high effectiveness on dimen- sions related to satisfaction and organizational morale, i.e., Student Educational Satisfaction, Faculty and Administrator Employment Satisfaction, and Organizational Health, while the academically oriented dimensions tended to be low, i.e., Student Academic Development and Professional Develop- ment and Quality of the Faculty. Institution 3, on the other hand, had relatively high effectiveness in Student Academic Development but was less effective in most other areas. Institution 2 had consistently high scores on the dimensions with the highest relative effectiveness being on the nonstudent-oriented dimensions. Institution 6 was almost exactly opposite to that pattern by being consistently low on the effectiveness dimensions but with the highest relative effectiveness being on the student oriented dimensions.

These results suggest that the institutions can be distin- guished, on the basis of their effectiveness profiles, as those having very high or very low effectiveness on external dimensions (institutions 4 and 5), those with very high or low effectiveness on morale dimensions (institutions 1 and 3),

62 1 /ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 20: Measuring Organizational Effectiveness in Institutions of Higher Education

and as those having high or low effectiveness on student oriented dimensions (institutions 2 and 6). Furthermore, whereas institutional effectiveness profiles differ signifi- cantly from one another and relative strengths and weak- nesses are evident, some institutions do achieve higher overall effectiveness than others.

The analyses of these two studies indicated that the hypothesized dimensions had acceptable reliability and that they were useful in differentiating among colleges and uni- versities for organizational effectiveness. Each institution was found to vary uniquely across the nine effectiveness dimensions, although certain patterns of effectiveness seemed to emerge. Furthermore, scores on the dimension were generally not significantly affected by different re- spondent categories.

Evidence for Validity

Supporting evidence for internal consistency and discrimi- nant validity in these studies still left questions unanswered about the external validity and construct validity of the effec- tiveness dimensions. There was a dilemma, however, in at- tempting to deal with validity. On the one hand, no generally accepted criteria exist against which to compare these per- ceptual dimensions; therefore, testing for concurrent or criterion validity was impossible. On the other hand, con- struct validity - an approach to validity used when no valid external criteria are available (Cronbach and Meehl, 1955) was similarly questionable since, as Nunnally (1 967) indi- cated, proof of construct validity comes from determining the extent to which measures of a construct fit into a net- work of expected relations. Inasmuch as organizational ef- fectiveness in institutions of higher education has never been measured, no theoretical or predictable network of re- lationships has been possible between effectiveness of col- leges and universities - particularly these nine dimensions - and other constructs. Campbell (1973) pointed out that much of the explanatory research on organizational effec- tiveness had been done using individual behavior or per- formance. Very few studies have used organizational units as degrees of freedom. Consequently, there is no well- defined nomological network for organizational effectiveness in general, let alone college and university effectiveness. This study was designed to begin the development of a network. Some indications of validity in this research project were needed, nevertheless, in order that followup research, in which explanatory data could be obtained and related to the effectiveness constructs, would prove meaningful and worthwhile. Two separate pieces of evidence were found which suggested that the effectiveness dimensions had some external and construct validity.

Objective indicators of the effectiveness dimensions had been obtained from each of the six institutions, and it was hypothesized that positive correlations between the two sets of data would provide some evidence for the external validity of the perceptual measures. Table 4 reports the nonparametric rank order correlations between the objective data and the perceptual ratings.

622/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 21: Measuring Organizational Effectiveness in Institutions of Higher Education

Organizational Effectiveness

Table 4

Rank-Order Correlations Between Objective and Perceptual Measures of the Effectiveness Dimensions

Dimension r pa

1. Student educational satisfaction .600 .10 2. Student academic development .829 .02 3. Student career development -.657 .08 4. Student personal development .771 .04 5. Faculty and administrator

employment satisfaction .314 .27 6. Professional development and

quality of the faculty .943 .002 7. System openness and community

interaction -.600 .10 8. Ability to acquire resources .714 .05 9. Organizational health No objective data collected

Moderate to high positive correlations for all but two of the effectiveness dimensions provided some support for exter- nal validity, although two of the dimensions had, unexpect- edly, negative correlations indicating that either the objec- tives measures or the perceptual measures were faulty, that different and negatively correlated concepts were being as- sessed, that the concepts being measured were not uni- dimensional and had complex relationships with each other, or that the constructs being measured in the two effective- ness dimensions were confusing to respondents.

There was no sure way to determine the reason for the negative correlations in these two studies, particularly given the small sample, but a close examination of the objective and perceptual items for the eight dimensions did suggest that two separate concepts may have been assessed. In the case of Student Career Development, objective items fo- cused on vocational counseling and work study, whereas the perceived items emphasized successful placement of stu- dents in desired post-college employment and the offering of a career oriented curriculum. The perceived items relating to Systems Openness and Community Interaction dealt mainly with community and professional activities of em- ployees, whereas the objective items focused on continuing education and extension programs. Close examination of the items also revealed, however, that other dimensions had the same problem. That is, differences in objective and per- ceptual concepts could be hypothesized for almost all of the dimensions. For example, the objective measures for the Student Academic Development dimension seemed to em- phasize continued academic attainment after leaving the in- stitution whereas the perceptual measures emphasized academic development of students within the institution. Yet, the correlation coefficient for that dimension was high and positive.

This is not an unusual difficulty when comparing objective and perceptual measures, and similar problems have been found in relation to other concepts, most notably environ- mental uncertainty (Tosi, Aldag, and Storey, 1 973; Downey, Hellreigel, and Slocum, 1975). Researchers on environmental

623/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 22: Measuring Organizational Effectiveness in Institutions of Higher Education

uncertainty have generally concluded that a choice should be made between the two types of measures and compari- sons between them avoided. The dilemma in this study was that some evidence of external validity was needed to help determine the amount of bias existing in the perceptual ratings, yet comparisons with the objective data was tenu- ous. Limited support for external validity seemed to be jus- tified for some of the dimensions since what appeared to be related concepts were being assessed by two types of measures, but no definitive conclusions can be drawn.

There is evidence that the objective measures of effective- ness in this research, furthermore, were not as reliable as would have been desirable. It was found, for example, that relatively little objective data were available on inputs, pro- cesses, and outcomes at the six institutions studied. Data were often in confidential files, in several offices, or unavail- able altogether. Answers to many of the items, con- sequently, were guesses by the responding administrator, particularly when the data were not readily available or had not been centrally compiled. This objective data gathering made it understandable, in fact, why most studies of higher education avoid multivariate objective data on effectiveness and rely instead on cost ratios.

A second indication of validity was found by comparing scores on the nine effectiveness dimensions of institutions with unionized faculties and those without a union. Figure 3 shows that the institutions with a faculty union (institutions 1, 3, 4, and 6) scored lower than each institution without a unionized faculty (institutions 2 and 5) on four of the effec- tiveness dimensions: Faculty and Administrator Employment Satisfaction, Professional Development and Quality of the Faculty, Ability to Acquire Resources, and Organizational Health. These findings are consistent with research con- ducted by Duryea et al. (1973), Hedgepeth (1974), Garbarino (1975), Kemerer and Baldridge (1975), and others, which found lower faculty satisfaction, more emphasis on collec- tive bargaining issues and less on faculty concerns, feelings of powerlessness or of being externally controlled, and less collegiality and organizational benevolence in unionized in- stitutions. In terms of construct validity, these relationships between the effectiveness dimensions and other external concepts, i.e., faculty unionism, in predictable directions pro- vides the beginnings of a nomological network that can be expanded with additional research.

CONCLUSION

Multidomain Character of Effectiveness.

Much of the lack of cumulativeness in past effectiveness research has resulted from confusion over what conceptual referent or effectiveness domain has been applied when referring to organizational effectiveness, and from the wide variety of types and sources of criteria used to indicate effectiveness. The emphasis on one best definition of organizational effectiveness that has been common in past literature has not advanced the development of studies of organizational effectiveness either theoretically or empiri- cally. While acknowledging the multidimensional character of organizational effectiveness, researchers continue to

624/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 23: Measuring Organizational Effectiveness in Institutions of Higher Education

Organizational Effectiveness

write as if a unitary concept is being considered (Hall, 1972; Mott, 1972; Child, 1974; Hannan and Freeman, 1977; Weick, 1977). In this study it is proposed that since the concept of organizational effectiveness differs with different constituencies, different levels of analysis, different aspects of the organization, and different research or evaluation pur- poses, effectiveness not only possesses multiple dimen- sions, but it is not a unitary concept. Rather it is a construct composed of multiple domains which are therefore operationalized in different ways. Effectiveness in one do- main may not necessarily relate to effectiveness in another domain. For example, maximizing the satisfaction and growth of individuals in an organization, the domain of ef- fectiveness for Argyris (1 962), Likert (1 967), Cummings (1977), and others, may be negatively related to high levels of subunit output and coordination, the domain of effective- ness for Pennings and Goodman (1977). Specifically, publish- ing a large number of research reports may be a goal indicat- ing high effectiveness to faculty members (on an individual level) while indicating low effectiveness at the subunit or organizational level (e.g., poor teaching quality, little time with students, little personal attention for students, graduate student teaching instead of professors) to legislators and parents of undergraduates.

Application of the Approach

This approach to the study of organizational effectiveness is probably most useful as a first step in approaching a fine- grained analysis of effectiveness in colleges and universities. Weick (1 974: 366) pointed out that:

We treat effects more crudely than we do causes. If we tried obsessively to discriminant subtle differences in effects, we would probably find more single-cause, single-effect relationships than we now see.

That is, one of the reasons for the lack of theoretical and methodological development in studies of organizational ef- fectiveness is the tendency of researchers to do a fine- grained analysis of causes but a coarse-grained analysis of effects.

It has been discovered that no institution operates effec- tively on all effectiveness dimensions, but that certain effec- tiveness profiles are developed in which particular dimen- sions are emphasized. No single profile is necessarily better than any other, since strategic constituencies, environmental domain, contextual factors, etc., help determine what com- bination is most appropriate for the institution. Once a pro- file of effectiveness is identified for an institution, however, a fine-grained analysis of effectiveness can then really be made. That is, once a particular college or university is found to have high effectiveness in Organizational Health and the Ability to Acquire Resources, for example, and low effec- tiveness in Student Academic Development and in faculty satisfaction, detailed examinations of the causes, correlates, and components of its strengths and weaknesses are possi- ble, whereas no such analyses can be made when general prestige rankings (Cartter, 1966) or internal efficiency ratios (Mood et al., 1 972) are relied on.

The instrument used in assessing these nine dimensions of

625/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 24: Measuring Organizational Effectiveness in Institutions of Higher Education

organizational effectiveness can be the first step in a fine- grained analysis of effectiveness on the institutional level in identifying relevant effectiveness dimensions. The instru- ment could now be developed into at least nine separate instruments in a fine-grained analysis of each of the nine dimensions in colleges and universities.

This approach to assessing organizational effectiveness also appears applicable to other types of loosely coupled organi- zations, particularly in the non-profit or public sectors. Rainey, Backoff, and Levine (1976), in reviewing differences between public and private organizations, suggested that one major difference lies in the availability of tangible, speci- fiable goals. In the private sector, goal accomplishment is more easily recognized, agreed upon, and quantifiable than in the public sector. It is suggested that by inductively deriv- ing criteria, by focusing on organizational attributes rather than operationalized goals, and by carefully selecting sources and types of criteria to indicate effectiveness, important di- mensions of effectiveness can be identified which can lead to more fine-grained analyses of public sector organizations.

REFERENCES

Argyris, Chris 1962 Interpersonal Competence and

Organizational Effectiveness. Homewood, IL: Irwin.

Astin, Alexander W. 1968 The College Environment.

Washington: American Council on Education.

1971 Predicting Academic Perform- ance in College. Riverside, NJ: Free Press.

1977 Four Critical Years. San Fran- cisco: Jossey-Bass.

Balderston, Frederick E. 1974 Managing Today's Universities.

San Francisco: Jossey-Bass.

Barney, Jay 1978 "The electronic revolution in

the watch industry: a decade of environmental changes and corporate strategies." In Robert H. Miles (ed.), Organizational Adaptation to Environment: 1-63. Working Paper No. 7, Government and Business Re- lations Series. New Haven, CT: Yale University.

Barro, Stephen M. 1973 "Toward operational accounta-

bility systems for colleges and universities." In Addresses and Proceedings: 32-47. Oakland, CA: Western College Associa- tion.

Bass, Bernard M. 1952 "Ultimate criteria of organiza-

tional worth." Personnel Psychology, 5:157-173.

Beckhard, Richard 1969 Organizational Development.

Reading, MA: Addison-Wesley.

Bennis, Warren G. 1966 "The concept of organizational

health." In Warren G. Bennis (ed.), Changing Organizations. NewYork: McGraw-Hill.

Beyer, Janice M., and Reuben Snipper 1974 "Objective versus subjective

indicators of quality in graduate education." Sociology of Edu- cation, 47: 541 -557.

Bidwell, Charles E., and John D. Kasarda 1975 "School districtorganization

and student achievement." American Sociological Review, 40: 55-70.

Blau, Peter M. 1973 The Organization of Academic

Work. New York: Wiley.

Blau, Peter M., and Rebecca Z. Margulis 1973 "America's leading profes-

sional schools." Change, 5: 21-27.

Blau, Peter M., and W. Richard Scott 1962 Formal Organizations. San

Francisco: Chandler.

Bowen, Howard R. 1973 "Holding colleges accounta-

ble." The Chronicle of Higher Education, March 12: 28.

1977 Investment in Learning, San Francisco: Jossey-Bass.

Bowen, Howard R., and Gordon K. Douglas 1971 Efficiency in Liberal Education:

A Study of Comparative In- structional Costs for Different Ways of Organizing Teaching- Learning in a Liberal Arts Col- lege. New York: McGraw-Hill.

Cameron, Kim 1978a "On the domains of organiza-

tional effectiveness." Work- ing paper, School of Business, University of Wisconsin.

1978b Organizational Effectiveness: Its Measurement and Predic- tion in Higher Education. Doc- toral dissertation, Department of Administrative Science, Yale University.

Campbell, John P. 1973 "Research into the nature of

organizational effectiveness: an endangered species?" Working paper, Department of Psychology, University of Min- nesota.

1974 "Sources of organizational indi- cators." Proceedings from the Symposium on the Utilization of Indicator Data, Institute for Social Research, University of Michigan.

1977 "On the nature of organiza- tional effectiveness." In Paul S. Goodman and Johannes M. Pennings (eds), New Perspec- tives on Organizational Effec- tiveness: 13-55. San Fran- cisco: Jossey-Bass.

Caplow, Theodore 1964 Principles of Organization. New

York: Harcourt Brace Jovanovich.

Cartter, Allan M. 1966 An Assessment of Quality in

Graduate Education. Washington: American Council on Education.

1977 "The Cartter report on the lead- ing schools of education, law, and business." Change, 9: 44-48.

626/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 25: Measuring Organizational Effectiveness in Institutions of Higher Education

Cattell, Raymond B. 1966 "The meaning and strategic

use of factor analysis." In Raymond Cattell (ed.), Hand- book of Multivariate Experi- mental Psychology: 174-243. Chicago: Rand McNally.

Ch ickeri ng, Arth u r W. 1971 "Research in action." In Paul

Dressel (ed.), The New Col- leges: 25-52. Iowa City, IA: College Testing Program and the American Association of Higher Education.

Child, John 1974 "What determines organiza-

tional performance?" Organiza- tional Dynamics (Summer): 2-18.

Clark, Burton R. 1970 The Distinctive College.

Chicago: Aldine.

Cohen, Michael, and James G. March 1974 Leadership and Ambiguity: The

American College President. New York: McGraw-Hill, Car- negie Commission for the Fu- ture of Higher Education.

Cronbach, Lee J., and Paul E. Meehl 1955 "Construct validity in psycho-

logical tests." Psychological Bulletin, 52: 281 -302.

Cummings, Larry L. 1977 "Emergence of the instrumen-

tal organization." In Paul S. Goodman and Johannes M. Pennings (eds.), New Perspec- tives on Organizational Effec- tiveness: 56-62. San Fran- cisco: Jossey-Bass.

Dornbusch, Sanford M., and William R. Scott 1975 Evaluation and the Exercise of

Authority. San Francisco: Jossey-Bass.

Downey, Kirk, Don Hellreigel, and John Slocum 1975 "Environmental uncertainty:

the concept and its operationalization." Adminis- trative Science Quarterly, 20: 613-629.

Dressel, Paul L. 1971 The New Colleges: Toward an

Appraisal. Iowa City, IA: Ameri- can College Testing Program and the American Association of Higher Education.

1972 Return to Responsibility: Con- straints on Autonomy in Higher Education. San Francisco: Jossey- Bass.

Dubin, Robert 1976 "Organizational effectiveness:

some dilemmas of perspec- tive." Organization and Ad- ministrative Sciences, 7: 7-14.

Duncan, Robert B. 1973 "Multiple decision-making

structures in adapting to en- vironmental uncertainty: the impact on organizational effec- tiveness." Human Relations, 26: 273-291.

Duryea, E. D., R. S. Fisk, and Associates 1973 Faculty Unions and Collective

Bargaining. San Francisco: Jossey-Bass.

Educational Testing Service 1970 Institutional Functioning Inven-

tory. Princeton, NJ: Educational Testing Service.

Etzioni, Amitai 1964 Modern Organizations. En-

glewood Cliffs, NJ: Prentice- Hall.

Feldman, Kenneth A., and Theodore M. Newcomb 1969 The Impact of College on

Students. San Francisco: Jossey- Bass.

Fincher, Cameron 1972 "Planning models and

paradigms in higher educa- tion." Journal of Higher Edu- cation, 43: 754-767.

Friedlander, Frank, and Hal Pickle 1968 "Components of effective-

ness in small organizations." Administrative Science Quar- terly, 13: 289-304.

Garbarino, Joseph W. 1975 Faculty Bargaining: Change

and Conflict. New York: McGraw-Hill.

Georgopolous, Basil S., and Floyd C. Mann 1962 The Community General Hos-

pital. New York: MacMillan.

Georgopolous, Basil S., and Arnold S. Tannenbaum 1957 "The study of organizational

effectiveness." American Sociological Review, 22: 534-540.

Ghorpade, Jaisingh V. 1968 Study of Relative Effective-

ness of Joint Stock and Cooperative Sugar Factories. Doctoral dissertation, Graduate School, University of California, Los Angeles.

Goodman, Paul S., and Johannes M. Pennings, editors 1977 New Perspectives on Organi-

zational Effectiveness. San Francisco: Jossey-Bass.

Gross, Edward 1968 "Universities as organizations:

a research approach." Ameri- can Sociological Review, 33: 518-544.

Haberstroh, Chadwick J. 1965 "Organizational design and

systems analysis." In James G. March (ed.), Handbook of Organizations: 1171-1211. Chicago: Rand McNally.

Hall, Richard P. 1972 Organizations: Structures and

Process. Englewood Cliffs, NJ: Prentice-Hall.

1978 "Conceptual, methodological, and moral issues in the study of organizational effective- ness." Working Paper, De- partment of Sociology, SU NY-Albany.

Hannan, Michael T., and John Freeman 1977 "Obstacles to comparative

studies." In Paul S. Goodman and Johannes M. Pennings (eds.), New Perspectives on Organizational Effectiveness: 106-131. San Francisco: Jossey-Bass.

Hartmark, Leif 1975 Accountability, Efficiency, and

Effectiveness in the State University of New York. SUNY-Albany: Comparative Development Studies Center.

Hayman, John, and Jack Stenner 1971 "Student performance." In

Darrell Bushnell (ed.), Planned Change in Education: 47-62. New York: Harcourt Brace Jovanovich.

Hedgepeth, Royster C. 1974 "Consequences of collective

bargaining in higher educa- tion." Journal of Higher Edu- cation, 45: 691-705.

Hirsch, Paul M. 1975 "Organizational effectiveness

and the institutional environ- ment." Administrative Science Quarterly, 20: 327-344.

Hutchins, Robert Maynard 1977 "Interview with Robert

Maynard Hutchins." The Chronicle of Higher Educa- tion, 14: 5.

Katz, Daniel, and Robert L. Kahn 1978 The Social Psychology of Or-

ganizations. NY: Wiley.

Kaufman, Herbert 1960 The Forest Ranger. Baltimore:

Johns Hopkins Press.

Kemerer, Frank R., and J. Victor Baldridge 1975 Unions on Campus: A Na-

tional Study of the Conse- quences of Faculty Bargain- ing. San Francisco: Jossey- Bass.

627/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 26: Measuring Organizational Effectiveness in Institutions of Higher Education

Kerlinger, Fred N. 1973 Foundations of Behavioral Re-

search. NY: Holt, Rinehart and Winston.

Kimberly, John R. 1976 "Contingencies in the creation

of organizations: an example from medical education." New Haven: School of Or- ganization and Management, Yale University.

Kirchhoff, Bruce A. 1975 "Examination of a factor

analysis as a technique for determining organizational ef- fectiveness." Proceedings: Midwest AIDS Conference, 6: 56-59.

1977 "Organizational effectiveness measurement and policy re- search." Academy of Man- agement Review, 1: 347-355.

Lawler, Edward E., Douglas T. Hall, and Greg R. Oldham 1974 "Organizational climate: rela-

tionship to organizational structure, process, and per- formance." Organizational Behavior and Human Per- formance, 11: 139-155.

Lawrence, Paul R., and Jay W. Lorsch 1969 Organization and Environment.

Homewood, IL: Irrwin.

Likert, Rensis 1967 The Human Organization.

New York: McGraw-Hill.

Mahoney, Thomas A. 1967 "Managerial perceptions of

organizational effectiveness." Administrative Science Quar- terly, 14: 357-365.

Mahoney, Thomas A., and William Weitzel 1969 "Managerial models of or-

ganizational effectiveness." Administrative Science Quar- terly, 14: 357-365.

Mahoney, Thomas A., and Peter J. Frost 1974 "The role of technology in

models of organizational ef- fectiveness." Organizational Behavior and Human Per- formance, 11: 122-138.

March, James G., and Johan P. Olsen 1976 Ambiguity and Choice in Or-

ganizations. Oslo: Uni- vesitetsforlaget.

McGregor, Douglas 1960 The Human Side of Enterprise,

NewYork: McGraw-Hill.

Meeth, Richard L. 1974 Quality Education for Less

Money. San Francisco: Jossey- Bass.

Merton, Robert K 1957 Social Theory and Social Struc-

ture. New York: Free Press.

Micek, Sidney S., and Robert A. Walihaus 1974 An Introduction to the Identifi-

cation and Uses of Higher Edu- cation Outcome Information. Boulder, CO: Western In- terstate Commission on Higher Education.

Miles, Robert H. 1979 Macro Organizational Behavior.

Santa Monica, CA: Goodyear (in press).

Miles, Robert H., and Kim Cameron 1977 Coffin Nails and Corporate

Strategies: A Quarter Century View of Organizational Adapta- tion to Environment in the U.S. Tobacco Industry. Working Paper No. 3, Business- Government Relations Series. New Haven, CT: Yale Univer- sity.

Molnar, Joseph J., and David C. Rogers 1976 "Organizational effectiveness:

an empirical comparison of the goal and system resource ap- proaches." Sociological Quar- terly, 17: 401 -413.

Mood, Alexander M., Colin Bell, Lawrence Bogard, Helen Brownlee, and Joseph J. McCloskey 1972 Papers on Efficiency in the

Management of H ig her Educa- tion. NewYork: McGraw-Hill.

Mott, Paul E. 1972 The Characteristics of Effec-

tive Organizations. New York: Harper & Row.

National Institute of Education 1975 Administration and Manage-

ment in Educational Organiza- tions: A Proposal for Research. Watsonville, CA: National Insti- tute of Education.

Negandhi, Anant, and Bernard Reimann 1973 "Task environment, decen-

tralization, and organizational effectiveness." Human Rela- tions, 26: 203-214.

Nunnally, Jum C. 1967 Psychometric Theory. New

York: McGraw-Hill.

O'Neill, June 1971 Resource Use in Higher Educa-

tion: Trends in Outputs and In- puts. New York: McGraw Hill.

Osborn, Richard N., and James C. Hunt 1974 "Environmentandorganiza-

tional effectiveness." Adminis- trative Science Quarterly, 19: 231-246.

Pace, C. R. 1969 College and University Envi-

ronment Scales: Technical Manual. Princeton, NJ: Educa- tional Testing Service.

Pennings, Johannes M. 1975 "The relevance of the

structure-contingency model for organizational effective- ness. Administrative Science Quarterly, 20: 393-410.

1976 "Dimensions of organizational influence and their effectiveness correlates." Ad- ministrative Science Quarterly, 21: 688-699.

Pennings, Johannes M., and Paul S. Goodman 1977 "Toward a workable

framework." In Paul S. Good- man and Johannes M. Pen- nings (eds.), New Perspectives on Organizational Effective- ness: 146-184. San Francisco; Jossey- Bass.

Perrow, Charles 1961 "Goals in complex organiza-

tions." American Sociological Review, 6: 854-865.

1970 Organizational Analysis: A Sociological View. Belmont, CA: Brooks/Cole.

Pfeffer, Jeffrey 1977 "Usefulness of the concept."

In Paul S. Goodman and Johan- nes M. Pennings (eds.), New Perspectives on Organizational Effectiveness: 132-143. San Francisco: Jossey-Bass.

Pfiffner, John M., and Frank P. Sherwood 1960 Administrative Organization.

Englewood Cliffs, NJ: Prentice-Hall.

Price, James L. 1968 Organizational Effectiveness:

An Inventory of Propositions. Homewood, IL: Irwin.

1972 "The study of organizational ef- fectiveness." Sociological Quarterly, 13: 3-15.

Rainey, Hal, Robert Backoff, and Charles Levine 1976 "Comparing public and private

organizations." Public Adminis- tration Review, 36: 233-244.

Reimann, Bernard C. 1974 "Dimensions of structure in ef-

fective organizations." Academy of Management Journal, 17: 693-708.

Reinhardt, Uwe E. 1973 "Proposed changes in the or-

ganization of health care deliv- ery: an overview and critique." Milbank Memorial Fund Quar- terly, 51: 169-222.

628/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 27: Measuring Organizational Effectiveness in Institutions of Higher Education

Organizational Effectiveness

Rice, A. K. 1963 The Enterprise and Its Envi-

ronment. London: Tavistock.

Rice, Charles E. 1961 "A model for the empirical

study of large social organiza- tions." General Systems Year- book, 6: 101-106.

Rushing, William 1974 "Differences in profit and

non-profit organizations." Ad- ministrative Science Quarterly, 19: 474-484.

Scott, W. Richard 1977 "Effectiveness of organiza-

tional effectiveness studies." In Paul S. Goodman and Johan- nes Pennings (eds.), New Per- spectives on Organizational Effectiveness: 63-95. San Francisco: Jossey-Bass.

Scriven, Michael 1967 "The methodology of evalua-

tion." In Ralph W. Tyler, Robert W. Gagne, and Michael Scriven (eds.), Perspectives in Cur- riculum Evaluation: 39-83. Chicago: Rand McNally.

Seashore, Stanley E. 1976 "Defining and measuring the

quality of working life." In Louis E. Davis, Albert B. Cherns, and Associates (eds.), The Quality of Working Life: 105-118. New York: Free Press.

Seashore, Stanley E., B. P. Indik, and Basil S. Georgopolous 1960 "Relationships among criteria

of job performance." Journal of Applied Psychology, 44: 195- 202.

Seashore, Stanley E., and Ephraim Yuchtma n 1967 "Factorial analysis of organiza-

tional performance." Adminis- trative Science Quarterly, 12: 377-395.

Steers, Richard M. 1975 "Problems in measurement of

organizational effectiveness." Administrative Science Quar- terly, 20: 546-558.

1977 Organizational Effectiveness: A Behavioral View. Santa Monica, CA: Goodyear.

Stewart, James H. 1976 "Factorsaccounting forgoalef-

fectiveness." Organization and Administrative Sciences, 7: 109-121.

Survey Research Center 1975 Michigan Organizational As-

sessment Package. Ann Arbor: University of Michigan.

Thompson, James D. 1967 Organizations in Action. New

York: McGraw-Hill.

Tosi, Henry, Ramon Aldag, and Ronald Storey 1973 "On the measurement of the

environment: an assessment of the Lawrence and Lorsch environmental uncertainty scales." Administrative Sci- ence Quarterly, 18: 27-36.

Van de Geer, John P. 1971 Introduction to Multivariate

Analysis. San Francisco: Freeman.

Van de Ven, Andrew 1977 "A process for organizational

assessment." Working Paper, Wharton School, University of Pennsylvania.

Warner, W. Keith 1967 "Problems in measuring the

goal attainment of voluntary or- ganizations." Journal of Adult Education, 19: 3-14.

Warner, W. Keith, and A. Eugene Havens 1968 "Goal displacement and the in-

tangibility of organizational goals." Administrative Science Quarterly, 12: 539-555.

Webb, Ronald J. 1974 "Organizational effectiveness

and the voluntary organiza- tion." Academy of Manage- ment Journal, 17: 663-677.

Weick, Karl E. 1969 The Social Psychology of Or-

ganizing. Reading, MA: Addison-Wesley.

1974 "Middle range theories of so- cial systems." Behavioral Sci- ence, 19: 357-367.

1976 "Educational organizations as loosely coupled systems." Administrative Science Quar- terly, 21: 1-19.

1977 "Re-punctuating the problem." In Paul S. Goodman and Johan- nes M. Pennings (eds.), New Perspectives on Organizational Effectiveness: 193-225. San Francisco: Jossey-Bass.

Yuchtman, Ephraim, and Stanley E. Seashore 1967 "A system resource approach

to organizational effective- ness." American Sociological Review, 32: 891-903.

629/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 28: Measuring Organizational Effectiveness in Institutions of Higher Education

APPENDIX A: Objective and perceptual items measuring the nine organizational effectiveness dimensions

Dimension Perceived Items Objective Items

1. Student educational satisfaction Manifested student dissatisfaction Number of student terminations Received student complaints Number of counseling center visits for problems Attrition resulting from dissatisfaction School spirit displayed

2. Student academic development Amount of extra work and study by students Percentage going on to graduate schools Level of academic attainment Number of library books checked out Number going on to graduate school Percentage of alumni holding graduate degrees Amount of academic development" Emphasis on outside academic activities

3. Student career development Number employed in major field Number receiving vocational or career counseling Extent to which career goals are met Number involved in work study Number of career oriented courses Number obtaining jobs of first choice Importance of career education and job attainment at sc hool g

4. Student personal development Opportunities for personal development Number of weekly extracurricular activities Nonacademic growth" Number in extramurals or intramurals Emphasis on nonacademic activities Number in student government Importance of personal development" Number in drama, music, art or dance presentations

5. Faculty and administrator employment satisfaction Faculty preference for this institution over others Number of faculty members leaving Administrator preference for this institution over others Number of administrators leaving Faculty satisfaction with employment" Percentage of faculty on policy-making boards or Administrator satisfaction with employment" committees Faculty satisfaction with the school" Administrator satisfaction with the school"

6. Professional development and quality of the faculty Faculty attendance at professional conferences Percentage of faculty earning a degree after being hired Faculty publications Percentage of budget for professional development Teaching at the cutting edge Number of new courses taught Awards received by faculty Percentage of faculty with doctorates Amount of professional development" Percentage of administrators with doctorates

7. System openness and community interaction Community service of employees Number of continuing education courses Professional activities outside the institution Numberof conferences and workshops for nonstudents Emphasis on community relations Attendance at extension courses Community programs sponsored Percentage of students with jobs in the community Adaptiveness to environment*

8. Ability to acquire resources National reputation of faculty Amount of general funds raised Drawing power for local students Previously tenured faculty hired Drawing power for national students Average student high school rank Drawing power for faculty Athletic teams placing first Drawing power for financial resources Number of transfer students Ability to acquire resources' Number of students holding outside scholarships

9. Organizational health Student-faculty relations None obtained Interdepartment relations Amount of feedback obtained Typical communication type Presence of cooperative environment Flexibility of the administration Levels of trust Amount of conflict and frustration Problem-solving styles used Items used in the second study but not in Use of talents and expertise the first. Types of supervision and control Note: Actual questions are not listed to Types and adequacy of recognition and rewards conserve space. Objective items were not Decision-making styles used unless all institutions provided data for Amount of power associated with participation them, therefore several items are not listed Equity of treatment and rewards here. The actual questionnaire and the Equanizatyiof altealtmn acomplete list of objective items are Organizational health * available from the author. Long-term planning and goal settings I ntellectual orientation *630/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 29: Measuring Organizational Effectiveness in Institutions of Higher Education

Organizational Effectiveness

APPENDIX B: Orthogonal factor analysis of the nine effectiveness dimensions - first study

Factors Dimensions Items 1 2 3 4 5 6 7 8 9 10

Student educational X11 .414 .331 .012 .208 .281 .215 .144 .112 .046 -.018 satisfaction X12 .686 .122 .109 .013 .040 .007 .277 .068 .102 .040

X13 .656 .199 .065 .100 -.188 -.072 -.027 -.066 .077 .158 X14 .419 .292 .108 .221 .385 -.065 .232 .128 .146 .138

Student academic X21 -.052 .496 .230 -.161 .084 .079 .232 .072 .020 -.067 development X22 .250 .469 .160 .001 .058 .247 -.108 .250 .030 .118

X23 -.171 .387 .093 -.115 .240 -.020 .031 .001 -.078 -.078

Student career X31 .160 .087 .472 .470 .141 .178 .001 -.082 .205 .038 development X32 .160 .011 .748 .039 -.026 .050 .163 .016 -.004 .253

X33 -.180 .175 .728 .089 .080 -.112 -.027 .079 .087 -.073 X34 .098 .073 .099 .848 .065 .064 .028 .073 .032 .073

Student personal X41 -.003 .095 .047 .069 .829 -.008 .153 -.008 .123 .098 development X43 -.028 .059 .014 .035 .818 -.017 -.056 .056 -.068 .005

Faculty and administrator X51 .018 .133 -.043 .046 -.010 .872 -.016 .071 .129 .097 employment satisfaction X52 -.025 .146 .028 .081 -.019 .872 .154 .072 -.039 .109

X53 .251 .208 .095 -.068 .124 .196 .358 .119 .234 .078 X54 .110 .349 .065 -.080 .159 .230 .472 -.231 .150 .225 X55 .237 .211 .118 .126 -.066 .111 .513 .072 .396 .058 X56 .134 .093 .097 .060 .016 .115 .661 .145 .274 .228

Professional development X61 -.021 -.101 .052 -.010 .279 .124 .118 .464 .416 .012 and quality of faculty X62 -.058 .195 .061 -.053 -.061 -.058 .146 .745 .073 -.008

X63 .300 .128 -.028 .114 .074 .184 -.056 .680 .121 .084 X64 -.180 .129 .023 .360 .127 .196 .238 .471 .034 .203

System openness and X71 .106 .006 .015 .045 .165 .112 .089 .166 .403 .039 community interaction X72 .122 .049 .194 .140 -.039 -.004 .030 .034 .533 .208

X73 .117 .113 .096 -.109 -.004 .086 .228 .075 .780 .219 X74 .111 .050 -.038 .167 .135 .163 .147 .153 .692 .099 X75 .098 -.002 .018 .213 -.089 -.055 .072 .008 .677 .235

Ability to acquire X81 .186 .513 -.048 -.075 -.022 -.097 .081 .268 .083 .107 resources X82 .280 .625 .144 .113 .058 .088 .073 -.038 .102 .304

X84 .040 .788 .061 .116 .089 .125 .183 .071 .126 .160 X83 .211 .759 -.026 .104 .051 .144 .078 .154 .111 .092 X85 -.044 .383 .017 .161 .139 -.048 -.001 .090 .639 .075

Organizational health X91 .395 .208 .187 -.133 .243 .072 .042 .092 .288 .387 X92 .196 .201 .014 -.022 .142 .079 .127 .000 .130 .142 X93 .168 .111 .089 .126 .007 .126 .113 .066 .139 .451 X94 .168 -.095 .064 .092 .087 .136 .055 .167 .077 .512 X95 .120 .155 .084 .041 .140 .006 .025 .101 .157 .336 X96 -.044 .113 -.082 .148 -.026 -.074 .202 .020 .329 .532 X97 .040 .037 .138 .138 .027 .035 .089 .136 .106 .316 X98 .024 .181 -.074 -.003 -.010 .091 .100 .135 .140 .314 X99 .058 .168 -.005 -.018 -.085 .069 .190 .073 .162 .604 X910 -.033 .212 -.083 .258 -.051 -.093 .249 -.009 .359 .349 X911 -.015 .012 -.032 -.059 .026 .054 .036 .186 .026 .735 X912 .035 .006 .105 .033 -.005 .031 -.029 .200 .215 .315 X913 .112 .094 .058 .054 .134 .100 .068 -.163 .085 .801 X914 .039 .177 .102 .121 .005 .107 .059 -.014 .214 .745 X915 .052 .053 .077 .048 .037 .050 .086 .024 .021 .529

Variance (%) 1.9 3.5 1.2 1.1 2.0 1.9 1.4 1.6 2.3 14.2 Eigenvalues 3.6 6.8 2.4 2.1 4.0 3.7 2.8 3.1 4.5 27.8

631 /ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions

Page 30: Measuring Organizational Effectiveness in Institutions of Higher Education

APPENDIX C: Orthogonal factor analysis of the nine effectiveness dimensions- second study

Factors Dimensions Items 1 2 3 4 5 6 7 8

Student educational X11 .591 .175 .222 .283 .112 .055 .069 .339 satisfaction X12 .123 -.017 .063 .350 .224 -.038 .174 .218

X14 .169 .175 .422 .341 .056 .153 .263 .218

Student academic X21 .613 -.061 .088 .005 .139 .011 .159 .153 development X22 .624 -.359 .283 .175 .152 .000 .047 .224

X23 .231 -.025 .188 -.027 .070 .095 .072 -.033 X24 .693 -.090 .272 .103 .117 .029 .234 .268 X25 .581 -.035 .066 .233 .118 .405 .029 .186

Student career X31 .017 .702 .185 .084 .122 .011 -.020 .014 development X32 -.201 .764 -.117 .003 -.088 .095 -.092 -.044

X33 -.294 .509 -.295 -.044 -.052 .237 -.188 -.228 X34 .122 .467 .186 .209 .117 -.099 .192 -.149 X35 .033 .675 .004 -.023 .016 .080 .066 -.023

Student personal X42 .267 .034 .707 .056 .096 .110 .093 .171 development X43 .035 -.028 .380 .199 .008 .308 .065 .201

X41 .182 -.125 .563 .177 .096 .348 .178 .302 X'4 4 .173 .013 .766 -.022 .132 .139 .100 .010

Faculty and administrator X51 .130 -.119 -.103 .809 .046 .108 -.032 .122 employment satisfaction X52 .069 .035 .242 .696 .039 -.016 .197 .197

X53 .058 -.029 .004 .728 .145 .133 .028 .379 X54 .178 .098 .075 .720 .007 .092 .065 .341 X55 .126 .168 .026 .590 .177 .150 .230 .481 X56 .060 .171 .227 .471 .115 -.011 .396 .368

Professional development X61 .021 .036 .220 .139 .649 .176 .016 .209 and quality of faculty X62 .155 -.032 .177 .056 .750 .022 .218 .091

X63 .273 -.022 -.062 .200 .706 .103 -.205 .213 X64 .074 .095 .202 -.006 .707 .100 .118 .113 X65 .086 -.028 -.088 .071 .728 .096 .230 .144

System openness and X71 -.164 .047 .070 .043 .175 .735 .020 -.010 community interaction X72 .204 -.029 .166 .190 .232 .551 .222 .407

X73 .211 .100 .097 .083 .072 .577 .297 .142 X74 .292 -.015 .122 .034 .136 .579 .289 .242 X75 .135 .096 .164 .202 .119 .670 -.061 .309

Ability to acquire X85 .133 -.207 .252 .151 .145 .178 .634 .253 resources X81 .344 .138 .327 .119 .274 .162 .120 .140

X82 .276 .175 -.054 .149 .162 .237 .392 .082 X83 .207 .079 .156 .176 .247 .212 .498 .323 X84 .370 .040 .130 .340 .163 .119 .432 .332 X86 .274 -.196 .307 .168 .202 .092 .649 .265

Organizational health X91 .474 -.117 .106 .218 .135 .017 .108 .432 X92 .059 .079 .079 .246 .236 .084 .174 .377 X914 .234 -.132 .080 .080 .005 -.041 -.064 .735 X915 .219 .124 -.060 .185 .139 .190 .024 .561 X912 .069 .198 -.024 .258 .096 .169 .081 .650 X913 .176 -.123 .050 .098 .014 .110 .090 .740 X93 .089 -.058 .138 .060 .179 -.069 .187 .646 X94 .135 -.055 .001 .276 .129 .049 .083 .739 X95 ..060 -.038 .093 .303 .101 .035 .021 .489 X96 ..005 -.144 .065 .026 .127 .114 .258 .741 X97 .138 -.062 .129 .314 .131 .149 .030 .662 X98 .129 -.020 .121 .447 .105 .149 -.033 .610 X99 .177 -.016 .110 .250 .131 .156 .050 .771 X910 .089 .055 .376 .199 .171 .171 .147 .611 X916 .182 .065 .350 .246 .062 .346 .221 .537

X917 .238 -.048 .515 .185 .143 .115 .208 .428 X918 .449 -.046 .207 .192 .246 .139 .285 .525

Variance (%) 3.7 2.8 2.4 2.8 5.7 3.1 1.9 34.3 Eigenvalues 2.56 1.99 1.66 1.94 4.00 2.17 1.32 24.03

632/ASQ

This content downloaded from 195.34.78.148 on Tue, 10 Jun 2014 00:14:09 AMAll use subject to JSTOR Terms and Conditions


Recommended