unniivveerrssiittyy iooff –pprre ettoorriaa eeettdd 0

61
University of Pretoria etd – Jansen van Vuuren, P (2002) 42 CHAPTER THREE CONCEPTUALISATION OF COMMUNICATION RESEARCH 3.1 INTRODUCTION The systems theory and the information gap theory provide the theoretical base for this study. These theories were discussed in Chapter Two. In Chapter Three the development and importance of communication research receives attention. The communication professional and research will be addressed. This includes measurement, reputation management and knowledge management. According to Du Plooy (1995: 1), we can describe scienc e broadly as a human enterprise carried out within a community of practitioners. The purpose of science is to study and explain human reality in an authoritative fashion, and in such a way that it acquires value and use for mankind. In the scientific study of communication, we emphasise a specific part of this reality, namely the communication process with its constituent parts (communicators, messages, recipients and the circumstances in which the process takes place). The structure and process of communication research can be broadly divided into two categories : the first deals with abstract issues like the origin of research traditions. The second category focuses on the practical dimensions of the research process, or how to conduct research (Du Plooy 1995: 2). Members of a research tradition tend to have the same view of reality : about what constitutes valid knowledge; about the way the communication process should be theoretically explained; and about the best research methods (Du Plooy 1995: 2).

Upload: others

Post on 25-Mar-2022

0 views

Category:

Documents


0 download

TRANSCRIPT

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 42

CHAPTER THREE

CONCEPTUALISATION OF COMMUNICATION RESEARCH

3.1 INTRODUCTION

The systems theory and the information gap theory provide the theoretical base

for this study. These theories were discussed in Chapter Two.

In Chapter Three the development and importance of communication research

receives attention. The communication professional and research will be

addressed. This includes measurement, reputation management and knowledge

management.

According to Du Plooy (1995: 1), we can describe scienc e broadly as a human

enterprise carried out within a community of practitioners. The purpose of

science is to study and explain human reality in an authoritative fashion, and in

such a way that it acquires value and use for mankind. In the scientific study of

communication, we emphasise a specific part of this reality, namely the

communication process with its constituent parts (communicators, messages,

recipients and the circumstances in which the process takes place).

The structure and process of communication research can be broadly divided

into two categories: the first deals with abstract issues like the origin of

research traditions. The second category focuses on the practical dimensions

of the research process, or how to conduct research (Du Plooy 1995: 2).

Members of a research tradition tend to have the same view of reality: about

what constitutes valid knowledge; about the way the communication process

should be theoretically explained; and about the best research methods (Du

Plooy 1995: 2).

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 43

3.2 POSITIVIST RESEARCH

The positivist research tradition in communication (sometimes called empirical or

quantitative research) originates in the approach to, or ideology of, science

known as positivism. Positivism originated in the 17th century Europe. By the

early 19th century this approach and ideology of science displayed three key

components:

• A belief that valid knowledge could only be gained from observable

(empirical) evidence. In this process the researcher had to be objective.

• A belief that researchers should strive towards methodological unity between

the social sciences and natural sciences, by applying the methods of the

natural sciences (like laboratory experiments) on the social sciences.

• A belief in the progress of human reason and the need to utilise the social

sciences to establish a new social order (a better world). The latter was also

referred to as the positivist ideal (Mouton 1996).

Positivist media research assumes an empirical theory of knowledge .

Developed by a succession of British philosophers, empiricism views experience

(observation) as the only source of real knowledge. To achieve the positivist

ideal of methodological unity with the natural sciences, empirical communication

research uses the scientific method. This method differs from other methods of

obtaining knowledge because it is based on observation and the testing of

assumptions (hypotheses) against the evidence of the “real world”.

3.3 THE CRITICAL RESEARCH TRADITION

While the positivist research tradition in communication developed in the United

States the critical research tradition started in Germany. By the turn of the

century, Germany – like the United States – was in a state of flux and upheaval.

Social critique and ideological power in particular typified the philosophical and

political thinking of the day. This was because of change and transformation in

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 44

the country brought about, mainly by industrialisation. These changes led to

political and economical inequality. To chart the development of this modern but

dislocated society, intellectuals relied mostly on Karl Marx’s theory of historical

materialism (Du Plooy 1995: 13).

As in the United States, people in Germany were unhappy with the destruction of

the old agrarian community and its replacement by the new mass society. Thus

when the Frankfurt School of Social Research started in 1923, it represented an

effort to incorporate the critique against mass society into a Marxist framework.

The academics at the School formed their own variant of Western Marxism,

which we call critical theory (Du Plooy 1995: 13).

The School’s idea of reality was informed by Marx. Marx himself was influenced

by the ideas of the German philosopher, Georg Hegel (1770-1831). Hegel

contended that social reality had a dialectical nature.

3.4 COMMUNICATION PRACTITIONERS AND RESEARCH

There has been a substantial amount of research by academics into the way in

which practitioners use research. Broom & Dozier (1990) drew a distinction

between two extremes of public relations practice at either end of a continuum.

At one extreme the practice is “intuitive”, and personal, with little use for social

and behavioural research. At the other end scientific research is conducted and

the practice is “objective and rigorous,” based on application of empirical

knowledge and theory (Puchan, Pieczka & L’Etang 1999: 168).

In this chapter, the positioning and the current status of communication research

and environmental scanning will be investigated. Both communication research

and environmental scanning, a research method, experienced notable growth

during the past years.

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 45

During the last two decades, rapid changes in the environment seriously

threatened many organisations, resulting in radical re-engineering, restructuring

or transformation for many. Management increasingly realises that it is

imperative to understand the organisation’s environment and change along with

it, in order to survive and prosper. They have come to understand the

importance of monitoring changes that are taking place, since an organisation

can only respond to those parts of the environment of which it is aware.

There is thus increasing pressure to make greater investments in personnel and

systems for gathering environmental information, to reduce some of the

uncertainty and risk encountered in strategic decision making. Technological

innovations are making it easier to acquire environmental information. However,

effective systems should be developed to acquire relevant environmental

information in a timely fashion.

In the current academic literature, there is a strong bias in favour of scientific

positivist research. Dozier (1990) contrasts various types of public relations

practitioners in his research into roles in which he identifies “the creative artistic

practitioner (who wants) more say in decisions but not at the expense of their

spontaneity and emotional involvement in the public relations process”. The

implication of Dozier’s research is that “creative artistic practitioners” who do not

use or understand scientific research are likely to be limited to a technician role

(Puchan, Pieczka & L’Etang 1999: 168).

3.5 THE MEASUREMENT OF RELATIONSHIPS IN CORPORATE

COMMUNICATION

3.5.1 Why is it important to measure relationships in public relations

(corporate communication)?

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 46

A growing number of corporate communication practitioners and scholars have

come to believe that the fundamental goal of corporate communication is to build

and then enhance on-going or long-term relationships with an organisation’s key

constituencies.

Techniques for measuring and evaluating the relatively short-term outputs and

outcomes of specific programmes, events and campaigns have existed for quite

a number of years. But up until now, measuring the success or failure of long-

term relationships stemming, in part from public relations efforts, have not

existed (Grunig & Hon 1999: 4).

Outputs are usually the immediate results of a particular public relations

programme or activity. They measure how well an organisation presents itself to

others and the amount of attention or exposure that the organisation receives.

Outcomes measure whether target audience groups actually received the

messages directed at them …paid attention to them …understood the messages

…and retained those messages in any shape or form (Grunig & Huang 2000:

23).

They also measure whether the communication materials and messages that

were disseminated have resulted in any opinion, attitude, and/or behaviour

changes on the part of those targeted publics to whom the messages were

directed (Grunig & Hon 1999: 4).

As important as it can be for an organisation to measure public relations outputs

and outcomes, it is even more important for an organisation to measure

relationships. This is because for most organisations measuring outputs and

outcomes can only give information about the effectiveness of a particular or

specific public relations programme or event that has been undertaken (Grunig &

Hon 1999: 4).

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 47

3.5.2 Why are successful relationships important to public relations?

For at least 25 years, public relations scholars have asked two fundamental

questions? “How do you measure the effects of public relations?” and “How do

you show the value of public relations to an organisation and to society?”

Communication researchers have known how to measure several effects of

public relations for many years. Nevertheless, they know how to evaluate the

effects of public relations techniques and programmes better than they know how

to measure the value of public relations to an organisation and to society (Grunig

& Hon 1999: 8).

Measures of the effects of public relations techniques and programmes indicate

whether they have achieved their communication objectives, but they fall short of

being able to measure the value of public relations to an organisation or society.

It is possible, for example, that a public relations programme could be based on

poor strategic thinking and change the cognitions, attitudes and behaviour of a

public that has little impact on the organisation (Grunig & Hon 1999: 8).

3.5.3 The value of public relations is in relationships

In the research project on excellence in public relations and communication

management conducted for the International Association of Business

Communicators’ Research Foundation, researchers searched the literature on

organisational effectiveness for ideas that could explain the value of public

relations.

They believed it was necessary to understand what it means for an organisation

to be effective before they could explain how public relations makes it more

effective. The search of the literature on organisational effectiveness revealed

that effective organisations achieve their goals. Effective organisations choose

and achieve appropriate goals because they develop relationships with their

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 48

constituencies, which public relations practitioners typically call publics.

Ineffective organisations cannot achieve their goals, at least in part, because

their publics do not support and typically oppose management efforts to achieve

what publics consider illegitimate goals (Grunig & Hon 1999: 8).

Public opposition to management goals and decisions frequently results in

“issues” and “crises”. As a result, the process of developing and maintaining

relationships with strategic publics is a crucial component of strategic

management, issues management, and crises management.

The process of incorporating the goals, interests and concerns of publics into the

strategic decision processes of organisations is never easy, because

organisations generally encounter multiple publics with multiple goals.

Public relations makes an organisation more effective, when it identifies the most

strategic publics as part of strategic management processes and conducts

communication programmes to develop and maintain effective long-term

relationships between management and those publics.

As a result, we should be able to determine the value of public relations by

measuring the quality of relationships with strategic publics.

3.5.4 What contribution does achieving short-term communication

objectives make to the building of long-term relationships?

Grunig & Hon (1999: 8) says strategic public relations consists of the

identification of the most strategic publics with which an organisation needs to

develop a relationship; planning, implementing and evaluating communication

programmes to build relationships with these publics and measuring and

evaluating the long-term relationships between the organisation and these

strategic publics.

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 49

They also say that knowledge of how to evaluate public relations largely is limited

to the second stage: we know how to determine the effects of specific

communication programmes on the cognitions, attitudes and behaviours of

publics in the short term. There is a link, however, between short- and long term

outcomes of public relations.

The International Association of Business Communicators Excellence study

provided evidence that there is a correlation between achieving short-term

communication effects and maintaining quality long-term relationships. The

research team classified public relations departments as excellent when the chief

executive officer’s of their client organisations assigned a high value to the

contribution of the department. The chief executive officer said they valued these

departments because of their ability to maintain relationships with key

stakeholders. The senior communicators in the excellent departments also

reported more often than those in less-excellent departments that their

programmes had “change-of relationship” effects such as changes in behaviour

of a public, greater cooperation between the organisation and the development of

a stable, long-term relationship.

3.6 LERBINGER’S CLASSIFICATION OF CORPORATE COMMUNICATION

In relating research to the development of a corporate communication strategy

(and plans), a very useful definition of the kinds of corporate communication

research is provided by Lerbinger (1977: 11).

3.6.1 Environmental monitoring

Environmental monitoring or scanning is research to detect trends in

stakeholders’ opinions and in the social-political, economic, technological,

ecological and legal environment. It is used to keep track of “what is going on

out there”. Environmental monitoring is the mainstay of corporate

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 50

communication and is often referred to as “assessing the corporate climate”

(Lerbinger 1977: 12).

3.6.2 Social auditing

Social auditing is research similar to environmental monitoring. Social audits

determine the effects the organisation has had on its stakeholders and the extent

to which those effects must be corrected. The primary purpose is to examine,

catalogue, systemise and measure the organisation’s performance as a

corporate citizen (Lerbinger 1977: 12).

3.6.3 Corporate communication auditing

Public relations auditing is research to define stakeholders, and to determine how

they perceive and evaluate the organisation. There are two basic types of audits:

audience identification and corporate image studies.

Audience identification:

• Identifies relevant stakeholders

• Evaluates the organisation’s standing with each relevant stakeholder

• Identifies issues of concern to those stakeholders

• Measures the power of each stakeholder.

Corporate image studies are an extension of the corporate communication audit

insofar as they determine:

• The familiarity of each stakeholder with the organisation

• The attitudes of each stakeholder toward the organisation

• The personality characteristics each stakeholder associates with the

organisation (Lerbinger 1977: 14).

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 51

3.6.4 Communication content auditing

This is research to evaluate corporate communication programmes or plans to

find out whether messages have actually reached the target audience. It

includes readership surveys, content analysis of messages, and the

measurement of the readability of messages (Lerbinger 1977: 15).

Research is generally divided into being quantitative and qualitative. Quantitative

projects are aimed at quantifying the number of people who have a specific

attitude or problem with the company. Qualitative research is where one is trying

to get a deeper understanding of the issues.

3.7. QUALITATIVE AND QUANTITATIVE RESEARCH

Qualitative research is broader than quantitative research and based more on

observation. Quantitative research provides statistical information. Used

together, both types of research complement each other (White 1998:15).

3.7.1 Qualitative research

Qualitative research aims at the development of theories (grounded theory) and

understanding. The objective of qualitative research is to promote better self-

understanding and increase insight into the human condition. Unlike quantitative

researchers, qualitative researchers do not regard themselves as collectors of

“facts” about human behaviour that will lead to verification and the extension of

theories and enable researchers to determine causes of and predict human

behaviour. In qualitative research the emphasis is on improved understanding of

human behaviour and experience (Garbers 1996: 283).

These researchers try to understand the ways in which different individuals make

sense of their lives and to describe those meanings. Empirical observation is

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 52

prominent, because researchers need to study actual cases of human behaviour

if they are to be in a position to reflect on the human condition with more

meaning and clarity (Garbers 1996: 283).

Qualitative research is mostly done in groups – for instance a number of people

(about 8) is brought to a venue and asked to discuss the issues being

researched. These venues are mostly equipped with a one-way mirror and

sound equipment for recording for the client to review the process. A

professional person will guide the group to ensure that they remain focussed on

the subject, that some people do not dominate, that they move through the

issues in an orderly fashion, and that interesting points raised by the participants

are pursued (Du Plessis 1994: 3).

Qualitative methodology includes direct observation, an overview of different

documents and artefacts, participant observation and open-ended, unstructured

interviewing. Researchers are led by an evolving and flexible design (Garbers

1996: 283).

“Qualitative” is an umbrella term for research based on different theoretical

orientations. Prospective qualitative researchers should familiarise themselves

with, inter alia,

• The phenomenological approach whereby researchers strive to understand

the meaning of events and interactions to ordinary people in specific

situations;

• Symbolic interaction, which is based on the assumption that objects, people,

situations and events do not have inherent meaning – meaning is attributed to

them and this process involves interpretation;

• Ethnomethodology, which refers to the study of how individuals create and

understand their daily lives; and

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 53

• Cultural studies , which are embedded in conceptual frameworks like neo-

Marxism, feminist materialism and feminist post-structuralism (Garbers 1996:

284).

The most important person in this process is the moderator. This person has to

know how to get the group to participate; who has to keep them going; who has

to get the shy respondents involved; who has to ensure that people remain

interested in the proceedings; who has to make sure that time is spent on

interesting topics when these come up spontaneously; who has to ensure that

the group moves onto new issues when an issue has been exploited; etc (Du

Plessis 1994: 3).

Generally the moderator does not start to put the case for the company - i.e.

when people have views the company would feel to be "wrong" the moderator

does not get into explaining why their views are wrong - which is often failing

when the moderator is attached to the company or its public relations

practitioner.

The strength of group discussions is using the inter group dynamics that occur in

any social occasion. People express views, raise new issues, argue differences

et cetera. But this is also the weakness of a group. In such a setting people's

views are influenced by the views of others, and the way in which other people

express views (Du Plessis 1994: 3).

The presentation of the dichotomous “formal” (quantitative) and “informal”

(qualitative) framework does not make the point that good or bad research may

be either quantitative or qualitative. The emphasis tends to be on quantifiable

outcomes (Puchan, Pieczka & L’Etang 1999: 168).

What are the characteristics of qualitative research?

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 54

• Words. Qualitative research focuses on words rather than numbers.

• Researcher involvement. The main research “instrument” in qualitative

research is the researcher who closely engages with the people being

studied.

• Participant viewpoints. A desire to explore and present the various

subjective perspectives of participants is associated with qualitative

research.

• Small-scale studies. Qualitative researchers are interested in deep

exploration in order to provide rich, detailed, holistic description – as well as

explanation.

• Holistic focus. Rather than directing their attention to one or two isolated

variables, qualitative researchers tend to be oriented to a wide range of

interconnected activities, experiences, beliefs etc.

• Flexibility: Although researchers have a topic and an agenda which fuel

their research progress, they are usually committed to exploring new and

often surprising avenues that emerge as informants reveal their

understandings and interests.

• Processual. Qualitative research rarely provides static portraits of

phenomena. Instead it aims to capture processes that take place over time.

• Natural settings. On the whole, qualitative investigations are carried out in

people’s natural environments such as in their offices or where they shop.

• Inductive then deductive. Qualitative research tends to start out with

inductive reasoning and then, through a sequential process, employs

deductive reasoning (Daymon & Holloway 2002: 7).

3.7.2 Quantitative (empirical) research

Quantitative researchers seek explanations and predictions that will generalise to

other persons and places. The intent is to establish, confirm, or validate

relationships and to develop generalisations that contribute to theory.

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 55

Some research data are in numbers. In fact, quantitative research methods are

inquiries in which observations are expressed predominantly in numerical terms.

Quantitative research has two major branches, surveys (of all varieties) and

experiments (Reinard 2001: 8).

Quantitative research represents the mainstream approach to research and

carefully structured guidelines exist for conducting it. Concepts, variables and

hypotheses tend to be defined before the study begins and remain fixed

throughout.

Researchers must remain detached from their “subjects” in order to make

unbiased, universal, context-free generalisations.

Quantitative researchers typically reduce their data to numbers, which they then

present as the results of statistical tests.

The advantage of quantitative research is that people are interviewed

individually, and they all answer the same questionnaire. Thus when they are

asked a question, they give their own opinion - unbiased by answers they have

just heard someone else give. From a quantitative study the client can read that

a certain percentage of a specific stakeholder-group has a specific view about

him (Du Plessis 1994: 4).

3.8 INFORMAL REVIEWS

Broadly, such practical assignments will fall into the subjective or objective

categories. Certainly if the goals for the public relations performance were

originally set only as aims, then there is not much point in taking the assessment

behind the subjective stage. Though, even this can be valuable.

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 56

One well-established subjective measure is to hold a performance review

towards the end of a programme or activity and before the detailed planning of

the next stage. This needs to involve all the people who contributed to the

original discussions and helped in preparing the aims. The meeting will look at

each of the aims originally defined and will discuss whether there is any evidence

to support any improvement related to the targets.

According to Brody & Stone (1989:6) informal research are all those processes

that do not employ statistical methods, thus implying that the results are not

necessarily reliable. The trend, however, is to acknowledge many research

techniques which do not use statistics, as equally reliable. Informal research can

therefore be regarded as research that has not been verified through accepted

research techniques.

3.9 PUBLIC RELATIONS EVALUATION

Both formal and informal research methods have a place in the practice of public

relations and although instinct and gut feelings remain important in the conduct of

public relations work, management also demands measurement, analysis and

evaluation at every stage of the public relations process (Walker 1997: 98).

The term measurement is often confused with evaluation. Measurement is the

process of assigning numerical values to some or all attributes of the study

object. Another primary concern with measurement is whether the measure is

“valid” and “reliable”. Validity means that the measurement instrument actually

measures what it is supposed to measure. Reliability refers to the consistency of

the measure.

The increasing move to quality management with its expectations of measurable

outcomes has intensified the demands for the public relations practitioners to

evaluate their work. Economic constraints and increasing competition for

scarce resources favour those who can demonstrate their effectiveness with

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 57

facts and figures. At the same time, attempts to professionalise the public

relations industry have emphasised professional characteristics as particular

expertise, a shared body of knowledge and continuous learning (Walker 1997:

98).

Setting measurable objectives is essential in order to know the value of public

relations but it is difficult to decide how much should be spent to determine this

value. Given that the “pressure to measure” seems to outweigh current practice

in public relations evaluation, and the fact that public relations is increasingly

compared with other marketing disciplines, it can be argued that between 3 and

12 per cent of the public relations budget should be assigned to public relations

measurement generally (Macleod 1998: 384).

According to Cole (1997: 49) the lack of communication measurement is one of

the reasons why organisations cannot “fix” their communication problems. Many

people automatically turn thei r minds to the “off position” upon hearing words

such as “measurement, numbers, or statistics”. Cole suggests that there are two

choices for quantifying successful communication. The first is observing the

actual outcome, and the second is to implement a regularly scheduled procedure

to measure communication.

Many people trained in the communication arts have an aversion to

measurement, while people trained in the science of engineering and finance

tend to have an affinity for it. Between this aversion and affinity lies a fair amount

of tension. Managers frequently insist that they measure everything, that “what

gets measured is what gets done” (Lee 1999:13).

With a sophisticated understanding, an organisation can become adept at

measuring communication in a wise and constructive way.

In recent years, more and more public relations professionals and chief executive

officers have come to realise not only that reputations matter, but also that they

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 58

can be managed. Indeed, many public relations firms have gone so far as to

substitute the words “public relations” for “reputation management” to describe

their work (Klein 1999: 32).

But whether you call it “public relations” or “relationship management,” the

struggle has been the same - how to prove the effects of your work. Now,

however, public relations professionals have access to a new generation of

research tools that can help measure the reputation they’re purporting to manage

and prove whether their work actually improved it (Klein 1999: 32).

What’s motivating this rush to reputation management and measurement is a

plethora of evidence that a good reputation can dramatically affect an

organisation’s results. A 1997 Ernst & Young study, “Measures That Matter,”

found that as much as 40 percent of the average company’s market value is

based on non-financial assets, including its reputation. But before you can

manage a company’s reputation, you have to measure it (Klein 1999: 32).

According to Lee (1999:13) the first touchstone of good measurement is to

understand why you want to measure at all. It is not really constructive to

attempt to quantify the value of strategic communication. As economists remind

us, the value of anything is whatever someone is willing to pay for it. To the

extent that management doesn’t know how much to spend on communication,

the real problem may be a poor understanding of communication as a leadership

and management process.

The second touchstone of good measurement is to agree on what to measure.

In general, one should resist the temptation to measure the inputs and outputs of

the work of communication. Instead, one should measure the consequences of

communication. Examples of inputs are the resources (such as time, money

and personnel) that go into communication. Examples of outputs are the

frequency and circulation of a publication (Lee 1999:14).

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 59

The alternative is to concentrate on how well the communication function is

building awareness, understanding and acceptance of the competitive

environment and strategic direction as a foundation to commitment. To do that,

measurement should focus on outcomes , the impact of communication within the

organisation on whatever strategic metric is of importance. The third touchstone

of good measurement is to understand how to measure (Lee 1999:14).

3.10 WHY PRACTITIONERS DO NOT EVALUATE

It is widely and well known that practitioners have not evaluated their work in the

past. In this section, the researcher will investigate why this is the situation. A

survey of IPR members’ attitudes to evaluation, done by Watson (1993)

discovered that while three-quarters of practitioners claimed to undertake some

form of evaluation, three-quarters of respondents also agreed that little money

was spent on evaluation – from zero to five per cent of the total budget.

When questioned about their motives for undertaking evaluation, “prove value of

campaign” came out a very clear leader, followed by “help campaign targeting

and planning” and “need to judge campaign effects”. Another reason, “help get

more resources” came a distant fourth (Watson 1993:139).

Watson’s research showed that practitioners were defensive about their

activities. They used evaluation techniques to present data on which they could

be judged rather than using evaluation to improve programmes. Output

measurement was seen to be more relevant than gauging impact or gaining

intelligence so that programmes could be improved (Watson 1993:139).

It is common that communication professionals feel that they need to prove the

value and efficiency of their department, but don’t know where to begin.

Because the work of the communication professional is largely devoted to formal

communication - newsletters, brochures, videos, speeches and so forth – they

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 60

are looking for yardsticks on such things as the readership and readability of

publications (Lee 1999:16).

According to Baron (1997: 32) there are three main reasons why communicators

don’t measure:

1. Nobody asks them to or gives them time and/or budgets to do so.

2. The influences of communication programmes are almost never isolated, so

it’s difficult to posit relationships between performance and communication

programmes only.

3. Most communication professionals are not comfortable with research

methodology.

Two of the main reasons why communicators do not measure are insecurity and

not understanding the need to measure. They are afraid that the results will be

bad news.

Elizabeth Howell, ABC, concurs by saying, “When I don’t measure, it’s usually

because I don’t know how; I have no budget to outsource the measurement

activity, or management doesn’t really want to know” (Baron 1997: 33).

Many communication managers shrink from measurement because of anxious

uncertainty over what to measure and how to measure it. That, along with

intimidation by colleagues in other disciplines who insist to measure everything

(Lee 1999:17).

One of the major difficulties in evaluating the effectiveness of public relations

activity is the sheer range of the activity. Public relations can, for example,

support business objectives, explain policies, increase awareness, focus

attention on issues, encourage informed discussion, help to change perception,

influence attitudes, motivate staff, reinforce the marketing and sales effort, build

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 61

and sustain a reputation over time, help restore credibility, have some effect on

the values of a particular group (Dyer 1998: 13).

It is indeed a broad range of activities. To accomplish assessment of these

activities, current public relations evaluation strategies are very dependent on

surveys of public opinion. According to Dyer (1998: 14), there are furthermore no

models of public relations productivity and performance that specify categorical

statements of what is good or bad public relations outcome.

The specific challenge we face is how to go about communication measurement

in a meaningful way. Communication cannot be measured meaningfully if the

quantifying metrics is the disputed value of communication. To have meaning,

communication measurement must assume the value of communication and

focus on effectiveness with respect to the organisation’s strategic direction. That

makes measurement of communication worthwhile (Lee 1999:17).

Lee (1999: 17) pinpointed the main reasons why programmes were not formally

evaluated. These were, first, lack of knowledge, second was “cost”, followed by

“lack of time” and “lack of budget”.

There are also other reasons why evaluation is seen to be problematic.

• Understanding what it is that has to evaluated. Often what is measured is

output and not outcome. So we will be very happy to see a nice, fat clippings

file and will spend money to pay a clipping agency to collate the file for us.

However, in the long run it doesn’t matter how heavy the clippings file is, what

matters is what those clippings achieved (the outcome) (Watson 1993: 141).

• Setting objectives. Objectives need framing in measurable terms. “Raising

awareness” is not a good objective unless you quantify by how much: 1 per

cent or 99 per cent? Research will show you what is possible. Some

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 62

objectives will be fairly simple to quantify. The achievement of objectives is

the clearest way to evaluate any programme or campaign.

• Understanding of what can be achieved. Public relations practitioners

should make realistic promises. What is required is an honest, sober

appraisal of what can be achieved. The over-promising problem is

exacerbated by a genuine lack of knowledge of the psychological art of the

possible (Watson 1993: 142).

• Range of evaluation techniques required. Public relations are unlike some

other forms of marketing communications, such as direct mail, where the

evaluation is relatively simple. You count the number of returns and the

business transacted. Public relations addresses many audiences in many

different ways and different types of evaluation techniques are needed

(Watson 1993: 142).

• The communication chain. The decisions that have to be taken all along

the communication chain affect the communication outcome. You have to

decide on the message, the medium, the form of words and/or images, and

ensure the target is receiving and interpreting the communication correctly.

“Evaluation” has to take place all along the chain. If one element is wrong,

the desired outcome will not be achieved. Thus evaluation just at the end of

the programme can be misleading.

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 63

Past and current research on the measurement and evaluation of communication

focussed on the evaluation of the communication process. The focus should

instead be on the distinction of the goals of the company, the alignment of the

communication processes with the goals and the measurement of the output. In

other words, the measurement of the effectiveness of the technique and process

is not sufficient.

The trick is to focus on impact, not activity. It is pointless to survey employees on

their readership of a newsletter lacking much strategic information and future

perspective. Rather, measure their awareness, understanding, acceptance, and

commitment to the organisation’s strategic success. It is one thing to ask

employees, in a survey, whether they know the strategy. It is quite another thing

to ask them to state the strategy (Lee 1999:19).

Other practitioners set only process or output objectives, which refer to the effort

extended by the practitioner, such as “to send out 10 press releases per month”

or “to invite 100 people to a cocktail party”. However, if none of the press

releases is published or if no organisational message is given to guests at the

party, the public relations activity had no impact (Steyn 1999:10).

Before launching a programme, it should be pre-tested (i.e. through focus

groups) to see if it communicates the intended message. It should also be tested

along the way (in-process evaluation) to see whether the programme is making a

difference, and to fine-tune it (Steyn 1999:10).

Many public relations practitioners serve as mere practitioners and do not use

scientific research. As a result, public relations practice tends to focus on the

means or strategies, paying little attention to the specifics to be achieved. Few

programmes have measurable objectives specifying measurable outcomes, and

even fewer use systematic research to determine the nature of problem

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 64

situations, progress toward achieving objectives, or programme success or

failure (Broom & Dozier 1990).

The outcomes of the programme are monitored by means of summative

evaluation research, to see whether the problem has indeed been solved,

whether another problem has inadvertently been caused or whether a new

problem has arisen itself (Steyn 1999:11).

Internal evaluation can be done among employees (informal research such as

talking to people involved in executing the programme) to identify hiccups

experienced during implementation, which can be avoided in the future. External

evaluation (formal research such as surveys or focus groups or perception

studies) can be done to assess the climate among external publics, to see if

attitudes and perceptions have indeed changed (Steyn 1999:11).

The described model should form the basis of any corporate communication

actions. Without a research orientation and the skills necessary to make

corporate communication truly part of an organisation’s adaptive subsystem, the

function cannot become part of the organisation’s strategic decision-making

process (Steyn 1999:11).

Measuring the effectiveness of communication with respect to strategic direction

requires first appreciating the rightful role of communication in business. It is just

this: to create and cultivate employee awareness, understanding, and

acceptance of the organisation’s competitive environment and strategic intent as

a foundation to a broad, real commitment to executive strategy, welcoming

change and achieving goals (Lee 1999:19).

Viewed this way, communication is an essential leadership tool. No longer is it

merely the “morale shop” or a harmless publisher of newsletters with free

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 65

classified ads and grainy snapshots of company picnics. Now its mission and the

company’s mission are one and the same (Lee 1999:19).

The organisation’s top leaders may or may not have a sophisticated

understanding of communication. But chances are they will listen to and

embrace an explanation of communication in a strategic context. Commitment of

the personnel is needed and that can only be obtained through communication.

Commitment, moreover, requires the acceptance of strategic intent. Acceptance

of strategic intent requires understanding of it, and understanding requires

awareness of it. This four -stage progression - from awareness to understanding

to acceptance to commitment – is what strategic communication is really all

about (Lee 1999:20).

According to Grunig (1992: 186), there are two kinds of corporate communication

research – evaluation research and environmental scanning. For corporate

communication to contribute to organisational effectiveness (make a contribution

to the bottom line), the most senior manager or practitioner heading the public

relations function has to take responsibility for providing strategic information to

top management. This information should be focused on the organisation’s

stakeholders and their concerns, as well as on identifying and managing the

response to issues emerging in the internal and external environment, and

anticipating their consequences for the organisation’s policies or stakeholders.

Evaluation research will be discussed now and environmental scanning will be

discussed in length in the next chapter.

3.11 EVALUATION RESEARCH

Evaluation research or programme evaluation is conducted primarily to

determine the effectiveness of public relations plans or programmes. Such

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 66

studies are among the most sophisticated research activities that practitioners

can undertake.

Evaluation research adds objective feedback on programme impact to the

subjective assessments and informal research now used to judge programme

effectiveness. Evaluation research uses scientific procedures to collect, analyse

and interpret information to help determine the worth of something.

Evaluative research uses techniques that allow for comparisons between

situations and measure changes.

Surveys or focus groups in themselves are key communication vehicles sending

strong messages across the organisation. Valid techniques can be effective

when implemented at the outset and incorporated into the design of the

initiatives. The fact that most people today still rely mostly on focus groups to

measure is our own limitation. There is an art to research techniques and, there

are people who have been trained in effective measurement methods (Baron

1997: 33).

With the establishment of evaluation as an academic discipline, researchers

began to investigate the evaluation process itself: were evaluations too costly

and too late? Did evaluators use the right tools, methods and measures to find

the right answers?

Geddie (1996: 24) further came to the conclusion that surveys are a waste of

time and money until you use the results. “Communicators who let surveys and

other evaluations gather dust on a shelf – rather than using results to plan – are

doomed to begin every project at the bottom of the hill”. Strategic

implementation can begin as simply as setting baselines for improvement.

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 67

In the late 1970s and early 1980s a significant debate in the field was the

quantitative-qualitative debate which was strongly related to the discussion of

knowledge construction in evaluation (Puchan, Pieczka & L’Etang 1999: 167).

According to Steyn (2000:10) evaluation research can be subdivided into:

3.11.1 Formative evaluation

Formative evaluation helps practitioners to plan public relations programmes and

design communication strategies to deal with opportunities and threats found in

the environment.

It takes place before a public relations programme or plan begins or during

implementation. It helps practitioners to better formulate plans and design

implementation strategies (Steyn 2000: 10).

According to White (1998:15), formative research establishes benchmarks and

clarifies situations and objectives. Precise and measurable objectives at the

outset of a programme are a prerequisite for later evaluation. Diagnostic

research can be used to evaluate and, if necessary, adjust programme activities.

3.11.2 Summative evaluation

Summative evaluation, which measures public relations programmes, is used

both to monitor their implementation and to evaluate programme performance

against stated objectives (Steyn 1998: 22).

It is used to measure the overall impact of public relations programmes – both to

monitor its implementation and to evaluate programmes to ascertain whether

goals and objectives were met (Steyn 2000:22).

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 68

Communication auditing is a well-known evaluation technique to most

practitioners. It is research to evaluate public relations programmes and plans to

find out whether messages have actually gotten through to the receivers. It

assesses communication activities, widely used to study the readability and

readership of corporate newsletters, and other routine communication, such as

annual reports and press releases (Steyn 2000: 10).

In the implementation phase public relations plans and programmes are

developed to implement the corporate communication strategy. Once again

research plays an important role in evaluating ongoing programmes, pre-testing

the effectiveness of certain tools, profiling a specific stakeholder and its attitudes,

accumulating information about effective uses of media and evaluating

completed programmes and campaigns (Steyn 2000: 10).

Apart from quantitative objective measures, subjective measures of performance

are also quite legitimate. These factors may be especially important in the client

and consultancy relationship, but are also highly prized in the relationships that

in-house departments build with other departments within the organisation.

Examples of these subjective yardsticks are enthusiasm; efficiency and

professionalism; creativity and initiative (Gregory 1996: 145).

Communication professionals have an unproductive instinct to hold a yardstick

up to inputs and outputs, or a combination of the two, and compare them with the

yardsticks of their companies. It’s easy to measure inputs, which are just the

resources (such as money and personnel) that go into the communication. It’s

equally easy to measure outputs, such as quantity and quality of media. The

alternative is to concentrate on how well the communication function is building

awareness, understanding, and acceptance of the competitive environment and

strategic intent as a foundation to the needed employee commitment. To do that,

measurement should focus on outcomes, the impact of communication within the

organisation on whatever strategic metric is of importance (Lee 1999:19).

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 69

Many communication personnel still think of communication as only newsletters

and brochures and regard formal communication as the only communication.

Perspectives should be broadened to include semiformal and informal

communication. For together, semiformal and informal communication account

for the preponderance of communication in any organisation. It is also the most

important (Lee 1999:19).

Semiformal communication includes those programmes and initiatives,

procedures, systems, and processes that carry the organisation’s official

endorsement. They function as communication because they do what

communication does or should do: convey information, messages, and meaning

to people and induce or impede certain kinds of thinking or behaviour in

response.

Informal communication consists of relationships between leaders and the led,

driven by routine conversation-comments, questions, complaints, humor – as

well as by all those myriad behaviours and decisions, attitudes, and choices on

the part of leadership that so often speak louder than words (Lee 1999:20).

Communication professionals often shy away from thinking in terms of

semiformal and informal communication because they assume it is beyond their

power to influence or measure. Their pessimism is unwarranted. Lee (1999:21)

suggests that the four-stage progression - from awareness to understanding to

acceptance to commitment - provide a much-needed structure for measuring

impact. There is a distinct relationship between the three rubrics of

communication and the four stages.

Formal communication works best at building awareness and understanding.

Semiformal communication, to the extent it is consistent with its formal brethren,

builds understanding and acceptance. Informal communication - again, to the

extent it reflects the themes and messages of formal communication stand alone

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 70

in building commitment. With this relationship in mind, one can measure levels

of awareness, understanding, acceptance and commitment as an indication of

the impact of formal, semiformal and informal communication (Lee 1999:21).

In practical terms how does this translate into reality? There are a number of

critical factors to consider when planning a campaign or programme:

• Set measurable objectives;

• Build in evaluation and quality checks from the start;

• Agree on measurement criteria with whoever will be judging the success of

the work;

• Establish monitoring procedures that are open and transparant; and

• Demonstrate results (Gregory 1996: 143).

In recent decades, the emergence of trends such as the demand for

accountability, the need to prove goal attainment and the development of modern

management practices have forced not only private business but also the public

sector to acknowledge the necessity of evaluation. Evaluation provides its user

with two major advantages: feedback and the documentation of effectiveness

(Puchan, Pieczka & L’Etang 1999: 164).

The origin of programme evaluation is difficult to pin down. Planned social

research, as we would call it today, was recorded in 2000 BC in China where

sophisticated personnel selection measures were established. The last 200

years have seen an enormous increase in attempts to carry out evaluation, but it

was not until the 1960s that modern programme evaluation emerged as a

discipline. The 1960s was also the decade when established professional

societies, especially in social sciences, became interested in evaluation at their

annual meetings and conferences, and informal groupings of researchers

amalgamated into professional organisations such as the Evaluation Network

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 71

and the Council for Applied Social Sciences (Puchan, Pieczka & L’Etang 1999:

166).

In the 1970s, the Professional Evaluation Research Society became established

and professional journals were released such as the Evaluation Quarterly in

1977, later re-titled Evaluation Review and Evaluation and Program Planning. All

these developments were very much United States-centered (Puchan, Pieczka &

L’Etang 1999: 166).

Evaluation practices in the United Kingdom began in the 19th century, with the

Royal Commission on the Poor Laws in the 1830s, and other social surveys.

Evaluation developed most strongly after the Second World War when the field of

evaluation was most strongly connected with educational evaluation (Puchan,

Pieczka & L’Etang 1999: 166).

In the United Kingdom “evaluation” means mainly media evaluation or content

analysis, which may in some instances be quantified in terms of a measure of

advertising equivalence. In 1996, the trade magazine PR Week commented that

the emphasis is on the growing professionalism of the practice. It is expressed

through increased evaluation practice and it is worth noting that the evaluation

also becomes an issue in the relationship between the occupations of public

relations and marketing (Puchan, Pieczka & L’Etang 1999: 168).

A useful device is the macro-model of evaluation demonstrated by Jim

Macnamara (Macnamara 1992). The model forms a pyramid. At the base are

the inputs, basically information and planning, and at the peak, objectives

achieved. Each activity is split down into the various steps of the communication

process.

It recognises inputs and asks the user to make a judgement on the quality of

information, the choice of the medium and the content of the communication. It

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 72

then considers outputs, the communication produced, for example the newsletter

or press release and then it considers the results or outcomes – what the

communication actually achieved (Macnamara 1992).

The mode needs to be customised for each project, but the basics remain the

same. Its strength is that it recognises a range of evaluation methods. The more

advanced evaluation methods further up the pyramid are recommended. They

measure actual outcomes, are more sophisticated and more expensive. The

ones lower down the pyramid are more basic and can be seen as tests that you

are doing things right, more akin to quality control (Macnamara 1992).

Despite the trend of increase in evaluation, in 1998 PR Week reports that still

less than 5 per cent of most public relations budgets are spent on any kind of

evaluation. In public relations textbooks evaluation is part of the planning

process. There is a sharp contrast between American texts (which specify a

range of techniques and approaches to evaluation) and the British public

relations field, which is likely to be less knowledgeable and produces less

research literature (Puchan, Pieczka & L’Etang 1999: 168).

Contemporary European interest in the field has seen the establishment of

specialised interest groups in the 1980s and 1990s and the launch of a new

journal, Evaluation, which takes a more European focus. More recently, there

are signs that evaluation is becoming of more interest to the Latin countries.

The interest in evaluation has been stimulated by several important trends in

society: the consumer interest movement, the implementation of measures to

improve managerial effectiveness and the tendency to professionalisation in both

the public and the private sector.

When White and Blamphin undertook their Delphi research study among United

Kingdom academics and practitioners to discover what the national priorities for

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 73

research within the public relations industry were, evaluation came out top of the

list. According to Gregory (1999:142), nothing has changed since then.

This is largely because there is increasing recognition in the communication

world that to be taken seriously, there has to be evidence of the contribution

made. There is growing evidence that organisations and clients require public

relations and the communications functions to demonstrate their effectiveness

and that showing a file of press cuttings is not enough (Gregory 1999:142).

White and Blamphin’s Delphi study in the United Kingdom placed the

“measurement and evaluation of public relations” as top priority for research.

Academics have been urging practitioners to use more research at both strategic

and planning levels and to get to grips with evaluation. They should investigate

the value of research to the development of public relations and potential

influence at boardroom level. There have been some debate about what it is in

public relations that has to be evaluated - “measuring the effectiveness of public

relations efforts has proved almost as elusive as finding the Holy Grail”. The

problem, it seems, is twofold. Firstly, what do we measure and secondly, how do

we measure it? (Puchan, Pieczka & L’Etang 1999:170).

Since the Delphi study, there have been a number of initiatives. In 1996, a

cross-industry body formed to devise a standard unit of measurement, the PR

Point. This was followed in November of the same year by a two-day workshop

attended by practitioners and media evaluation companies from Europe and the

United States of America. The aim of the workshop was to devise a minimum

international industry standard. In October 1997 the International Committee of

Public Relations Consultancies Association (ICO) launched a client guide to

designing measurable communication objectives. At the same time the

Association of Media Evaluation Companies (AMEC) introduced a guide on

media evaluation (Gregory 1999:143).

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 74

At a forum in November 1999 hosted by PR Week it was agreed to develop an

evaluation tool kit endorsed by the participating bodies. As a result, the Public

Relations Research and Evaluation Toolkit was launched in May 1999.

Although there is no standard methodology that can be employed to make

evaluation simple, there is general agreement on a number of basic principles.

Evaluation starts at the beginning of a programme, not at the end. There has to

be clear understanding about the starting point. Next comes the setting of

objectives. For substantial programmes these objectives need to contribute

towards the achievements of the organisation’s overall goals. The final stage is

an assessment of whether or not the objectives were met, that is summative

evaluation (Gregory 1999:143).

3.12. FORMAL RESEARCH

Market research specialists will confirm that monitoring the effectiveness of a

public relations campaign is not as difficult as many people imagine; nor need it

be an extremely expensive activity. It is essential that the measure is of the

public relations effectiveness and that other changes are not inadvertently

monitored; for example other influences could include an increase in the sales

force, the effectiveness of a new sales manager, the impact of an advertising

campaign etc.

Therefore, as the original objectives will probably have been phrased in relation

to some form of quantifiable awareness or attitude factor, it is that which needs to

be measured. In fact, attitude surveys are probably the most valuable method of

assessing the effectiveness of the campaign. Attitude research should also be

the benchmark against which future effectiveness can be measured.

Without the completion of the communications loop, communication is one-sided

and one must then depend only on “gut feel” for guidance. But, what is “gut

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 75

feel”? It is the experience and background knowledge brought to bear in making

business decisions. Research, in providing the feedback link, does not detract

from this very important experience, nor does it replace it. Research adds to

one’s market knowledge, aiding communication managers in making more

informed decisions.

The use of research to evaluate public relations programmes has been growing

as client management steps up its demand for accountability throughout its

operations. This has accelerated the growth of research departments at major

public relations firms that measure everything from media coverage to

behavioural shifts. The new bottom-line focus among clients has also led to the

creation of strategic alliances between public relations and research firms (Harris

1998: 267).

Unlike the advertising business, which banded together decades ago to create

standards for research, public relations has no unified measurement standards.

Harris embraces the measurement model of Dr Walter K. Lindenmann, Senior

Vice President and Director of Research at Ketchum Public Relations (Harris

1998: 267).

In 1988, Dr Lindenmann devised a survey of public relations research,

measurement and evaluation to assess the full extent of public relations research

in the United States. This survey provided the first clear picture of what

practitioners believed or thought about research. As well, it gave insights into

what they used research for, when they used it, and which techniques they

mainly used (Walker 1997: 98).

According to Lindenmann (1998: 66) anyone can count press clippings and the

mistake that many public relations practitioners make, is to do the same thing

and pass the numbers and/or percentages that they come up with on to their

senior management. They assume that by doing so they have “measured” the

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 76

effectiveness of their public relations programmes and activities. This is only an

elementary first step in the very complex and involved process of seeking to

measure and evaluate impact.

All they are doing, is measuring what many in the field call measuring public

relations “outputs”. Outputs are usually the short-term results of a particular

public relations programme. It measures how well an organisation presents itself

to others, that amount of attention or exposure that an organisation might receive

(Lindenmann 1998: 66).

Measuring outputs need to be viewed as only the first step in the entire public

relations evaluation process. By counting press clippings and reviewing the

subjects that are covered in those clips practitioners can begin to measure

possible exposure to public relations messages, but cannot measure whether

target audiences actually saw the messages and respond to them at all

(Lindenmann 1998: 67).

What are really needed are mechanisms for measuring awareness and

comprehension, recall and retention, opinion and attitude change and

behavioural patterns. The Ketchum Public Relations and Measurement

Department concluded that to really determine, as fully as possible, whether a

particular public relations programme or activity had an impact, or not, a two-step

process was needed:

- setting in advance very specific and clearly defined public relations goals and

objectives, and

- pin pointing those levels of measurement that are crucial to the organisation

in determining to what extent those specific public relations goals and

objectives have been met.

The Department found three levels for measuring public relations effectiveness:

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 77

Level 1 is the basic level for measuring public relations outputs. This measures

the amount of exposure that an organisation receives from the media, the total

number of placements et cetera. It includes content analysis, segmentation

analysis etc (Lindenmann 1998: 67).

Level 2 is the intermediate level for measuring public relations outgrowths. It

measures whether or not target audience groups actually received the messages

directed at them, paid attention to them, understood the messages, and retained

those messages in any shape or form (Lindenmann 1998: 67).

Level 3 is the advanced level for measuring outcomes. This measures opinion,

behavioural change, determining if there has been a shift in views and/or how

people actually act when it comes to an organisation, its products or its services

(Lindenmann 1998: 67).

The different levels of measuring public relations impact can be plotted on a

yardstick in a hierarchical fashion.

Permission was obtained to use the Lindenmann survey for an Australian study

to identify behaviour, beliefs and attitudes using an existing instrument. The

study of public relations research in Australia has attempted to uncover current

practice in the industry and to identify the attitudes and beliefs of practitioners

about the place of research in their work. It provides some new information

about the beliefs underlying particular actions relating to public relations

research, measurement and evaluation (Walker 1997: 98).

Organisations need a performance measurement management framework that

allows for the choice of three important measures: effectiveness; efficiency and

cost-effectiveness. While the discussion of communication product and

programme effectiveness measurement still dominates communication

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 78

measurement literature, communication managers are finding that executive

management is just as concerned with efficiency and cost-effectiveness as it is

with effectiveness. It can be measured on three levels, being the level of

product, programme and positioning.

3.13 MAJOR EVALUATION COMPONENTS

What are the major components that need to be evaluated? For any public

relations evaluation research to be credible, five major components of the

process or steps need to be taken into consideration. It is first setting specific

measurable public relations goals and objectives. Then targeting public relations

outputs, next targeting public relations outgrowths, eventually targeting public

relations outcomes, and finally taking steps to link what has been accomplished

in public relations to the ultimate business goals, objectives, and

accomplishments of the organisation as a whole (Lindenmann 1998: 67).

According to Lee (1999:45) the sophistication of the measurement instrument

can vary from a stratified random-sample survey with response sheets of

computer-read bubbles to a quick pulse- and temperature telephone survey of

arbitrarily selected employees. Responses to the questions are analysed in

terms of awareness, understanding, acceptance and commitment and are then

tracked back to the relevant kinds of formal, semiformal and informal

communication they represent.

One fundamentally important point in all this stands out. The communicative

power of programmes and policies (semiformal) and of decisions, behaviours

and actions (informal) is a legitimate aspect of the communication professional’s

work. It is not realistic to expect the communication function to take responsibility

for the organisation’s communication processes and environment without it also

addressing the critically important semiformal and informal communication (Lee

1999).

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 79

3.14 SETTING SPECIFIC MEASURABLE GOALS AND OBJECTIVES

This component has to come first as no one can really measure the effectiveness

of anything, unless they first figure out exactly what it is they are measuring that

something against. In setting public relations goals and objectives, it is usually

important to recognise that measuring public relations effectiveness per se – that

is, the management of an organisation’s overall communications activities with its

target audience groups or publics – can be difficult to do unless the individual

elements or components of the programme are clearly defined. Instead of trying

to measure public relations as a total entity, steps should be taken to measure

the effectiveness of individual or particular public relations activities.

3.14.1 Measuring public relations outputs

Outputs are usually defined as the short-term, or immediate, results of a

particular public relations programme or activity. Media content analysis is one

of the principal methodologies used to measure media outputs. Outputs might

also be the assessment of a specific event, a direct mailing campaign or how a

chief executive officer handles himself or herself at a press conference. In any

event, both the quantity and quality of outputs can be measured and evaluated

(Lindenmann 1998: 69).

Macleod (1998:382) considers input as the first step. “Input provides

benchmarks to answer the question: Where are we starting from?” Input

research is a prerequisite to determine whether and to what extent attitudes and

behaviour have altered (outcome) in line with objectives.

Output measures whether the message was sent and aimed at the target

audience. One of the principal outputs is media evaluation – what message was

sent, where did it appear, who would have seen it, and what impact did it have?

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 80

3.14.2 Measuring public relations outgrowths

As important as it might be to measure public relations outputs, it is far more

important to measure public relations outgrowths. The usual starting point for

any public relations outgrowth measurement is to determine whether target

audience groups actually received the messages directed at them, paid attention

to them and understood the messages (Lindenmann 1998: 69).

Macleod (1998: 383) labels this step “out-take” – the degree to which the

audience has understood and retained the message imparted by public relations.

It is important to obtain benchmark data against which to measure any possible

changes in awareness and/or comprehension levels. To determine whether

there have been any changes at all in audience awareness and comprehension

levels usually requires some type of comparative study, either a “before and

after” survey or a test and control group study.

Two other outgrowth measures that are important for public relations

practitioners to examine are whether those in the target audience group can

recall the messages that are being disseminated, and whether they are retaining,

in any shape or form, the information that is being directed at them (Lindenmann

1998: 69).

Although recall and retention studies have not been conducted that frequently by

public relations practitioners, they clearly are important forms of outgrowth

measurement that ought to be seriously considered by public relations

professionals. Various data collection techniques can be used when conducting

such studies, including telephone, face-to-face, postal etc (Lindenmann 1998:

69).

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 81

3.14.3 Measuring public relations outcomes

Outcomes measure whether the communication materials and messages that

were disseminated have resulted in any opinion, attitude, preference and/or

behavioural changes on the part of those targeted audiences to whom the

messages were directed. “Opinion research” generally measures what people

say about something. “Attitude research” on the other hand, is far deeper and

more complex. Usually, attitude research measure not only what people say

about something, but also what they know and think, what they feel and how they

are inclined to act (Lindenmann 1998: 70).

According to Macleod (1998: 383) “Outcome is to what degree attitudes and

behaviour have been affected”.

The ultimate test of effectiveness - the highest outcome measure possible – is

whether the behaviour of the target audience has changed, at least to some

degree, as a result of the public relations programme or activity. It is general

knowledge that it is hard to measure behaviour, because it is difficult to prove

cause-and-effect relationships.

3.14.4 Measuring business and/or organisational outcomes

It is imperative that public relations practitioners take steps to link their

accomplishments to the ultimate goals, objectives and accomplishments of the

organisation as a whole. The objective is to relate public relations outcomes to

such desired business and/or organisational outcomes as increasing market

penetration, market share, sales and ultimately, increasing an organisation’s

profitability (Lindenmann 1998: 72).

Obstacles of evaluation in public relations practice include:

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 82

• Lack of understanding on the part of practitioners of the role of research and

evaluation, and of the techniques of research

• Lack of understanding on the part of clients

• Lack of time and resources .

There is a second process embedded within the overall scheme outlined above.

Evaluation takes place on a continuous basis to ensure the programme is on

track and will meet the required objectives. There are three stages at which

evaluation takes place.

The first stage is at the input level: was the information gathered for the

campaign platform adequate; were the channels of communication selected

correctly; was the message presented properly? (Gregory 1999:144).

The second stage is at the output level: how many messages were sent; how

many people received them; how many were placed in the media; who

considered them? (Gregory 1999:144).

The final stage is at the outcome level; what were the effects of these

communications; how many changed their attitude or responded positively to the

message? For some theorists there is an intermediate stage between output

and outcome and that is outtake (Gregory 1999:145). Given that the process is

not complicated, why is so little evaluation undertaken? Watson (1997:145) in

his study of IPR members discovered that it was due to ignorance of methods,

lack of budget and a lack of confidence in promoting evaluation to employers

and clients.

What is clear is that evaluation is becoming a requirement. It can help put the

professional communicator on an equal footing with other business professionals

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 83

who have to be accountable. What is incumbent on the professional

communicator is that they learn about the different types of public relations

research and how to conduct research properly (Gregory 1999:145).

Many of the factors that have shielded communicators from the need to evaluate

are changing rapidly. Also, leading-edge communicators are engaging in more

integrated and comprehensive interventions, so they can tie their multiple-tactic

programmes to performance. Professionals in a related career, training, are

already doing this. If employee and public communication professionals don’t

start, they will find their roles being taken over by others who are practicing a

newer and more appropriate form of communication/knowledge management

system (Baron 1997: 33).

3.15 SCORECARDS

Scorecards focus on many common elements, including: leadership; strategic

management; customer focus; internal processes; and employee learning and

change. Of these, strategic management, organisation and continuous

improvement are the most important measures for the communication function.

3.15.1 Strategic management

This can be done, in part, by examining the correlation between corporate and

business line objectives and actual work activities, and between communication

function objectives and work activities.

3.15.2 Organisation

We examine organisational structures, roles and resource utilisation against

planning priorities, production processes and client satisfaction. Does the

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 84

organisation of our work mirror the needs and priorities of our clients? (Likely

2000: 27).

3.15.3 Continuous improvement

We can conduct an examination of individual and group professional

development plans, function knowledge management systems and archival or

corporate memory mechanisms, and then compare them to client satisfaction,

product and programme results, internal and external recognition programmes

and industry benchmarks and trends (Likely 2000: 27).

For a communication manager to help position his or her organisation correctly in

the market, it is important to be aware of and monitor the reputation of the

organisation.

3.16 REPUTATION MANAGEMENT

In this section, the researcher will investigate what is meant by the term

"reputation management". Intuition will always play a role in understanding your

organisation’s image. But you don’t have to fly by instinct alone. More scientific

measures are now appearing as a growing number of research firms and

consultants, who see reputation management as a smart business opportunity,

are offering more affordable and accurate reputation management tools (Klein

1999: 33).

A public relations executive who’d like to measure his company’s reputation has

many choices of products and providers. Among the firms offering reputation

management services are Ernst & Young and PriceWaterhouseCoopers. But no

matter which tools you use, some general criteria apply as you move into a

reputation management study or survey. For example:

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 85

1. Get buy-in from the top. A company’s top management needs to understand

the importance of reputation management and measurement.

2. Measure all constituencies. A good gauge of a company’s reputation

considers the views of all its different stakeholders. “We say reputation is the

reflection of an organisation over time, and as seen through the eyes of all its

stakeholders”.

3. It’s not brand equity. Reputation must not be confused with brand equity,

which is typically tied to how one audience – the consumer – views a

company’s products or services. Or with corporate identity, which really

represents how the company defines itself. Reputation is all stakeholders’

view of the entire organisation.

4. Measure over time. You ought to look at reputation as financial people look

at the quarterly results. Once a benchmark study has been done, it’s possible

that some of the department’s other regular duties – like monitoring media

relations results – can contribute to regular updating of an understanding of

its reputation.

5. Apply what you learn. Understanding how stakeholders view a corporation

allows a public relations department to use limited resources more effectively

and increase its effectiveness as a strategic business tool. The bottom line is

that companies can’t afford not to know their reputation and other

constituencies (Klein 1999: 33).

The enormously challenging world of managing in the rapidly changing and

highly competitive markets, the world of complex thinking, the speed at which

change happens and evolutionary theories all have an impact on individuals in

various work environments (Hardijzer 2000: 22).

3.17 KNOWLEDGE MANAGEMENT

Achieving success in the knowledge economy is one of the fundamental forces

shaping the future of business. Knowledge is a fundamental driver of business

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 86

success and is effectively becoming an organisation’s most valuable asset and

its chief tool for creating wealth (Kelleher & Seekings 2000: 24).

In the knowledge economy a key source of sustainable competitive advantage

and superior profitability within an industry is how a company creates and shares

its knowledge.

Professionals can no longer know everything they need to know in order to do

their jobs well and compete in the marketplace. They require instant access to

critical information. In the future, organisations able to provide their members

with rapid access to the full repository of knowledge are the organisations that

are most likely to succeed.

3.17.1 The purpose of knowledge management

Every employee builds a wealth of knowledge through learning, skills

development and daily behaviour. Every employee possesses knowledge of

beneficial value to the company, yet most of this knowledge is not leveraged to

the collective benefit of the organisation (Breedt 2000: 22).

The purpose of knowledge management is to integrate the collective knowledge

of employees in such a way that the whole is greater than the sum of its parts.

This information must then be made accessible to all relevant individuals within

the organisation so that it may be utilised and applied to enhance the

competence and competitive advantage of the organisation as a whole. In

addition, knowledge management aims to provide the organisation with the ability

to learn, and thus, undergo a continual process of change. With this change

process comes the opportunity to improve and enhance the performance of the

organisation (Breedt 2000: 22).

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 87

3.17.2 Knowledge Management Objectives

The objectives of knowledge management is to develop a knowledge base equal

to all employees’ knowledge, skills, behaviours, perceptions, values, principles

and education within the organisation’s specific culture, strategy and structure.

Following an integrated approach to identifying, capturing, retrieving, sharing and

evaluating the enterprise’s knowledge assets, enables the organisation to

develop a knowledge base. These assets may include databases, documents,

policies and procedures as well as uncaptured, tacit expertise and experience

resident in the individual workers’ minds (Breedt 2000: 23).

It has been suggested that we have gone beyond the human capacity to process

information. The information overload currently experienced is the cause of more

than a third of all reported stress-related illnesses. Symptoms of information

overload include paralysis of analytical ability, feelings of being overwhelmed and

a loss of control, increased anxiety and decreased self-confidence, and an

increased tendency to blame others (Hardijzer 2000: 22).

Wurman (1989: 32) says: “Information anxiety is produced by the ever-widening

gap between what we understand and what we think we should understand.

Information anxiety is the black hole between data and knowledge. It happens

when information doesn’t tell us what we want or need to know”.

A weekday edition of the New York Times contains more information than the

average person was likely to come across in a lifetime in seventeenth-century

England (Wurman 1989: 32).

According to Wurman (1989: 35), more new information has been produced in

the last 30 years than in the previous 5,000. About 1,000 books are published

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 88

internationally every day, and the total of all printed knowledge doubles every

day.

Wurman (1989: 35) goes on to point out that, as we are confronted with

increasing amounts of information, we are increasingly unable to cope with it.

“The glut has begun to obscure the radical distinctions between data and

information, between facts and knowledge… And the more images with which we

are confronted, the more distorted is our view of the world”.

Information overload is certainly not diminishing. Information technology has

created a tidal wave of change to the extent that planning has become almost

unrealistic due to market volatilities, complexity and the various evolutionary

theories. The above is clearly indicated in the following figure.

Figure 3.1 Escalation of environmental turbulence

Source: Ansoff 1984

The explosion of information technology places a great deal of stress on our

cognitive capacity to deal with so much choice. The high technology, high-

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 89

pressure society has made it difficult for individuals to cope with choices, with the

ultimate impact of poorer choices due to time constraints.

According to Wurman (1989:39) “Meaning requires time-consuming thought, and

the pace of modern life works against affording us the time to think”.

It seems clear that it is not only the pace of modern life militating against

understanding and wisdom today, it is also the sheer volume and variety of

information we are confronted with. People have only so much time. How they

choose to spend their time is an important economic decision.

While computers have become faster and more powerful, there have been

comparatively few resources put into educating people about how to cope with

the mass of information with which they are confronted. Succeeding in the

knowledge economy is not just about technological brilliance and innovation; an

aspiring knowledge enterprise has also to overcome the biggest challenge of all

– culture (Kelleher & Seekings 2000: 24).

Knowledge is power, but what about information? Can there be anything more

dis-empowering than the torrent of information that cascades across our desks

each week? It’s hard enough to keep one’s head above the unstoppable flow of

information, much less make sense of it all.

The management of knowledge goes far beyond the storage and manipulation of

data, or even of information. It is the attempt to recognise what is essentially a

human asset buried in the minds of individuals, and leverage it into an

organisational asset that can be used by a broader set of individuals on whose

decisions the firm depends (More 1998: 353).

The Internet reveals that knowledge management is one of the hottest

management topics on companies’ agendas. Why such intense interest in

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 90

knowledge management? The old adage, follow the money, provides some

initial insight. With the increasingly complex problems that global organisations

face every day, knowledge has become the currency of competitiveness and

success.

Knowledge as an asset is beginning to supplant the traditional factors of

production – land, labour and capital, and has become the most important new

corporate and competitive resource. On average, about three-fourths of

companies’ market value stems from intangible assets of which the mainstay is

intellectual capital – patents, copyright, trade secrets et cetera (More 1998: 353).

We should ask: “how do we know that we know what we know?” and “how does

what we know shape our practical, daily activities?” Recognising the increasing

value of information and knowledge to business success, many organisations

have launched knowledge-management initiatives to leverage their competitive

advantages (Mudge 1999: 25).

The spirit of knowledge management becomes reality in an organisational

context when the principal knowledge management enablers – content,

community and computing are brought together.

3.17.3 Content, community and computing

3.17.3.1 Content

Content focuses on managed content, which means that things we know about

and know how to do are handled in a disciplined manner. Knowledge (content)

can be classified as “tacit” or “explicit” and the difference shapes how the

knowledge content will be managed (Mudge 1999: 6).

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 91

Explicit knowledge is what has been written or otherwise recorded. It includes

books, manuals, patents, databases et cetera. Tacit knowledge, on the other

hand, is personal, hard to formalise and communicate to others. Tacit

knowledge often takes the form of a mental model: beliefs and perspectives so

ingrained that they are difficult to articulate. It’s the wisdom and expertise in

people’s heads (Mudge 1999: 6).

3.17.3.2 Community

Community is largely about collaboration. Often the collaboration centres round

a business issue in which all community members have a vested interest.

Community is the most significant differentiator between knowledge management

and information management. A community might be defined by geography,

organisational units, functional specialities, or shared topics of interest (Mudge

1999: 7).

3.17.3.3 Computing

Computing includes the gathering, storage and maintenance of content; it also

provides access to content upon request of a community member. Computing

can be thought of as both the requisite backbone of knowledge sharing and one

of the most significant enablers (Mudge 1999: 7).

3.17.3.4 Content, community and computing working together

Millions of dollars are spent on collecting and moving content around within

companies – increasingly over company Intranets. Companies must develop the

processes, disciplines and know-how to focus on relevant, high-quality

knowledge – to become knowledge gourmets rather than knowledge gluttons .

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 92

Collaborative advantage will go to those who use technology to create mutable

and permeable boundaries between communities and to establish trust

relationships with external communities (Mudge 1999: 8).

Companies need to manage their knowledge because in part knowledge has

become their most important product. Effectively leveraging their knowledge

assets is key to understanding their customers’ needs and meeting those needs

in innovative ways.

Once information is identified, collected and managed, it must be transformed

into knowledge. This requires classification, analysis and synthesis – which

require human intervention. Knowledge is not captured. What is captured is

information that is more easily transformed into knowledge by the recipient.

Knowledge management exists at two levels, namely:

3.17.4 Management of information

Knowledge equals objects. This entails the construction of information

management systems. This track is new and growing rapidly, assisted by

developments in information technology.

3.17.5 Management of people

Knowledge equals processes. The focus here is primarily on assessing,

changing and improving human individual skills and behaviour. This track is not

growing as fast as the track on information.

Organisational knowledge and memory can be retained in six places: namely

individuals, organisational culture, organisational transformations, organisational

structures, organisational ecology and external archives. Individuals do,

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 93

however, remain the prime location for retention of the organisation’s knowledge

(Hardijzer 2000: 22).

Most employees of knowledge companies are highly qualified and highly

educated professionals - they are knowledge workers. Their work consists

largely of converting information to knowledge, using their own competencies for

the most part, sometimes with the assistance of suppliers of information or

specialised knowledge (Hardijzer 2000: 22).

The effort of knowledge management is not new at all. Successful organisations

have always made strategic use of the information at their disposal. The term is

used rather widely to describe a host of quite different practices and activities,

ranging from simple document management to business intelligence gathering to

complex efforts to create “learning organisations” (Mudge 1999: 25).

In some companies knowledge management is led by the information technology

unit; in others the impetus comes from specific business units such as sales and

marketing.

Information technology can support organisation knowledge and memory in two

ways: either by making recorded knowledge retrievable or by making individuals

with knowledge accessible.

Managers in some of the fastest-growing and more profitable businesses focus

on knowledge, see their business from a knowledge perspective and act as if

their intangible assets are real assets. By freeing themselves from the mental

straightjackets of the industrial age, some of these pioneer managers have found

a wellspring of limitless resources arising from the infinite human ability to create

knowledge and benefit from the convenient fact that, unlike conventional assets,

knowledge grows when it is shared. Knowledge on its own then cannot

represent power, only knowledge sharing will (Hardijzer 2000: 22).

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 94

Knowledge management is both a discipline and an art. It’s a management

discipline insofar as processes can be defined and implemented to capture and

tend the knowledge, to make it available to the workers, to keep track of who is

contributing to the knowledge arsenal and who is applying it well, and so forth

(Mudge 1999: 6).

But it’s the art part that actually causes knowledge management to work.

Knowledge management takes aim at evolving people’s attitudes and work

behaviours to effect new heights of collaboration – the intentional sharing of

ideas, information, knowledge and work itself – in support of a business need.

It’s about changing people’s value paradigm from “my information is power” to

“sharing is power” (Mudge 1999: 6).

According to (Mudge 1999: 6) we might define the spirit of knowledge

management as:

• Knowing individually what we know collectively and applying it.

• Knowing collectively what we know individually and making it (re) usable.

• Knowing what we don’t know and learning it.

Knowledge management is about changing the culture of an organisation in

order to exploit the full breadth and depth of experience and expertise contained

within. Traditionally, knowledge has been seen as a source of power and

competitive advantage for the individual – something to be hoarded. The

overriding aim of any knowledge management programme is to get people to

share their knowledge with the rest of their community and to see that it is in their

best interest to do so (Alexander & Ward 1997: 167).

The ability to manage human intellect and its conversion into useful “product” is,

perhaps, the most critical management skill in our age, providing the ultimate in

competitive edge. In most organisations managing intellectual capital and more

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 95

specifically the human side of knowledge management, remains largely

unexplored territory (More 1998: 353).

3.17.6 The nature of knowledge management

It’s helpful to understand knowledge management’s relationship to information

management before delving further into how knowledge management works.

Information management typically is associated with the industrial era. It focuses

on using information technology to enable the collection and management of

explicit business information, largely for purposes of management reporting

within hierarchical organisations (Mudge 1999: 6).

Knowledge, in contrast to information, could be thought of as the best

understanding that we have about a particular topic at a given point in time.

Because knowledge is rooted in human experience and social context, managing

it will mean paying attention to people, culture and organisational structure, as

well as to the information technology that is an essential tool for knowledge

sharing and use in large organisations.

There should be a differentiation between data, information and knowledge

before defining knowledge management. The relationships between these

entities are illustrated in the following knowledge creation process.

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 96

Figure 3.2: The Knowledge Creation Cycle

Source: Breedt (2000)

Knowledge, in theory, is a product of a continuous cyclical process that starts by

organising ubiquitous facts through specific stimuli into structured data and then

developing information through a process of aggregation. Once information is

internalised by a knowledge worker it is used to create knowledge and

understanding. Once the knowledge worker expresses this understanding in a

structured form it becomes information, e.g. advice, which is then disseminated

to yet again form data and is finally dissipated to become mere facts (Breedt

2000: 16).

Knowledge is what is learned or retained after data or information is forgotten or

the product is sold. Examples of knowledge include knowing where things are,

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 97

how they can be found, understanding and comprehending as well as developing

believes and values that are used to facilitate decision making. Knowledge

promotes understanding and proposes action whereas intelligence is the efficient

use of information to produce more knowledge (Breedt 2000: 16).

According to a variety of authors (Miller 1999: 12; Mitchell 2001: 8; Bednar 1999:

22) knowledge management is in trouble, to the benefit of the communication

professional. Companies have been ploughing huge organisational resources

into knowledge management for a couple of years, only now to hit what they call

“cultural blockages”. This is a polite way of saying that, even though all the

information technology systems are in place, our people still won’t talk to each

other.

This initial failure may prove a blessing. Now that knowledge management

initiatives are losing momentum, there’s a glorious opportunity for the internal

communication function to step in and invigorate the field. By galvinising

knowledge management activity, you will gain the boundless gratitude of

information technology, knowledge managers and of the chief executive officer

(Miller 1999: 12).

According to Miller (1999: 12) there are six ways that internal communication can

mobilise knowledge sharing in the company.

1. Show people what’s in it for them

2. Create neighbourhoods everywhere

3. Get those neighbours talking

4. Generate terrific content

5. Show why trust is vital in any neighbourhood

6. Show people how knowledge management can benefit your internal public

relations.

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 98

3.17.7 Knowledge creation

The target of a process for knowledge creation is to enhance the potential of

creating innovations. Firstly, knowledge domain members start by creating

collective tacit knowledge by jointly experiencing new work processes, tasks,

technological characteristics, use of technologies et cetera. This is not an easy

process. Members of the knowledge domain must spend considerable time

together, discuss and reflect upon their experiences, observe how their

colleagues solve tasks and interact with technologies and explain and give sense

to their actions (Aben, Nonaka & Von Krogh 2001:424).

In the next phase, the team attempts to make these collective experiences

explicit, through agreeing on proper, just and accurate descriptions of their

experiences. In the third step, this concept then becomes subject to scrutiny. It

is matched against market data, consumer trends and technological

requirements. A concept that successfully passes through this phase is

transformed into a prototype process, product or service (Aben, Nonaka & Von

Krogh 2001:424).

While these four steps typically cover the major steps of knowledge creation, the

fully-fledged process goes further by integrating the newly created knowledge in

existing manufacturing, marketing and sales. An important issue of knowledge

creation is to enhance the pace of innovation and to reduce the time span to

commercial success in the market (Aben, Nonaka & Von Krogh 2001:424).

3.17.8 Knowledge transfer

Knowledge transfer with external partners is also important. Strategic

partnerships provide mutual access to other companies’ knowledge. Research

and training agreements with universities and other research institutions provide

companies with access to recent research knowledge.

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 99

3.17.9 The relationship between corporate communication and knowledge

management

3.17.9.1 The nature of our work is changing

While even the most technologically savvy organisation will always need carefully

crafted messages and skillful practitioners to convey those messages, the

Internet, e-mail, and Intranets are spurring profound changes in organisational

communication. Participating in knowledge management initiatives will enable

communicators to develop and extend these new competencies (Mudge 1999:

27).

3.17.9.2 Communicators already know something fundamental about

knowledge management

“True knowledge management is taking all that information out there and

structuring it in ways that work to the advantage of both the individual and the

organisation”.

In addition to crafting messages, communicators provide structure and context

for an audience, and that is the fundamental act in knowledge making.

According to Mudge (1999: 27), departments like information technology or

finance have their own agendas. But a communicator’s agenda is to be the glue.

It is all about your audience and not about you.

3.17.9.3 Knowledge management is interesting

According to Mudge (1999: 27), this new discipline attempts to address

fascinating and complex technological and organisational issues that have large

implications for a business’s bottom line. By getting involved in these initiatives

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 100

now, communicators can help shape the discipline in ways that are useful to an

increasing number of people and organisations.

3.17.10 Strategic knowledge management

In order to manage competence (human knowledge) there must be an

understanding of how employees, and in particular professionals, should be

recruited, developed, motivated, and rewarded. To achieve this, dependence on

external experts needs to be decreased as the knowledge and wisdom exists

within the internal structure. Some other strategies that could be considered are:

• Implement distinct strategies for information and knowledge. An information-

focused strategy will address the development of information technology,

while a knowledge-focused strategy addresses the intangible assets, assets

that convert invisible revenues from a large number of activities into tangible

revenues (Hardijzer 2000: 27).

• Establish a learning culture. If people cannot learn how to learn, they run the

risk of becoming “walking encyclopaedias” of outdated information. The

learning process therefore will require a creative destruction of barriers to

learning and the broadening of access to new sources of knowledge and

experience. It also needs to be understood that learning can be formal,

informal of incidental (Hardijzer 2000: 27).

• Expand the virtual office. The knowledge worker can be accommodated in

terms of working from home or anywhere else for that matter.

If you want to make the most out of your company’s knowledge, knowledge

strategy formulation and choice need to be tightly coupled with other strategic

activities within the company. The development of a knowledge-based

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 101

advantage requires adequate attention and resource allocation paralleling the

development of other types of company advantages.

A company benefits from taking a proactive approach to its knowledge and

expertise, rather than just letting knowledge drift and evolve at the periphery of

management’s attention. In this sense, strategising in the knowledge economy is

about moving away from “driving ahead by looking in the rear-view mirror” to

“driving ahead by knowing what is around the corner.”

As was discussed, the organisation must provide an incentive for employees to

share knowledge and apply it to the mutual benefit of employees, departments

and the organisation.

3.18 CONCLUSION

According to Hardijzer (2000: 27), knowledge management is an abstract

concept that have become over-hyped and misunderstood. It is not a technology

tool but an amalgamation of strategy, technology and human application.

To take advantage of knowledge management requires investment. Knowledge

exists in people, not in technology, and requires human effort. Technology can

help to capture and store information, but it cannot create knowledge.

Knowledge management should not be allowed to become a stand-alone fad. In

order to successfully build a knowledge-sharing culture, it’s vital to integrate

knowledge management messages with an organisation’s wider corporate

communication efforts.

Knowledge is partially a result of research and the scanning of the internal and

external environment. Employees will not share their knowledge capital unless

UUnniivveerrssii ttyy ooff PPrreettoorriiaa eettdd –– JJaannsseenn vvaann VVuuuurreenn,, PP ((22000022)) 102

they feel safe and feel that they will receive benefit from their actions. Sound

corporate communication will enhance this process.

In many instances employees that are not on a managerial level can make

considerable contributions to the knowledge base of the organisations. A

specific culture of learning, and failure being part of the learning and knowledge

creating process, will add value.

The primary objective of knowledge management is to create and maintain

superior knowledge by making it available at point-of-need and enabling a

learning-organisation culture for the future.

In this chapter, the researcher investigated and conceptualised communication

research. Attention was also given to the application and management of

communication research.

Environmental scanning is the focus of Chapter Four. The origin of

environmental scanning and it’s correct application and management also

receives attention. The valuable contribution of this research technique is also

discussed.