Modern Economy, 2011, 2, 371-382
doi:10.4236/me.2011.23040 Published Online July 2011 (http://www.SciRP.org/journal/me)
Copyright © 2011 SciRes. ME
Testing Validity of Using Sample Mean in Studies of
Behavioral Facts
Yongchen Zou, Runqi Hu
School of Finance and St at i stics, East China Normal University, Shanghai, China
E-mail: rustyzou@gmail.com, rqhu712@gmail.com
Received Janarury18, 2011; revised April 15, 2011; accepted April 28, 2011
Abstract
In this paper we argue that a couple of taken-for-granted methods employed in studying behavioral facts of
human risk preference are mistakable. We call for within-subjects experiment design and propose a simple
statistical method that might be used to test the validity of using sample mean in interpreting as well as gen-
eralizing risk preferences.
Keywords: Risk Preferences, Behavioral Economics, Prospect Theory, Heuristics
1. Introduction
Decision making under uncertainty involves monetary
decision between a certain amount of gain/loss and a
gamble which gives you a certain chance to win the prize
/lose the money, taking the form of “Choose between a
sure gain/loss of x or a p chance to win/lose y”, where y
is always of greater magnitude a number than x. An
enormous amount of efforts have been invested into this
field by a great many behavioral economists and psy-
chologists, who produced various views and theories on
the mechanics of mankind’s decision making under un-
certainty.
Pioneering problems of this kind were first presented
by two psychologists, Daniel Kahneman and Amos
Tversky [1] in their renowned work of prospect theory,
revealing that people are generally loss averse and gain
certainty-seeking. Challenging the Subjective Expected
Utility (SEU) theory, prospect theory replaced the utility
function with the S-shaped value function, probability
with the weighting function. This critical step forward in
economics received enormous attentions.
Although theories on decision making under uncer-
tainty varies, they generally share methods such as
through questionnaires responses provided by agents.
Despite the many insightful views on how people make
decisions, little attention has been placed on the validity
of methods. We find that the ways in which data is ana-
lyzed in studies of this field are imperfect.
The main purpose of this paper is to aware of defects
lurking in the methods used in studies of decision mak-
ing under uncertainty, especially the misuse of sample
means, and introduce an approach of experiment data
analysis.
The remainder of this paper is organized as follows:
Section 2 provides a review of decision making under
uncertainty. Section 3 points out two major fallacies exist
in the studies of this field. Section 4 presents our ex-
periment on human risk preferences toward possible
gains and losses and introduces a new approach in ana-
lyzing data that circumvents the fallacies. Section 5 pro-
poses a theory of testing validity of use of sample aver-
age in deriving regression models. Section 6 is a conclu-
sion.
2. Decision under Risk
Precursors in studies of human decisions under risk,
Kahneman and Tversky concocted series of questions,
and through responds from the agents they found an in-
clined pattern in which people make monetary decision
under uncertainty. Typical problems presented by pros-
pect theory are as follows:
Problem 1: Imagine you are now $1000 richer. Choose
between:
A. Sure gain of $500; B. 50% chance to win $1000
Problem 2: Imagine you are now $2000 richer. Choose
between:
C. Sure loss of $500; D. 50% to lose $1000
For the matter that the majority of respondents chose
A and D, and that A and C (as well as B and D) are in
fact technically identical-namely if people choose A (B),
Y. C. ZOU ET AL.
372
they should have chosen C (D) to stay consistent in their
decision — a law in human decision making seemingly
obvious arises that people tend to seek certainty in gains
and averse to risk when it comes to losses, as is con-
cluded by prospect theory. Also, people focus on gains
and losses, which means that they take the reference
point of zero, or status quo, rather than concerning about
the final situation they will be in. For instance, if people
choose A and C, they will end up in both problem with a
final situation of $1500 richer.
Kahneman and Tversky (1979) replaced the utility
function with what they called value function, as is plot-
ted in Figure 1. Concave in the domain of gains and
more steeply convex in that of losses, the function indi-
cates that people are loss averse.
In addition, the two psychologists plotted a non-linear
weighting function, advanced in their later work (Kah-
neman and Tversky, 1992), that indicating people’s dis-
tortion in interpreting probability. See Figure 2. Particu-
larly, Kahneman and Tversky asserted, people overesti-
mate very low probability, which justifies people’s actions
Figure 1. Prospect theory value function.
Figure 2. Prospect theory weighting function.
in gambling and buying insurance. This argument is
supported by results of the following problems in their
1979 paper:
Problem 3: 0.1% to win $5000 or a certain gain of $5
where most people choose the former, and
Problem 4: 0.1% to lose $5000 or a sure loss of $5
where the majority favors the latter.
Similar brief reviews of prospect theory value function
could be found in Terrance Odean [2], Richard H. Thaler
and Eric J. Johnson [3], Richard Thaler [4], Daniel Kah-
neman and Mark W. Riepe [5].
The Striking of prospect theory is a tipping point
where a great amount of intellectual powers are evoked
and invest their time and efforts into exploring of human
decision making under risk. But as prototypes, pair
choices in the original paper left much to be desired: 1)
vast majority of the problems used in the paper is con-
cerned about monetary gains and losses. And things,
when associated with money, always become more com-
plicated than they are under other situations in that spe-
cial parts of people’s mind will be activated by sway of
human nature, such as greediness and fear of embar-
rassment for winning the least (losing the most) among
peers; 2) all problems in that paper is hypothetical, which
means that respondents won’t actually win the prizes or
suffer the losses, whatever choices they make. This could
give rise to differences in decision making, thus leading
to errors in statistics, as Steven J. Kachelmeier and Mo-
hamed Shehata [6] proved by their experiments con-
ducted in China, U.S., and Canada, which, however,
Amos Tversky and Daniel Kahneman [7] asserted insuf-
ficient and unnecessary; 3) a potential assumption hard
to be detected in the original paper is that, it presume
choice makers have “reset buttons” on their body, and
after every single choice is made, respondents will press
the button, reset themselves to the status quo physically
and psychologically, and go on with the next question.
This presumption is flawed in that in a series of decision
makings, risk-taking behavior is affected by the previous
choices the choice maker made. Risk-taking behavior
alters when previous gains or losses are incorporated into
decision making. This has been discussed by Staw [8],
Laughhunn and Payne [9], Tvesky and Kahneman [10],
Fredrick [11].
3. Fallacies in Methods
Although theories of decision making vary, they gener-
ally share methods of experimental nature usually in-
volving subjects, questionnaires, instructions, and inter-
pretation of subjective responses. But economics, distin-
guished from psychology, is not experimental. The idea
of generalizing what is said by myopic experiment re-
Copyright © 2011 SciRes. ME
Y. C. ZOU ET AL.373
sults into a theory of distant future has itself suffered
from cognitive biases. We find that discussion of ex-
periments results in studies of behavioral economics
were long done under a couple of major assumptions,
both have much to do with the use of sample mean,
which require justification.
3.1. Majority Heuristic
Consider the process in which prospect theory value
function is plotted: problem sets were concocted, choices
were made by the subjects, and an S-shaped function was
plotted in accordance with the majority’s risk preferences.
What are majority’s preferences? In studies of decision
making such as prospect theory, they generally refer to
the options favored by most people. But we should be
careful in deriving quantitative conclusion, such as the
value function, because it is constituted with two parts of
gain and loss domains. We argue that the intuitively
natural approach of connecting two branches up needs
reconsideration.
To illustrate the issue, we suppose the following two
problems and the results of them, mimicking the birth of
value function. Note that the case is intensely simplified.
Problem 7. Imagine you are now $2000 richer. Choose
between:
A. Sure gain of $500; B. 50% chance to win $1000
[70]* [30]
Problem 8: Imagine you are now $3000 richer. Choose
between:
C. Sure loss of $500; D. 50% to lose $1000
[35] [65]*
Assume that 100 subjects are involved in this study.
Note that the numbers in parentheses indicate the per-
centage of the subjects’ responses (e.g. of the re-
spondents choose A over B). Now responses of Problem
7 shows a sign of risk aversion, while those of Problem 8
indicates loss aversion. A function concave in the do-
main of gains and convex in the domain of losses is
therefore produced, as is shown in Figure 1, representing
the preference of the majority.
70%
Here the fallacy emerges. It has long been overconfi-
dently assumed that those who constitute the majority in
Problem 7 stay the majority in Problem 8. That is, ex-
perimenters mistakingly think that most, if not all, of the
70 subjects who chose in Problem 7 chose in
Problem 8, which is obviously erroneous for there exists
intersection in choices among different options. It is rea-
sonable to call this logic fallacy majority heuristic, and
somehow evade cognitive guard.
A D
Consider an extreme case of the responses of Problem
7 and 8: of the 70 subjects who chose in Problem 7,
only 35 D in Problem 8; and all 30 subjects who chose B
in Problem 7 favored D in Problem 8. This means that
the majority in Problem 8 is constituted by 35 MAJOR
subjects and 30 MINOR subjects in Problem 7 whose
responses were against the “trend”. Similarly, of 70 sub-
jects who chose A, 35 of them had chose C in Problem 8.
A
We maintain that studies in this field could change for
the better if results are interpreted in an innovative and
more statistically reliable way such that choices made are
documented, rather than marginally, in an integrate way.
For instance, of a problem set constituted by Problem 7
and Problem 8, instead of merely recording the percent-
ages of subjects who chose A, B, C, and D, more precise
preferences of every individual are documented as the
percentages of subjects choosing AC, AD, BC and BD.
Note that in the extreme case mentioned, the percentages
were 35% for AD, 35% for AC, 30% for BD, and 0% for
BC.
To prove that the majority heuristic matters, we draw
four typical problems (see Appendix) in the Kahneman
and Tversky’s 1979 paper and replicated the experiment
in which 32 subjects are involved. Rather than merely
analyzing the average preference, we recorded the result
individually, and compare each of them to the average
outcome.
Percentages in the parentheses shows the ratio of num-
ber of subject who made the corresponding choice to the
sample size.
Note that prospect theory value function indicates two
preferences of human decision making, namely risk
averse in the domain of gains and risk seeking in losses.
We note them R.A. in gains and R.S. in losses respec-
tively. Now subjects are said to be R.A. in gains if they
choose B in both PROB1 and PROB3, and R.S. in losses
if A are preferred in both PROB2 and PROB4. Define a
trend as existing if more than 50% of the subjects have it.
A perfect subject whose preference is consistent with
that is predicted by prospect theory will answer “B A B
A” in the two pair of problems.
Table 1 tabulates the result of this replication experi-
ment. We first draw an average trend of the result by
counting how many people chose A in the four questions
and then divide the numbers by the sample size. It turns
out that, as can be seen in the table, in the first pair of
problems, nearly 43.8% subjects chose A in PROB1,
which means 55.2% chose B, and 53% subjects chose A
in PROB2. Thus the average trend is “B A”, indicating
R.A. in gains and R.S. in losses. In the second pair of
problems, 81.3% answered B in PROB3, and 40.6% an-
swered A in PROB4. This, then, represents an average
trend of “B B”, being R.A. in both gains and losses do-
mains (note that this is a trend slightly disagree with
prospect theory’s prediction).
If we just look at the average trends, we see at least
Copyright © 2011 SciRes. ME
Y. C. ZOU ET AL.
374
Table 1. Result of a replication experiment of four prospect
theory problems.
Subject PROB1 PROB2 PROB3 PROB4
1 B A B A
2 A B B A
3 B A B B
4 B B B B
5 B A B B
6 A A B A
7 A A B A
8 B A B B
9 A A B B
10 A B B B
11 A B A B
12 A B A B
13 A A B B
14 B B A B
15 B A B A
16 B A B A
17 B B A A
18 B B B A
19 B B B A
20 B A B A
21 B B B B
22 A A B B
23 B B B B
24 A B B B
25 B A B A
26 A A B B
27 A A A B
28 A B B B
29 B B B B
30 B A A B
31 A B B A
32 B A B A
Choose A 14 [43.75%] 17 [53.13%] 6 [18.75%] 13 [40.63%]
Choose BA 10 [31.25%] 12 [37.5%]
Choose
BA,BA 6 [18.75%]
one of them is consistent with S-shaped value function,
and one might argue that if the sample size (or the num-
ber of questions) expands, prospect theory prediction
could be proved right. However, if the result of this rep-
lication experiment is interpreted in a different way, the
conclusion would be widely contradicting.
Instead of calculating an average answer, we now look
at within-subjects, or integrate, results by checking how
many subjects answered “B A” in both pair of questions,
and recall that “B A” indicates a preference of R. A. in
gains and R.S. in losses. As is shown in Table 1, only
31.3% and 37.5% of subjects answered “B A” in the
pairs of questions respectively, a result far cry in indi-
cating the trend of R.A. in gains and R.S. in losses. And
the majority heuristic is proved existing here because
although the average result shows a trend of R.A. and
R.S. in the first pair of questions, it is actually supported
by only 31.3% of the subjects.
Remarkably, this paradox becomes more irony when
we try to find “perfect subject” who answered “B A B
A”: there are only six of them, namely 18.8% of the
sample. If a trend is buttressed by no more than 20% of
the subjects, it can never be called a trend.
We argue that the majority heuristic has taught us a
paradoxical lesson that, people might be risk averse in
gains; people might be risk seeking in losses; but it
would be erroneous to assert that people are risk averse
in gains and risk seeking in losses. As a matter of fact,
we are to show in section III that even in the domains of
gains and losses respectively, human preferences cannot
be simply generalized as risk averse or risk seeking.
We should however point out that intersection of pref-
erences doesn’t happen all the time. Think of another
two problems and a pair of possible results:
Problem 9. Choose between:
A. Sure gain of $500; B. 50% chance to win $1000
[70]* [30]
Problem 10: Choose between:
C. Sure gain of $600; D. 50% chance to win $1,000
[75]* [25]
It is reasonable in this case to predict that all those
who chose A in the former problem favored C in the lat-
ter — no intersection happens. We speculate two pairs of
principles in human preferences here. Suppose A and B
are two outcomes,
a positive amount, and p a prob-
ability that 0p1
, then
a) If A is preferred to pB, A
is preferred to pB;
but A
is not necessarily preferred to BpP



;
b) If pB is preferred to A,
Bp
is preferred to A;
but
Bp
is not necessarily preferred to Ap
.
This indicates that an increment of benefit on the op-
tion favored would firm up the chooser’s preference,
Copyright © 2011 SciRes. ME
Y. C. ZOU ET AL.375
while increments on both sides don’t.
3.2. Misuse of Sample Mean
Studies that use statistical methods require appropriate
understanding of statistics used. When we require sub-
jective certainty equivalents for lotteries out of subjects,
we must make sure that the subjects understand what
mathematical expectation is because it would be unfair
otherwise to use the statistic that subjects don’t under-
stand later as a criterion testing subjects’ preferences.
More importantly, we should sure that we experimenters
ourselves fully understand the statistics we are using.
However, there exists the logic leap from existence of
statistic to apprehension of statistic. Just the fact that the
statistics are invented doesn’t mean they could be fully
understood. Typical example here is, we are of the opin-
ion, the basic and long been misused statistic of average.
Average, arithmetic or compound, is a basic statistic
that is used everywhere as a estimator of general stan-
dard. Broadly as it has long been used, human mind
could hardly grasp what an average really is. When it is
said that Class A bares an average score of 75 in a
mid-term examination, it is interpreted immediately by
human mind, or at least System 1, that almost every stu-
dent in Class A got test results in the neighborhood of 75,
which is obviously erroneous. And when the second
news came that Class B has an average score of 82, again
it is translated into a statement that almost every student
in Class B got a better score than students in Class A.
Ludicrous as it may sound, and despite almost everyone
would deny these kinds of thoughts had ever come to
them and they really understand what is an average, the
very logical fallacy is repeated frequently in studies of
human preferences, especially in those which experi-
ments involving subjects are conducted.
The common use of average is like this: conduct the
experiment to the subjects, have their responses docu-
mented, average the answers and analyze the average
answer as a supposed individual who represents the
whole sample. The basic idea of using average is to av-
erage out unwelcome volatility and thus get a general
trend. But this could be perilous in studying behavioral
facts of human preferences, for in these kinds of studies,
what we need is exactly the opposite: human preferences,
such as risk preferences, are very volatile, to say nothing
of the fact that an individual’s preference vary under
arbitrary situations. And this very divergence in prefer-
ences is exactly what worth investigating. This is not to
say average should be abandoned in preferences analysis.
As a matter of fact, one can never avoid using it because
it would be invalid and unnecessary to focus on every
individual. However, we can revise the sample mean
method by dividing the whole sample into groups with
obviously different characteristics, thus saving these di-
vergences from being willy-nilly averaged out. We are to
employ this method into our experiment in Part III.
Failure of human minds’ apprehension of other statis-
tics, such as expected value are entertainingly discussed
by Taleb [12], where he asked readers to imagine a com-
bination of pleasure of a vacation that will happen 85%
in Paris and 15% in Caribbean. In fact, he asserted, that
“[our] brain can properly handle one and only one state
at once — unless [you] have personality troubles of a
deeply pathological nature”.
4. Risk Preference Families
So far we have shown that the linkage of the two
branches of the value function and the use of total aver-
age in studying human preferences are mistakable. And
we claimed in Section II that not only risk preferences
should be studied separately in gain and loss domain,
even in the domains of gain or loss per se, they cannot be
simply generalized as “risk averse” or “risk seeking”.
Our experiment following is to test the different risk
preferences under gambles of different prize level.
4.1. The Experiment and Results
The process employed a two-stage method proved unbi-
ased by Becker et al. [13]. But rather than eliciting sub-
jective certainty equivalents under different probability
to win, we require certainty equivalents under different
prizes and losses at a same probability 50%. Prizes and
losses provided in the experiment ranged from 2 to
1000000 (2, 6, 10, 15, 25, 30, 50, 100, 200, 400, 750,
1000, 5000, 100000, 1000000). The experimental in-
structions are written in English. We assume this to be a
semi-“back translation” (a concept put forward by Brislin
[14]) for we, the designers, are Chinese).
The reason why we didn’t divide the probability into
smaller minor unity is that we fear that gradual increases
in probability make people confused and thus inducing
mindless answers. We assume that people get vague idea
about the difference between chances of 30% and 40%,
but they surely feel different among a half chance to win,
less than a half to win, and more than a half to win.
Hence we choose 50% as our constant probability, pre-
suming that our subjects stay cognitively conscious dur-
ing the test.
Note that all subjects are students major in either sta-
tistics, actuarial science, risk management, or financial
engineering, hence it is reasonable to assume than they
know quite well the concept of mathematical expectation,
justifying our using it later as a criterion in discussing
Copyright © 2011 SciRes. ME
Y. C. ZOU ET AL.
376
risk preferences.
To rule out the majority heuristic and the misuse of
total average discussed in Part II, we employ a different
method in interpreting the results. Risk preferences to-
ward possible gains and losses are discussed separately.
We are to show the risk preferences of gain domain and
those of loss domain, but we don’t and cannot provide a
single type of risk preference that summarizes the whole
picture. We avoid believing in average of the total.
Rather, responses are categorized, according to their
characteristics, into groups. In fact, we intuitively ex-
pected, rather than predicting, a general but nothing spe-
cific trend before we get the results: respondents may
show a tendency toward risk seeking when the possible
prizes/losses are small, and they may tend to be risk
averse when the possible prizes/losses surge to certain
higher levels. However, we didn’t plan to derive the
general trend which represents the average (and we don’t
believe any single one could represents all); instead, we
focus on detecting different patterns of risk preferences
among subjects. We even tried to analyze anomalies
among the results and explain why they are like that.
Defects in our experiment is that it provided no real
money rewards. Since we are trying to elicit subjects’
preference under big prize, we cannot afford to actually
pay them what they won (For experiment that involving
real subjects losses, see Etchart-Vincent and l’Haridon
[15]). But they are informed that rewards are to be paid
according to their final results. Subjects are urged to be-
have as if the gambles are real and independent
throughout the experiment.
We used in the gain domain the ratio of certainty
equivalent to expected value, or CE/E, as major criterion
of subjects risk preferences. If CE/E of a single response
exceeds unity, the subject is said to show a sign of risk
seeking toward the gamble, while if CE/E is lower than
one, it signals risk averse. In the loss domain, on the
other hand, the ratio of maximum premium (which means
that the maximum certain quantity the subject is willing
to pay to eliminate the risk of suffering potential loss) to
expected value, or MP/E, is used so that if MP/E of a
response toward a gamble exceeds one, it shows risk averse,
and if MP/E is less than one, it shows risk seeking. For
example, if the subject is extremely risk seeking toward a
potential loss, then she is not willing to pay even a penny
to an insurer to have her asset underwritten. In this case,
the maximum premium will be 0, thus MP/E is equal to 0.
For the purpose of illustrating the misuse of total av-
erage that has been discussed in Part II, Figure 3 mani-
fests the total average CE/E and MP/E values under the
prize/loss series toward gains and losses respectively. It
seems obvious, from the graphs, that there is a trend of
decreasing CE/E ratio and increasing MP/E ratio. What’s
Figure 3. Total average of CE/E ratio and MP/E ratio.
more, average CE/E ratio of our experiment rarely ex-
ceeds unity conspicuously and so does MP/E ratio. This
indicates that except for extremely small prizes (such as
50% to win 2) and extremely big losses (such as 50% to
lose 1000000), respondents show preferences of risk
averse in gain domain and risk seeking in loss domain,
consistent with what is told by prospect theory and its
value function.
We categorized according to their features the final
results of gain domain and loss domain respectively into
three families. As is shown by Figure 4 some typical
individual responses, these categories are named “Gen-
eral”, “Arch”, and “Bounce” in gains, and “General”,
“Neutral” and “Seeking” in the losses. Do notice that
these results are not paired with each other and must be
viewed separately, which means that, for instance, a
subject in “General” family of gain domain can either be
in “General”, “Neutral” or “Seeking” in the loss part.
The first three captures of Figure 4 manifests three
preference families that we detected from the subjects in
questions involving possible gains. “General” family
shows a preference pattern that fits the majority of the
respondents, in which agents show risk seeking toward
relatively small prizes, and they gradually become risk
averse as the prize become more and more attractive.
Note that although “General”, like the total average CE/E,
decreases as the prize level sour, the line does not fall
below unity until the prize level reaches somewhere
around 200, whereas the total average CE/E rarely sur-
passes one. Respondents in “Arch” family typically start
from strict risk averse, then becomes increasingly risk
seeking, but when the prize level seems attractive enough
to them, they show a sign of risk averse again. The CE/E
Copyright © 2011 SciRes. ME
Y. C. ZOU ET AL.
Copyright © 2011 SciRes. ME
377
Figure 4. Categorizing of final results in gain and loss domains.(1.General)
line of this family crosses the unity twice. Subjects in
“Bounce” family, albeit not large in number, shows an
interesting and sharp risk preference reversal when the
prize level gets extremely high, before which they gener-
ally share risk preferences with those in “General” family.
The last three captures, on the other hand, shows the
three risk preference families of the loss domain. Agents
in “General (Loss)” family are shown to have increasing
MP/E ratio as the loss level approaches extreme. But
note that MP/E values of a typical respondent in “Gen-
eral (Loss)” family exceeds unity when the possible loss
level reaches 200 or so, while in total average ratios,
MP/E only goes beyond 1 when the possible loss be-
comes astronomical. “Neutral” family possesses MP/E
ratio hovering around one whatever the possible loss
would be. Finally, subjects in “Seeking” family have
MP/E ratio always below unity; they are always risk
seeking, tending to run the risk of losing.
Counterintuitive finding of our experiment is that risk
preferences of gain and loss domains barely cross paths.
For example, not a single one of the subjects showed a
sign of risk neutral in gains part, and no one in losses
part presented a sudden preference reversal as is in
“Bounce” family. We find that subjects behaved sharply
different toward risks of certain gains and losses, and
their psychologies, influenced by various factors, varied
individually and were very volatile.
We interviewed after the experiment some of the sub-
jects and asked for what were their psychologies when
providing answers. To our surprise, every subject inter-
viewed gave exactly the same reason justifying their
subjective certainty equivalent toward small prized and
losses (such as a 50% to win/lose 2), which goes like:
The possible gains/losses are so small that I don’t care if
I win or lose them. But their explanations diverse. For
those who have CE/E ratio of 2 or bigger toward possible
gain, their explanations were typically “I don’t care
about the little gain so that I just gamble. Either win or
not win, no big deal”; for those who have CE/E ratio of
smaller value, their justifications were like “I don’t care
the winning of the little amount of money”. In the loss
part, none of the subjects has high MP/E ratio toward
small possible loss; they chorused a preference of risk
seeking, with MP/E ratios ranging from 0 to 1. Similarly,
they asserted in the aftermath interview that they “an-
swered so because they didn’t care about the small
Y. C. ZOU ET AL.
378
losses”. Rationally speaking, if subjects “don’t care”
whether they gamble or not, namely they think them-
selves indifferent between gambling or not gambling,
that should be interpreted as risk neutral thus their sub-
jective certainty equivalent should be equal to the
mathematical expectation, entailing a CE/E or MP/E
ratio of 1. However, the data just showed that behavior
of most subjects wasn’t consistent with this rationality.
Thus we have reason to doubt that when CE/E or MP/E
ratio does equal to unity, that doesn’t necessarily mean
the subjects are risk neutral toward the gamble.
The sudden CE/E ratio bounce back of subjects in
“Bounce” family reflects gambler’s psychology. When
respondents in the family were asked why they suddenly
wanted to take risk in the high prize gamble, they typi-
cally answer that when the prize level became extremely
high, the lottery turns into a gamble. And a gamble is a
gamble, either one win the big, or walk away with noth-
ing.
Most subjects in “Neutral” family of loss domain said
they feel horrible on the thought of losing their money,
be it the premium they pay or the money they lose if they
choose to gamble. Rather than endless dithery, they sim-
ply chose to stay indifferent: write down certainty
equivalents entered around mathematical expectations,
and whatever happens is destiny.
Agents in “Seeking” family of loss domain usually
assert that they would rather run the risk of losing the
gamble than paying anything to eradicate the risk, so that
their fortunes still have a good 50% chance to stay intact.
Although we highly doubt that this kind of psychology is
resulted partly from the fact that the possible losses in
the experiment are hypothetical.
4.2. Models
Kachelmeier and Shehata (1992) introduced a linear re-
gression model that includes subject effects, incorporat-
ing correlated errors entailed by individual differences in
risk preferences. Our models bases essentially on the
Kachelmeier and Shehata’s CERATIO model but a) we
ignore the percent effects for lotteries in our experiment
share the constant winning/losing chance of 50%, b) we
used logarithm of prizes/losses in the model and c) we
introduce a similar MPRATIO model for MP/E ratios
with logarithm of prize factor replaced by logarithm of
loss. Our regression models are shown as followed:

1
1
2
1
CERATIOlog Prize
n
jj
jSubject



(1)

**
1*
2
1
CERATIOlogLoss
i
n
jj
jSubject



(2)
where CERATIO and MPRATIO respectively means
ratio of certainty equivalent to expected value and ratio
of maximum premium to expected value;
and *
are intercepts of regressions; log(Prize) and log(Loss) are
logarithm of Prizes and Losses series in the lotteries;
j
Su
n
bject refers to subjects factor, which equals 1 if the
observation is from the j th observation, –1 if from the
th observation, and 0 otherwise;
is the disturbance.
The regression coefficients and the results of tests are
tabulated as follows.
Where CONSTANT= the intercept parameter;
i
S
*
p = the j th subject
= logarithm of Prize
Advantage of categorizing risk preference into fami-
lies, and thus deriving the categorized regression models,
could be observed first in terms of the increase in both R
square, namely the goodness of fit, and adjusted R square
in both “General” families of gain and loss parts as is
shown by Table 3 and Table 7, compared to that of the
sample average data in Table 2 and Table 6. R square
increases from 0.433 in sample average to 0.758 in
“General” family of gain domain, and from 0.784 in
sample average to 0.839 in “General” family of loss do-
main. And the corresponding adjusted R square as well
increase from 0.392 to 0.533 and from 0.586 to 0.683
respectively. Besides, estimates derived by OLS tech-
nique in CERATIO and MPRATIO model show statisti-
cal significance by and large, as can be seen particularly
in Table 3, Table 5, Table 7, and Table 8. Table 4 and
Table 9, corresponding to “Arch” family of Gain Do-
main and “Seeking” Family of Loss Domain, show in-
significance in estimates of parameters; we attribute this
insignificance to the small sample size, and it can be that
they are not representative to the population.
Second, in most families, the variances of estimates of
both CERATIO and MPRATIO drop from original vari-
ances of estimates of CERATIO and MPRATIO in sam-
ple average regression models, indicating better stability.
See Table 10.
5. “D” for Testing Validity of Sample
Average
It is mentioned without evidence in Sec II.C that the use
of total average in analyzing behavioral facts of human
preferences is mistakable for it averages out not only
unwelcome volatility of the data but also the divergence
between subjects preferences which worth investigating.
This is observable from Figure 3 and Figure 4, espe-
cially in the gain part, for the three risk preference fami-
lies, “General”, “Arch”, and “Bounce”, viz., of gain do-
main bear sharply different shaped CE/E lines.
To test whether the use of sample average averages
Copyright © 2011 SciRes. ME
Y. C. ZOU ET AL.379
Table 2. Results of CERATIO model for sample average of
gain domain.
Unstandardized
Coefficients
Standardized
Coefficients
Model
B Std. ErrorBeta
t Sig.
(Constant) 1.240 0.124 9.974 0.000
P* –0.056 0.008 –0.341 –6.9700.000
S1 0.233 0.165 0.095 1.4170.158
S2 0.143 0.165 0.058 0.8700.385
S3 –0.118 0.165 –0.048 –0.7170.474
S4 0.236 0.165 0.096 1.4330.153
S5 0.599 0.165 0.244 3.6390.000
S6 0.126 0.165 0.051 0.7650.445
S7 –0.837 0.165 –0.341 –5.0800.000
S8 –0.051 0.165 –0.021 –0.3120.755
S9 0.037 0.165 0.015 0.2240.823
S10 0.256 0.165 0.104 1.5520.122
S11 0.210 0.165 0.085 1.2730.204
S12 0.037 0.165 0.015 0.2250.822
S13 0.399 0.165 0.163 2.4220.016
S14 –0.491 0.165 –0.200 –2.9810.003
S15 0.358 0.165 0.146 2.1740.031
1
S16 0.189 0.165 0.077 1.1460.253
a. Dependent Variable: CERATIO
Model Summary
Model R R Square Adjusted R
Square
Std. Error of the
Estimate
1 0.658a 0.433 0.392 0.4510
a.Predictors: (Constant), S16, P*, S5, S2, S1, S4, S3, S8, S7, S6, S9, S10,
S12, S11, S13, S14, S15.
Table 3. Results of CERATIO model for “general” family
of gain domain.
Coefficientsa
Unstandardized
Coefficients
Standardized
Coefficients
Model
B Std. ErrorBeta
t Sig.
(Constant) 1.435 0.106 13.537 0.000
P* –0.092 0.009 –0.601 –10.7
32
0.000
S8 –0.051 0.135 –0.029 –0.38
2
0.703
S9 0.037 0.135 0.021 0.2740.784
S10 0.256 0.135 0.142 1.8960.060
S11 0.210 0.135 0.117 1.5560.122
S12 0.037 0.135 0.021 0.2750.784
S13 0.399 0.135 0.222 2.9600.004
S14 –0.491 0.135 –0.274 –3.64
3
0.000
S15 0.358 0.135 0.200 2.6570.009
1
S16 0.189 0.135 0.105 1.4000.164
a. Dependent Variable: CERATIO
Model Summary
Model R R SquareAdjusted R
Square
Std. Error of the
Estimate
1 0.751a 0.564 0.533 0.3691
a.Predictors: (Constant), S16, P*, S15, S14, S13, S12, S11, S10, S9, S8.
Table 4. Results of CERATIO model for “arch” family of
gain domain.
Coefficientsa
Unstandardized
Coefficients
Standardized
Coefficients
Model
B Std. Error Beta
t Sig.
(Constant) 1.2460.152 8.216 0.000
P* –0.0130.017 –0.101 –0.7880.434
S1 –0.0030.170 –0.002 –0.0160.988
S2 –0.0930.170 –0.085 –0.5460.588
1
S3 –0.3540.170 –0.326 –20.0840.042
a. Dependent Variable: CERATIO
Model Summary
Model R R Square Adjusted R
Square
Std. Error of the
Estimate
1 0.323a 0.104 0.039 0.46513940
a.Predictors: (Constant), S3, P*, S2, S1.
Table 5. Results of CERATIO model for “bounce” family of
gain domain.
Coefficientsa
Unstandardized
Coefficients
Standardized
Coefficients
Model
B Std. Error Beta
t Sig.
(Constant) 0.0610.182 0.333 0.741
P* 0.0070.022 0.033 0.3250.747
S5 10.4360.193 0.866 7.4560.000
1
S6 0.9630.193 0.580 4.9980.000
a. Dependent Variable: CERATIO
Model Summary
Model R R Square Adjusted R
Square
Std. Error of the
Estimate
1 0.765a 0.585 0.555 0.527
a.Predictors: (Constant), S6, P*, S5.
Table 6. Results of MPRATIO model for sample average of
loss domain.
Coefficientsa
Unstandardized
Coefficients
Standardized
Coefficients
Model
B Std. Error Beta
t Sig.
(Constant) 0.8750.088 9.895 0.000
S1 –0.3740.117 –0.177 –3.1880.002
S2 –0.2140.117 –0.101 –1.8290.069
S3 –0.0820.117 –0.039 –0.6970.486
S4 –0.1020.117 –0.048 –0.8720.384
S5 –0.3870.117 –0.183 –3.3020.001
S6 –0.7880.117 –0.372 –6.7270.000
S7 –0.6420.117 –0.304 –5.4820.000
S8 –0.9740.117 –0.460 –8.3160.000
S9 –0.8370.117 –0.395 –7.1400.000
S10 0.0420.117 0.020 0.3620.718
S11 –0.4660.117 –0.220 –3.9730.000
S12 –0.0870.117 –0.041 –0.7400.460
S13 0.2630.117 0.124 2.2470.026
S14 –0.3530.117 –0.167 –3.0130.003
S15 –0.3190.117 –0.151 –2.7210.007
S16 –0.6400.117 –0.303 –5.4650.000
2
P* 0.0580.006 0.410 10.1660.000
a. Dependent Variable: MPRATIO
Model Summary
Model R R Square Adjusted R
Square
Std. Error of the
Estimate
2 0.784a 0.614 0.586 0.32091842
a.Predictors: (Constant), P*, S16, S5, S2, S1, S4, S3, S8, S7, S6, S9, S10,
S12, S11, S13, S14, S15.
Copyright © 2011 SciRes. ME
Y. C. ZOU ET AL.
380
Table 7. Results of MPRATIO model for “general” family
of loss domain.
Coefficientsa
Unstandardized
Coefficients
Standardized
Coefficients
Model
B Std. ErrorBeta
t Sig.
(Constant) 0.600 0.094 6.381 0.000
P* 0.109 0.008 0.678 13.1330.000
S10 0.042 0.117 0.025 0.3640.717
S11 –0.466 0.117 –0.272 –3.9890.000
S12 –0.087 0.117 –0.051 –0.7430.459
S13 0.263 0.117 0.154 2.2560.026
S14 –0.353 0.117 –0.206 –3.0250.003
S15 –0.319 0.117 –0.186 –2.7320.007
2
S16 –0.640 0.117 –0.375 –5.4870.000
a. Dependent Variable: MPRATIO
Model Summary
Model R R Square Adjusted R
Square
Std. Error of the
Estimate
2 0.839a 0.705 0.683 0.31963043
a.Predictors: (Constant), S16, P*, S15, S14, S13, S12, S11, S10.
Table 8. Results of MPRATIO model for “neutral” family
of loss domain.
Coefficientsa
Unstandardized
Coefficients
Standardized
Coefficients
Model
B Std. ErrorBeta
t Sig.
(Constant) 0.700 0.067 10.468 0.000
P* 0.019 0.007 0.268 2.7320.008
S1 0.013 0.078 0.021 0.1710.864
S2 0.173 0.078 0.275 2.2170.030
S3 0.305 0.078 0.486 3.9220.000
2
S4 0.285 0.078 0.453 3.6580.000
a. Dependent Variable: MPRATIO
Model Summary
Model R R Square Adjusted R
Square
Std. Error of the
Estimate
2 0.581a 0.337 0.289 0.21312713
a.Predictors: (Constant), S4, P*, S3, S2, S1.
Table 9. Results of MPRATIO model for “seeking” family
of loss domain.
Coefficientsa
Unstandardized
Coefficients
Standardized
Coefficients
Model
B Std. ErrorBeta
t Sig.
(Constant) 0.326 0.071 4.601 0.000
P* 0.005 0.008 0.074 0.6350.528
S6 0.048 0.079 0.087 0.6090.545
S7 0.194 0.079 0.350 2.4480.018
2
S8 –0.138 0.079 –0.248 –1.7360.088
a. Dependent Variable: MPRATIO
Model Summary
Model R R Square Adjusted R
Square
Std. Error of the
Estimate
2 0.499a 0.249 0.195 0.21731274
a.Predictors: (Constant), S8, P*, S7, S6.
Table 10. SDs of CERATIO and mpratio in different mod-
els.
Gain averageArch Bounce General
ˆ
0.4510 0.4651 0.527 0.3691
Loss averageNeutral Seeking
General
(Loss)
ˆ
0.3209 0.2131 0.2173 0.3196
out volatility or divergence, namely to test whether the
use of sample average is desirable, one should look into
the difference between estimates of target values, in this
paper the CERATIO and MPRATIO, before and after the
sample is categorized into families.
We here focus on the difference between the “dis-
tance” between estimates of CERATIO/MPRATIO using
categorized models (the families models) and sample
average model to the real sample average.
Define Distance Function, denoted DF, which refers to
the distance between regression model estimates of
CERATIO/MPRATIO to the real sample average of
CERATIO/MPRATIO:
2
15
,
11
CERATIO CERATIO
ni
ij
ij
DF 
 (3)
15
,
11
MPRATIO MPRATIO
ni
ij
ij
DF 
 2 (3)
where n = subjects sample size;
, = regression model estimate of CE/E
ratio of the jth subject, towards the th lottery;
CERATIOij i
CERATIO
MPRATIOij
i = the sample average of CE/E ratio to-
wards the ith lottery and similarly are the meanings of
and
,MPRATIOi.
Define D as:
12
1
S
i
i
DFDF n
(4)
where s = number of families (in this paper three, for
gain and loss parts respectively);
n = subjects sample size;
DF = the value by Distance Function in (3) and (3)’,
then the difference between Ds of uncategorized model
and categorized models, denoted , is the “Avera-
geout Degree” of the sample average regression model,
measuring the difference between estimates of the
CERATIO and MPRATIO before and after the sample is
categorized into families. Thus the averageout is more
desirable,which means that it is more likely the use of
sample average averages out unwelcome volatility rather
than individual divergences, if the absolute value of
D
D
,
the averageout degree, is more close to zero.
For brevity, here the process of calculations are
skipped. Results are shown in Table 3 below:
Note that in the gain realm, the averageout degree,
Copyright © 2011 SciRes. ME
Y. C. ZOU ET AL.381
the , as is shown by Table 11, of the test bears a
value of great bigger magnitude than that of the loss
realm, consistent with our arguments earlier in this sec-
tion that the use of sample average in the gain domain
here might have averaged out not only unwelcome vola-
tility, but also individual divergences. However, it is still
premature for us to assert that the use of sample average,
the uncategorized model, is undesirable, because to do
that one need a benchmark
D
D
level upon with one
could judge the desirability of the use of uncategorized
sample average, which needs further studies.
6. Conclusions
There is no such a single function that can represent the
volatile and divergent preferences of mankind. The
prospect theory value function mistakably links its two
branches in gain and loss domains together, which, albeit
right intuitively, falls in the pit in the majority heuristic.
Also, using sample mean in analyzing data in the studies
of behavioral economics could be perilous because this
method treats the whole sample as an individual and
willy-nilly presumes that the average preference could
stand for the whole, while in studying the behavioral
facts of human preferences, individual divergences are
pivotal and worth deep investigation. We proposed in the
paper an approach toward modeling and data analyses,
which involves categorizing the sample into different
preference families and developing the sub-regression
models. We suggested that this approach is statistically
more reliable than the mere analysis of the sample means.
But we should also point out that our theory still is based
on experiments. It provides a way of revealing the fragil-
ity of uncategorized regression model in behavioral
studies, but it is incapable of testing the validity of using
Table 11. DF, D, and Δ
D
values of the test.
DF uncategorized DF categorized
Gain-Arch 2.030457834 3.214633228
Gain-Bounce 17.76499969 19.21976959
Gain-General 10.02344834 12.43379891
D 1.324406768 1.432155959
D
0.107749191
Loss-Neutral 33.68938576 4.105989609
Loss-Seeking 82.30766511 18.68931585
Loss-General 61.35279015 154.5656861
D 3.229913867 3.230015403
D 0.000101536
experiments in behavioral economics, which, we believe,
calls for further philosophical discussion.
7. Reference
[1] Kahneman, Daniel. and Tversky, Amos., “Prospect The-
ory: An Analysis of Decision under Risk”, Econometrica,
Vol. 47, No. 2, 1979, pp. 263-92
[2] Odean, Terrance., “Are Investors Reluctant to Realize
Their Losses”, The Journal of Finance, Vol. 53, No. 5,
1998, pp. 1775-98
[3] Thaler, Richard H., and Johnson, Eric J., “Gambling with
the House Money and Trying to Break Even: The Effects
of Prior Outcomes on Risky Choice”, Management Sci-
ence, Vol. 36, No. 6, 1990, pp. 643-60
[4] Thaler, Richard H., “Mental Accounting and Consumer
Choice”, Marketing Science, Vol. 4, No.3, 1985, pp.
199-214
[5] Kahneman, Daniel., and Riepe, Marke W. “Aspects of
Investor Psychology: Beliefs, preferences, and biases in-
vestment advisors should know about”, Journal of Portfo-
lio Management, Vol. 24, No. 4, 1998, pp. 67-91
[6] Kachelmeier, Steven J. and Shehata, Mohamed., “Exam-
ining Risk Preferences Under High Monetary Incentives:
Experimental Evidence from the People’s Republic of
China”, The American Economic Review, Vol. 82, No. 5,
1992, pp. 1120-41
[7] Tversky, Amos. and Kahneman, Daniel., “Advances in
Prospect Theory: Cumulative Representation of Uncer-
tainty”, Journal of Risk and Uncertainty, No. 5, 1992, pp.
297-323
[8] Staw, B. M. “The Escalation of Commitment to a Course
of Action,” Acad. Management Rev., Vol. 6, No. 4, 1981,
pp. 577-87
[9] Laughhunn, D. J. and J. W. Payne. 1984. “The Impact of
Sunk Outcomes on Risky Choice Behavior,” INFOR
(Canad. J. Per. Res. and information Processing), No. 22,
1984, pp. 151-81
[10] Tversky, Amos. and Kahneman, Daniel., “Judgement
under Uncertainty: Heuristics and Biases”, Science, New
Series, Vol. 185, 1974, pp. 1124-31
[11] Fredrick, Shane., “Cognitive Reflection and Decision
Making”, The Journal of Economic Perspectives, Vol. 19,
No. 4, 2005, pp. 25-42
[12] Taleb, Nassim N. “Fooled by Randomness: The Hidden
Role of Chance in Life and in the Markets”, W. W. Nor-
ton, ISBN-10: 1587990717, October 2001, pp. 182-83
[13] Becker, G. M., DeGroot, Morris H., and Marschak, Jacob,
“Measuring Utility by a Single-Response Sequential
Method,” Behavioral Science, No. 9, 1964, pp. 226-32.
[14] Brislin, Richard W., “Back-Translation for Cross-Cul-
tural Research,” Journal of Cross-Cultural Psychology,
No. 1, 1970, pp. 185-216
[15] Etchart-Vincent, N. and l’Haridon, O., “Monetary Incen-
tives In the Loss Domain and Behavior Toward Risk: An
Experimental Comparison of Three Reward Schemes In-
Copyright © 2011 SciRes. ME
Y. C. ZOU ET AL.
Copyright © 2011 SciRes. ME
382
cluding Real Losses”, Journal of Risk and Uncertainty, No. 42, 2011, pp. 61-83
Appendix: Replication Experiment of Pros-
pect Theory Problems
PROB1: In addition to whatever you own, you have been
given 1,000. You are now asked to choose between
A: 50% chance to win 1,000 B: a sure gain of 500
PROB2: In addition to whatever you own, you have
been given 2,000. You are now asked to choose between
A: 50% chance to lose 1,000 B: a sure loss of 500
PROB3: Choose between:
A: 25% to win 6,000
B: 25% to win 2,000 and another 25% to win 4,000
PROB4: Choose between:
A: 25% to lose 6,000
B: 25% to win 2,000 and another 25% to lose 4,000