 Open Journal of Modern Linguistics  2012. Vol.2, No.4, 159-169  Published Online December 2012 in SciRes (http://www.SciRP.org/journal/ojml)                    http://dx.doi.org/10.4236/ojml.2012.24021   Copyright © 2012 SciRes. 159  The Effect of Interviewers’ and Respondents’ Accent and Gender  on Willingness to Cooperate in Telephone Surveys  Marie-José Palmen, Marinel Gerritsen, Renée van Bezooijen  Department of Business Communication Studies, Radboud University, Nijmegen, The Netherlands  Email: M.Gerritsen@let.ru.nl    Received August 29th, 2012; revised September 24th, 2012; accepted October 3rd, 2012  This article presents two real-life experiments that investigate whether an interviewer’s accent and gender  combined with a respondent’s accent and gender have an impact on telephone survey cooperation rates.  Expectations were based on the authority and liking principles of the compliance theory. In Study 1, 12  standard-speaking interviewers (6 men, 6 women) and 12 interviewers with a regional accent (6 men, 6  women) called 1925 male and female respondents (speaking either the standard or the regional variety).  In Study 2, a female interviewer who mastered the standard accent and the regional variety, called 120  respondents from the same categories as in Study 1. The expectations were not confirmed. Interviewers  with authority (male, speaking standard Dutch) had no more success than interviewers with less authority  (female, speaking a regional accent), and agreement of gender and accent between interviewer and re-  spondent had no impact on the level of cooperation of the respondents. The results seem to indicate that it  is not necessary for research bureaus to reject potential employees with a regional accent or with a less  authoritative voice, and that they do not need to make an effort to match interviewers and respondents in  characteristics such as gender and accent.    Keywords: Response Rate; Cooperation; Authority Principle; Liking Principle; Agreement Interviewer  Respondent; Selection Interviewers  Introduction  Since the 1970s, research bureaus all over the world have  observed serious decreases in response rates (Steeh, 1981; De  Heer, 1999; De Leeuw & de Heer, 2002; Couper & De Leeuw,  2003; Curtin et al., 2005; Groves, 2011), especially in metro- politan areas (Steeh et al., 2001; Feskens et al., 2007). This is  alarming in view of the possible consequences for the gener- alizability of the results of surveys because nonrespondents  may differ systematically from respondents and consequently  increase the potential for bias due to nonresponse error (Groves  & Couper, 1996a). This may lead, for instance, to faulty policy  decisions by politicians. Visscher (1999a, 1999b) gives an ex-  ample of the detrimental effect that nonresponse may have. In a  study into living conditions and political interests, nonresponse  among people with a low level of education was twice as high  as among people with a high level of education. The results,  therefore, showed a distorted picture.   There is no simple relationship between a nonresponse rate  and nonresponse bias. Nonresponse does not necessarily lead to  nonresponse bias, and there is neither a minimum response rate  below which a survey estimate is necessarily biased nor a re-  sponse rate above which it is never biased (Curtin et al., 2000;  Keeter et al., 2000; Bethlehem, 2002; Merkle & Edelman, 2002;  Groves, 2006; Groves & Peytcheva, 2008; Stoop et al., 2010).  Stoop (2005) states that the hunt for the last respondents is  becoming increasingly difficult and costly, and may not be  altogether effective, since it does not necessarily decrease se- lective nonresponse (Van Ingen et al., 2009) and may even  increase nonresponse bias. An exit poll experiment by Merkle  et al. (1998) in the US showed, for example, that a pen incen- tive increased response rates, but also led to higher response  bias, since it only increased the response among Democratic  Party voters and not among Republicans.    Problems associated with the increasing nonresponse rates  exercise minds all over the world as shown by the fact that in  2011 the 21st anniversary of the International Workshop on  Household Survey Nonresponse was held. This workshop has  been held every year, and in 1999, culminated in the large In-  ternational Conference on Survey Nonresponse (1999). A se-  lection of the papers from this conference was published by  Groves et al. (2002). Another indication is that special issues on  nonresponse have been published (i.e., the Journal of Official  Statistics of June 1999, 2001 and 2011, the Journal of the Royal  Statistical Society: Series A (Statistics and Society) of July 2006,  the  Public Opinion Quarterly of October 2006) and that an  avalanche of articles on reducing nonresponse have appeared:  among others, Singer et al. (1999), Groves & McGonagle  (2001), Feskens et al. (2007), de Leeuw et al. (2007), Snijkers  et al. (2007), Durrant et al. (2010), Wenemark et al. (2011).  These publications focus on three aspects (Singer, 2006): a)  compensating for nonresponse by imputation and/or weighting;  b) measuring nonresponse bias; and c) reducing nonresponse.  The research reported in this article is an attempt to contribute  to the latter.  Two types of nonresponse can be distinguished: the part of  the target group that is not reached (noncontacts), and those  people who are reached but refuse to take part in the interview  (refusals). Nonresponse in the second sense; i.e. refusals, is the  topic of our studies. Singer (2006) shows that refusals contrib- ute more to the decrease of response rates than noncontacts; she  suggests that investigation of the interaction between inter- viewer and respondent at the initial contact might aid in gaining   
 M.-J. PALMEN  ET  AL.  insight as to how to reduce refusals, a plea made earlier by  Groves et al. (1992), Groves & Couper (1996b) and later by  Groves et al. (2008). Examples of studies of such interactions  are Maynard & Schaeffer (1997), Snijkers et al. (1999), Groves  & McGonagle (2001), Dijkstra & Smit (2002), Hox & de  Leeuw (2002), and Pondman (1998). The latter, for example,  shows that interviewers who react to externally attributed re- fusals such as “I have no time because I have just cooked our  dinner,” with participation-directed persuasion strategies such  as “Bon appétit. May I call you back after dinner?” have fewer  refusals than interviewers who react with refusal-directed  strategies, “Oh, of course, dinner can’t wait.” Other aspects of  the initial contact that could have impact on refusals are voice  characteristics of the interviewer (Groves, 2008). In the present  article, we used two real-life experiments to investigate whether  interviewers’ accent and gender in combination with respon- dents’ accent and gender had an impact on the refusal rate in  telephone surveys. An overview of the multitude of (socio)  linguistic and socio-psychological factors related to refusals in  telephone surveys can be found in Pondman (1998), Palmen  (2001), and Groves (2006).    In the next section, we discuss the theoretical backgrounds of  survey participation using a socio-psychological model called  compliance theory, and subsequently we present the expecta-  tions formulated regarding the impact of interviewers’ gender  and accent on the refusal rate. For the purpose of our study—  gaining insight into the impact of gender and accent on will- ingness to participate in a survey—a more detailed instrument  than usual for establishing the refusal rate had to be developed,  a cooperativeness scale. We discuss this instrument in the fol- lowing section. We then present the design and the results of  two real-life experiments. In the last section, we discuss our  results and outline the practical implications they might have  for research bureaus that conduct telephone surveys.  Theoretical Background: Compliance Theory,  Voice, and Refusals   Compliance Theory is a socio-psychological theory that de-  scribes the information people are guided by in forming opin-  ions, and the processes underlying the formation of those opin-  ions. Since the refusal rate is determined by the respondent’s  willingness to participate, this theory can be used to identify the  decision-making processes that make some respondents more  willing to participate in surveys than others are.    According to Groves et al. (1992), there are six psychology-  cal principles by which people are guided as they decide  whether to comply with a request. These six principles are au-  thority, liking, scarcity, consistency, reciprocation, and social  validation. In our study, authority and liking seem to be most  important; for that reason, we limit the following to a brief  discussion of the possible impact of these two principles on  refusal during initial contact between interviewer and respon- dent. For a detailed description of the relationships between the  other principles and survey participation, we refer to Palmen  (2001) and Gerritsen & Palmen (2002), and for empirical tests  of these principles, to Dijkstra & Smit (2002), Hox and de  Leeuw (2002) and Van der Vaart et al. (2005).  The authority principle has to do with the fact that people are  more inclined to comply with a request from someone whom  they regard as a legitimate authority. In telephone surveys, this  usually involves the well-known name and reputation of the  research bureau. Usually, (semi-) government institutions and  large market research bureaus have such a reputation. The for-  mer also have the advantage of an aura of reliability; for exam-  ple, official agencies that carry out censuses, such as the US  Census Bureau or the Statistisches Bundesamt in Germany.  Whether an interviewer has authority depends not only on the  organization for which he works, but also on personal aspects.  According to the authority principle, respondents are likely to  be more willing to comply with a request from an interviewer  with authority than from an interviewer with less authority.  Liking means that people are more inclined to comply with a  request from a person they like. In a survey situation, this  means a person will more likely participate when the request to  participate is made by an interviewer whom he finds sympa- thetic. Whether this is the case depends on many factors, but  according to Groves et al. (1992), the chances that the respon- dent will like the interviewer increase when they are more alike;  i.e., in terms of gender, age, social and regional background,  personality, or language use. They call this agreement. We find  this agreement concept also in other socio-psychological theo- ries like Byrne’s (1971) similarity attraction hypothesis. The  results of the study of Durrant et al. (2010) confirm the effect of  the agreement concept to a certain extent: interviewers and  respondents who shared attributes, especially educational back- ground, tended to produce higher cooperation rates. According  to the liking principle, respondents should be more willing to  comply with a request from an interviewer with whom they  share characteristics than from an interviewer with whom they  have little in common.  The decision to refuse or comply with the request to partici- pate in cold-call telephone surveys is usually made during the  first few sentences spoken by the interviewer (Oksenberg &  Cannell, 1988; Maynard & Schaeffer, 1997; Houtkoop-Steen-  stra & Van den Bergh, 2000) and this suggests that the decision  to refuse an interview is taken along the peripheral route of  decision (Cacioppo & Petty, 1982; Van der Vaart et al., 2005).  For that reason, it is plausible that the interviewer’s voice and  the information deduced from it by the listener plays a primary  role in the listener’s evaluation of the interviewer per the au- thority and the liking principles. Indications that the voice of  the interviewer indeed plays an important part in the decision of  the respondent to take part in a telephone survey are found in a  number of studies. Oksenberg et al. (1986) and Oksenberg &  Cannell (1988) found that interviewers rated by experts as hav- ing relatively high pitch, great variation in pitch, loudness, fast  rate of speaking, clear pronunciation, speaking with a standard  American accent, and perceived as sounding competent and  confident had fewer refusals than interviewers with the opposite  characteristics. The studies of Van der Vaart et al. (2005) and  Groves et al. (2008) partly supported this. Their study showed  that louder sounding and higher pitched voices were indeed  correlated with lower refusal rates, but faster speaking and be- ing judged as more confident were not. Since findings of the  effect of interviewers’ voice characteristics on refusal rates  were not consistent over models, Groves et al. (2008) argue for  further research, especially into the interaction of interviewers  and respondents. One of the lines of research they do not men- tion, but that seems important to us in light of the liking and the  authority principle, is the role of voice characteristics of the  respondent.   Coupland & Giles (1988) apply the agreement concept from  the liking principle to voice: the more people speak in a similar  Copyright © 2012 SciRes.  160   
 M.-J. PALMEN  ET  AL.  way, the more they like each other. In the context of a tele- phone survey, this could imply that agreement in voice might  decrease the refusal rate.  There might also be a relationship between voice and the au- thority principle; low voices, for example, have more authority  than high voices (Tielen, 1992; Van Bezooijen, 1995; Biemans,  2000). In light of the authority principle, we can expect that  interviewers with voice characteristics that are associated with  authority will have fewer refusals than interviewers without  those voice characteristics.    An interviewer’s voice comprises four elements: accent (re- vealing a possible regional and/or social origin), voice height  (revealing gender), voice quality (i.e., personal voice “colour”),  and prosodic characteristics (speech rate, etc.). We examined  the effect of accent and gender of both interviewer and respon- dent on the refusal rate. We chose these two factors because  they are easily noticed by the respondent, and because it is rela- tively easy to collect information about gender and accent from  refusers based on their voices.    Expectations  Gender  According to the authority principle, interviewers with more  authority should have a higher response rate than interviewers  with less authority. There is a very clear relation between au- thority and gender: men have more authority than women (Hol-  mes & Meyerhoff 2005; Barret & Davidson 2006), and male  voices have more authority than female voices (Tielen, 1992;  Van Bezooijen, 1995; Biemans, 2000). Moreover, Baruffol et al.  (2001) and Hansen (2007) found that male interviewers have  lower refusal rates than female interviewers in telephone sur- veys. This leads to Expectation 1.  Expectation 1: Male interviewers encounter higher response  rates than female interviewers.  According to the liking principle, agreement between the in- terviewer and the respondent would lead to a lower refusal rate.  As regards gender, this implies that male interviewers who call  male respondents should be more successful than male inter-  viewers who call female respondents, and that female inter-  viewers who call female respondents should be more successful  than female interviewers who call male respondents. The study  of Durrant et al. (2010) presents some support for this since  there was a tendency for female respondents to be more likely  to cooperate than male respondents were when the interviewer  was a woman. The liking principle leads to Expectation 2.    Expectation 2: The response rate is higher when interviewers  call a person of the same sex than when they call a person of  the opposite sex.  Accent  We use the term accent to designate a non-standard variety  of a language; for instance, a socially determined variety (so- ciolect), a regional variety (dialect), or an ethnic variety (eth- nolect). According to the authority principle, interviewers with  more authority should have a lower refusal rate than interview- ers with less authority. There is a clear relationship between  authority and accent. Much research shows that people who  speak the standard variety are viewed as more intelligent, more  self-confident, and more competent than people with a regional  accent. Standard speakers generally have higher status and  more authority than speakers with an accent (Cacioppo & Petty  1982; Milroy & Milroy, 1999; Heijmer & Vonk, 2002; Kra-  aykamp, 2005; Grondelaers et al., 2010). In the context of a  telephone survey, this could mean that interviewers speaking  the standard variety will have more success than those with a  regional accent. Oksenberg & Cannell (1988) also found this in  their US study. In the Netherlands, survey agencies usually  adopt this view and ask for interviewers who speak standard  Dutch in their employment advertisements (Eimers & Thomas  2000: p. 28). The use of a regional variety would be perceived  as less professional and would, therefore, result in a higher  refusal rate. This leads to Expectation 3.  Expectation 3: Interviewers who speak the standard variety  have a higher response rate than interviewers who have a re- gional accent.  According to the liking principle, agreement between inter-  viewer and respondent leads to lower refusal rates. Apropos of  accent, this implies that interviewers who use the standard vari- ety should have more success with respondents who also speak  the standard variety than with respondents who have a regional  accent, and interviewers with a regional accent should be more  successful with respondents who have the same accent than  with respondents who speak the standard language. The posi-  tive effect of agreement between people in a negotiation situa- tion is, for example, found by Mai & Hoffmann (2011) in their  studies of language use in service selling. When there was a fit  between the dialect of the salesperson and the dialect of the  client, purchase intention increased. This leads to Expectations  4a and 4b.  Expectation 4a: The response rate is higher when interview- ers who speak the standard variety of a language call respon- dents who speak the same standard variety than when they call  respondents with a regional accent.  Expectation 4b: The response rate is higher when interview-  ers with a regional accent call equally accented respondents  than when they call respondents who speak the standard vari-  ety.  Cooperativeness Scale   In order to gain a deeper insight into the linguistic aspects  that could play a part in refusals, we analyzed almost 1000  interviewer-respondent interactions recorded by an important  Dutch survey agency prior to our real-life experiments (cf. next  section). We discovered that the traditional classification as  applied to interview outcomes (refusal, appointment, or success)  was not adequate for our research into the effects of gender and  accent on respondents’ cooperativeness on initial contact be- tween interviewer and respondent. Respondents may be coop- erative without, in the end, being able to or allowed to partici- pate because they do not meet the selection criteria. For in- stance, a family member of the selected respondent answers the  phone and states that the selected respondent is not available.  This family member is cooperative in providing the information  that leads to selection, but he or she cannot take part in the sur- vey. Such calls are then written off as noncontacts. This is not a  problem for surveys, because there was indeed no contact with  the person who was selected for the survey. In order, though, to  be able to gain insight into the impact of gender and accent on  willingness to participate in a survey, a distinction has to be  made between respondents who are cooperative and respon- dents who are not. For this reason, we developed a scale of  Copyright © 2012 SciRes. 161   
 M.-J. PALMEN  ET  AL.  Copyright © 2012 SciRes.  162  0 20 40 60 80 100 12345 Score Cooperativeness   cooperativeness. The reactions of respondents (approximately  1000) to requests to participate in a survey could be classified  into 12 different reaction patterns (Palmen, 2001). These 12  reaction patterns were submitted to 22 communication special- ists who assigned to each situation a cooperativeness score  between 0 and 100. The higher the score assigned, the more  cooperative the respondent was considered to be in the reaction  described. The interrater reliability of the 22 raters was in ac- cordance with Kendall’s W 0.84 (p < .01). Figure 1 presents  the means for each of these twelve reaction patterns.  Figure 1.   Mean of the cooperativeness scores assigned by 22 subjects to  twelve reaction patterns (0 = not cooperative at all, 100 = very  cooperative).  Figure 1 clearly shows that some reaction patterns are very  close to each other. In order to determine which patterns dif- fered significantly, a one-way analysis of variance with a post  hoc analysis (Tukey HSD) was performed with reaction pattern  (1 - 12) as an independent variable and cooperativeness as a  dependent variable. The results are presented in Table 1.     Table 1.  Differences between the twelve reaction patterns.    Category1 2 3 4 5 6 7 8 91011 1             2 -            3 - -           4 * * *          5 * * * -         6 * * * * -        7 * * * * - -       8 * * * * * - -      9 * * * * * * * *     10 * * * * * - - - *    11 * * * * * * * * - *   12 * * * * * * * * * * *  Table 1 shows clearly that a number of categories do not  differ significantly in cooperativeness. For example, the sub- jects did not make a distinction in cooperativeness between a  respondent who interrupts the interviewer and slams the re- ceiver down (1); a respondent who interrupts the interviewer  and terminates the call decently (2); and a respondent who lets  the interviewer finish his introductory phrases, but then frankly  finishes the call (3). The post hoc (Tukey HSD) analyses show-  ed that regarding cooperativeness, the original 12 reaction pat-  terns could be reduced to five categories. Table 2 shows these  five categories (column 1), their cooperativeness with an exam- ple (columns 3 and 4), and which of the 12 reaction patterns  were taken together (column 2).  The scale’s five cooperativeness categories are represented  graphically in Figure 2. The numerical value assigned to each  of the five categories is the mean of the initial (non-compressed)  twelve categories. This informs us of the order: a respondent  who is given a 3 for cooperativeness is more cooperative than a  respondent who is given a 2; 2 is more cooperative than 1, etc.  Figure 2 shows that the distance between the five categories  are nearly similar; the lines that connect the points have almost  the same slopes; and the distance between 2 and 3 is similar to  the distance between 3 and 4. This is an indication that the co-  operativeness scale has interval features. According to Fienberg  (2007), the variable cooperativeness can, therefore, be treated  as an interval variable.  Note: *p < .05.    0 20 40 60 80 100 12 3 4 56 7 89101112 Category Cooperativiness   Instead of classifying a request as leading to success, refusal,  or an appointment (the classic tripartite system), we assigned a  cooperativeness score on a scale from 1 to 5. In the traditional   Figure 2.  The 5-point cooperativeness scale.    Table 2.  The cooperativeness scale.  New score  Old scores Degree of cooperation Result Example (I = interviewer; R = respondent)  1 1, 2, and 3 None Flat refusal, without a prospect of changeR: No, I don’t want to participate.  2 4 and 5 Minimal Refusal with explanation, possibly  convertible  R: Well, I really don’t feel like it, and those surveys always  take up so much time…  I: Oh, but this is a very interesting study, and it takes onl   five minutes.  R: No, thanks.  3 6, 7, 8, and 10 More than minimal Helpful, but with obstacles  R: I have to leave the house now.  I: Can I call you back some other time?  R: Yes, sure.  4 9 and 11 Less than maximal Helpful  R: How long will it take?  I: Five minutes at the most.  R: Go ahead then.  5 12 Maximal Full cooperation I: Would you answer some questions about that?  R: Yes, sure.   
 M.-J. PALMEN  ET  AL.    classification system, 1 and 2 would be classified as refusal, 3  as an appointment, and 4 and 5 as success.  The cooperativeness scale makes different degrees of coop-  erativeness—after initial contact is made—more visible than  the traditional classification system does, but it is definitely not  better than the traditional classification scale for reporting the  response rates of surveys. The cooperativeness scale should be  seen as a research instrument to be used to gain insight into  which characteristics of an interviewer in combination with  which characteristics of a respondent might play a part in will-  ingness to take part in a survey after initial contact between  interviewer and respondent has been made.     We defined our expectations in (3) in terms of response rates.  To assess all aspects of respondents’ cooperativeness, we de-  veloped the Cooperativeness Scale to replace the traditional  refusal or response rate. To compare with our expectations, the  term response rate should be read as cooperativeness score.   Design and Results of Two Real-Life   Experiments   To examine and verify our expectations, we carried out two  studies: a real-life study in which 24 interviewers (divided  equally into four categories: male/female and standard/regional  accents) called 1925 respondents from the same categories  (Study 1), and a matched guise study with one female inter-  viewer who mastered both linguistic varieties; she called 120  respondents from the same categories as in the first study  (Study 2).  Study 1: Real- Li fe  Study into the Effects of Gender   and Accent on Cooperativeness  In this section, we report on the experiment designed to sys- tematically investigate the cooperativeness scores achieved  between four different types of interviewers and respondents:  male and female interviewers speaking either standard Dutch or  a regional variety of Dutch called respondents in the same four  categories. On the basis of their cooperativeness scores, we  were able to determine whether certain respondent groups were  more cooperative than others, and whether their cooperative-  ness depended on the interviewer’s accent and gender.  Accents  Two accents were investigated in this study: standard Dutch  and Dutch with a Limburgian accent (a regional accent from the  south of the Netherlands). The Limburgian variety was selected  for a number of reasons. Just as standard Dutch, it is a clearly  recognizable variety of Dutch (Grondelaaers & Van Hout,  2010), and according to dialectologists it is the most deviant  compared to standard Dutch (Daan & Blok, 1969; Hoppen-  brouwers, 2001). It is also a variety that is viewed by its speak-  ers and by speakers of standard Dutch as enjoyable and pleasant  (Grondelaers et al., 2010; Grondelaers & Van Hout, 2010). It  seemed, therefore, to be an interesting variety to test the liking  principle. If respondents with a regional accent do not value  their own variety positively, they will be less likely to be will-  ing to comply with a request from an interviewer from the same  area than with a request from a speaker of the standard variety.   Interviewers  The interviewers were screened by two experienced socio- linguists to determine the pleasantness of their voices and par- ticular speech characteristics such as loudness, pitch, and rate  of speaking in order to avoid that interviewers would differ  with regard to voice characteristics that affect refusals (Van der  Vaart et al. 2005). Only interviewers with “normal”, non-devi-  ant voices were selected. In addition, the sociolinguists deter- mined whether the speech of the interviewers could be qualified  as one of our research categories: clearly male or clearly female,  and standard Dutch or Dutch with a Limburgian accent. For the  Limburgian interviewers, this meant that their speech had to be  Dutch in both grammar and vocabulary, but with a clear Lim-  burgian colour in their pronunciation. The twelve Limburgian  interviewers selected had lived in Limburg for most of their  lives, and their mother tongue was a Limburgian dialect. For  the standard Dutch speakers, it was important that their speech  did not reveal any specific regional origin; they had to speak  neutral Dutch without a regional color. The accents of the in- terviewers with a Limburgian accent and of the interviewers  with a standard Dutch accent were easy to understand for all  speakers of Dutch.  The 24 selected interviewers (6 standard Dutch men, 6 stan-  dard Dutch women, 6 Limburgian men, and 6 Limburgian wo-  men) were all younger than 25 years. They represented a rather  homogeneous group, and corresponded to the average survey  workforce in the Netherlands. Twenty-two of them were ex-  perienced poll-takers or telemarketers. The two interviewers  who did not have this experience could easily get work in these  jobs, according to the sociolinguists who screened their lan-  guage; their voices and way of speaking did not betray that they  were inexperienced in these jobs.    Before analyzing the data any further, we established whe-  ther there were major differences between individual inter-  viewers or whether it was possible to consider same-sex, same-  accent interviewers as one homogeneous group. To that end, we  compared the mean of the cooperativeness scores the inter-  viewers obtained with each group of respondents. The means of  the cooperativeness scores were very similar among interview-  ers within a group. One-way analyses of variance confirmed  that the differences were not significant. The results of these  analyses can be found in Table 3. In our subsequent analyses,  therefore, we combined the interviewer results per group.      Table 3.  Interviewers: One-way analyses of variance (per respondent group).  Respondents N  Interviewers  (N = 6 for every cell) F (df1, df2)p  St. Dutch male101 St. Dutch male 1.14 (5, 95).34   104 St. Dutch female 1.08 (5, 98).38   102 Limb. male 1.28 (5, 96).28   96 Limb. female 0.40 (5, 90).85  St. Dutch  female 135 St. Dutch male 0.50 (5, 129).77   132 St. Dutch female 0.12 (5, 126).99   108 Limb. male 0.55 (5, 102).74   119 Limb. female 0.62 (5, 113).68  Limb. male 120 St. Dutch male 0.71 (5, 114).62   117 St. Dutch female 0.28 (5, 111).92   105 Limb. male 0.56 (5, 99).73   109 Limb. female 0.78 (5, 103).56  Limb. female150 St. Dutch male 1, 55 (5, 144).18   159 St. Dutch female 0, 63 (5, 153).68   140 Limb. male 1, 40 (5, 134).23   128 Limb. female 1, 72 (5, 122).13  Copyright © 2012 SciRes. 163   
 M.-J. PALMEN  ET  AL.  Respondents and Pr ocedure   In order to ensure that the respondents were as similar as  possible in all aspects other than accent and gender, we selected  four medium-sized cities (between 50,000 and 100,000 inhabi-  tants) in the Netherlands with similar socio-economic charac-  teristics and infrastructure based on the statistical community  data of the national statistical agency. Two Limburgian cities  (Roermond & Sittard) were paired with two cities from the  region in which the most standard Dutch is spoken according to  Daan & Blok’s (1969) and Hoppenbrouwers’ (2001) map (Ri-  jswijk & Velsen). For each of the four cities, we selected the  phone numbers of 1500 private persons by using the elec-  tronic directory. Each interviewer was provided with a mixed  list of 250 phone numbers obtained from the four cities. The  interviewers alternately called standard Dutch-speaking areas  and Limburgian areas. When the phone was answered by a  respondent, the interviewer used the following introduction  (translated from Dutch): “Good evening sir/madam, this is  [name of the interviewer] of BC Research. We are currently  conducting a study to assess people’s opinions of certain topics  relating to the Dutch language. Could I ask you a few questions  on that?” Our survey agency was called “BC Research”, BC  standing for Business Communication, the department respon-  sible for the study. To investigate the impact of gender and  accent on cooperativeness, we had to ask some questions of  cooperative respondents. Since we did not want to ask ques-  tions without a purpose, we asked colleagues at the Radboud  University to suggest questions for which they would like an- swers. A survey created by Smakman (2006) was ideal for this  purpose as it comprised questions on the position of Dutch  attitudes towards standard Dutch and regional varieties, what  standard Dutch really is, and who uses that variety. We be-  lieved this topic would be equally interesting for all four cate-  gories of respondents, and would consequently avoid a nonre-  sponse bias (Keeter et al., 2006).    The survey was carried out on Monday through Thursday  nights, between 6.30 and 9.30 p.m. All interviewers worked the  same days and the same hours over the survey period. Inter-  viewers were instructed to try to persuade hesitant respondents  to participate and to give them additional information, as would  be done by a real survey agency. They were told not to try any  further in the case of firm, radical refusals without any cues for  further discussion.    For every call, the interviewers noted whether the telephone  was picked up or not. When the phone was picked up, they  noted whether they had spoken to a man or a woman, and the  outcome of the call (success, refusal, or appointment). When  the phone was not picked up, the interviewer made a note of the  telephone number, and phoned that number three other times on  other days. The interviewers kept track of the number of re-  spondents in each of the four groups, and they were instructed  to continue until they had spoken to at least 25 persons in each  group: 25 men and 25 women from the standard Dutch area,  and 25 men and 25 women from the Limburgian area. Based on  Cohen (1992), we needed an α-level of .05, a power of .8, and  16 groups with 15 respondents per cell to detect a large effect  size. We decided to choose a large effect size, because small  effect sizes are, in our view, not relevant for practical use in  telephone surveys. The interviewers were asked to conduct 25  conversations in order to leave enough margin to be able to  leave out of consideration respondents with an accent other  than the two varieties investigated, children, and clearly dis- abled persons (hard of hearing, mentally addled, or handi- capped). The number of 15 was achieved for every group; when  there were more than 15 respondents, the additional data were  also entered into the analysis.    Since the aim of our study was to determine whether inter-  viewer accent and gender in combination with respondent ac-  cent and gender had an impact on the cooperativeness rate in  telephone surveys, respondents’ gender and accent had to be  determined. All calls—whether they resulted in success, re-  fusal, or appointment—were taped and assessed for gender and  accent by two experienced sociolinguists. Assessing gender was  not problematic. Accent was assessed on a scale of 1 to 10:1  meaning Dutch with a very strong regional colour, and 10  meaning perfectly neutral Dutch. Respondents who were attrib- uted 7 or higher on this scale were qualified as speakers of  standard Dutch for the purposes of this study. Respondents who  were attributed 5 or less, and who were identifiable as Limbur-  gian, were assigned to the group of Limburgian respondents.  In the end, 1925 calls were examined for the impact of gen-  der and accent on cooperativeness. According to the traditional  response classification scale, the results were 589 (31%) suc-  cessful interview, 150 (8%) appointment, and 1187 (62%) re-  fusal. Since we were only interested in whether interviewer and  respondent gender and/or accent had an impact on respondents’  cooperativeness, the number of non-contacts was not relevant  for our study, and for that reason they were not noted.    Results Study 1  Table 4 shows the results of a 2 × 2 × 2 × 2 analysis of vari-  ance with the cooperativeness scores as the dependent variable,  and the following four factors: interviewer gender (male, fe-  male), interviewer accent (standard Dutch, Limburgian), re-  spondent gender (male, female), and respondent accent (stan-  dard Dutch, Limburgian). Table 5 presents the means and  standard deviations for each of the respondent groups.     Table 4.  Interviewers’ and respondents’ gender and accent: results of analysis of  variance.      F  df = 1,1909p  Expectations    Gender interviewer (Expectation 1) 1.70 .19  Gender interviewer*gender respondent  (Expectation 2) .011 .91  Accent interviewer (expectation 3) 2.85 .09  Accent interviewer*accent respondent   (Expectation 4) 1.31 .25  Possible interactions    Gender interviewer*accent respondent .75 .38  Gender interviewer*accent interviewer*gender  respondent .35 .55  Gender interviewer*accentinterviewer*accent    respondent .32 .57  Gender interviewer*gender respondent*accent  respondent .05 .82  Gender interviewer*accent interviewer*gender  respondent * accent respondent .00 .98  Accent interviewer*gender respondent 1.44 .23  Accent interviewer*gender respondent*accent  respondent 1.64 .20  Copyright © 2012 SciRes.  164   
 M.-J. PALMEN  ET  AL.  Copyright © 2012 SciRes. 165    Table 5.  Mean cooperativeness scores (M) and standard deviations (SD) for all respondent groups, per interviewer group (N = 1925) (1 = no cooperation, 5 =  full cooperation).   Respondents   St. Dutch male St. Dutch female Limb. Male Limb. Female Total  Interviewers n  M  SD n M  SD n M  SD n M  SD n M  SD  St. Dutch male 101 2.28  1.56 135 2.27  1.52 120 2.44  1.56 150 2.22  1.42 506 2.30  1.51  St. Dutch female 104 2.33  1.58 132 2.35  1.57 117 2.50  1.57 159 2.38  1.55 512 2.39  1.56  Limb. Male 102 2.48  1.58 108 2.55  1.59 105 2.22  1.50 140 2.44  1.48 455 2.42  1.53  Limb.  Female 96 2.54  1.61 119 2.47  1.58 109 2.45  1.64 128 2.60  1.63 452 2.52  1.61  Total 403  2.40  1.58 494 2.40  1.56 451 2.41  1.57 577 2.40  1.52 1925 2.40  1.55    Expectation 1, that male interviewers encounter higher co-  operativeness rates than female interviewers, was not confirmed.  There was no significant main effect for gender of the inter-  viewer on the cooperativeness rate. Male interviewers (M =  2.36, SD = 1.52) had as much success as female interviewers  (M = 2.45, SD = 1.59). Nor was Expectation 2 confirmed; there  was no significant interaction effect between gender of the  interviewer and gender of the respondent. Male interviewers  had as much success with male respondents (M = 2.36, SD =  1.55) as with female respondents (M = 2.35, SD = 1.50), and  female interviewers had as much success with female respon-  dents (M = 2.45, SD = 1.58) as with male respondents (M =  2.46, SD = 1.59).    Expectation 3, that interviewers who speak the standard vari- ety achieve a higher cooperativeness than interviewers who  have a regional accent, was also not confirmed. There was no  significant main effect for the accent of the interviewer on the  cooperativeness. Interviewers speaking standard Dutch (M =  2.34, SD = 1.54) had as much success as interviewers speaking  Limburgian (M = 2.47, SD = 1.57).There was also no signify-  cant interaction effect between the accent of the interviewer and  the accent of the respondent. Interviewers speaking standard  Dutch had as much success with standard Dutch speaking re-  spondents (M = 2.31, SD = 1.55) as with Limburgian-speaking  respondents (M = 2.38, SD = 1.52), and Limburgian-speaking  interviewers had as much success with Limburgian-speaking  respondents (M = 2.44, SD = 1.58) as with standard Dutch-  speaking respondents (M = 2.51, SD = 1.56) (Expectation 4).  None of the expectations based on either the authority prin-  ciple (Expectations 1 and 3) or the liking principle (Expecta- tions 2, 4a and 4b) were confirmed. Although we had no ex- pectations regarding interactions of interviewers’ and respon- dents’ gender and accent—for example, that the cooperative- ness with male interviewers would be different than with fe- male interviewers for Limburgian respondents only—we test-  ed all possible interactions of interviewers’ and respondents’  gender and accent in the hope of gaining more insight into the  factors that play a role in cooperativeness. Table 4 shows there  are no significant interactions between interviewers’ and re- spondents’ gender and accent. Our results show clearly that in-  terviewers’ and respondents’ gender and accent did not influ-  ence cooperativeness. Our real life study among 1925 respon-  dents did not confirm any of our expectations; this led us to  perform a second experiment.   Study 2: Matched-Guise R eplication, On e Female   Interviewer with Two Accents   Table 4 showed no significant differences between the inter-  viewers in the level of success they obtained; all 24 interview-  ers achieved similar cooperativeness scores. One could raise the  objection that 24 interviewers could not possibly present iden-  tical behavior on all levels as they are human beings, not ma-  chines. Although they were instructed to observe the instruc-  tions strictly, small deviations may have occurred. In some  cases, for instance, the respondent’s argument that he or she  was probably too old to participate was not always handled in  the same way; e.g., the interviewer might forget to mention the  counter argument, or use different phrasing. Although these  deviations do not seem to have been systemic (every inter-  viewer had some interviews presenting slight differences), they  resulted in some (inevitable) variation in the data.   Also, the voices and speech characteristics of 24 interviewers  were not completely equal on all levels. Although two socio-  linguists assessed the interviewers’ voices prior to our study,  their voices were not identical. These differences may have  been minimal, or they may possibly have affected the inter- viewers’ scores, but did they even out across all interviewers?  In order to find out whether our results were clouded by such  voice and/or behaviour differences between interviewers, we  did a second real-life investigation in which we used a well-  tried method from sociolinguistics: the matched-guise tech- nique (Lambert, 1967; Giles, 1973).  In the matched-guise technique, a speaker is used who mas- ters two varieties of a language on a native speaker’s level: for  example, the standard variety and a dialect variety. Two speech  fragments are recorded by this perfectly bilingual person, one in  the standard language and one in the dialect variety. The frag- ments differ only regarding language variety; apart from that,  they are the same (i.e., in voice height, voice quality, and pro- sodic characteristics) because both fragments are uttered by the  same person. The fragments are presented to listeners who in- dicate their appreciation of the speaker in a scaled manner in  terms of intelligence, friendliness, status, etc. Since the frag- ments are uttered by the same speaker, differences in apprecia- tion of the speaker of the fragments can only be attributed to the  variety they speak. This matched-guise technique was found  suitable to investigate whether the results of the first study  could have been clouded by voice and/or behaviour differences  between interviewers. If a bilingual interviewer attained similar   
 M.-J. PALMEN  ET  AL.  results as the interviewers from Study 1, that would constitute a  confirmation of the findings of that study.  In regular matched-guise studies, every subject is exposed  twice to the speaker, once when the speaker uses one variety,  once when he uses the other (within-groups design). This was  not possible in our study, since it was not possible to call the  same respondent twice with the same request. For that reason,  we used a between-groups design: each respondent was con-  tacted once and heard one variety.    The female bilingual interviewer we found, according to two  sociolinguists, matched the criteria applied to the standard  Dutch and Limburgian interviewers in Study 1. Unfortunately,  no male interviewer who met these criteria could be found. As a  consequence, only the results for the female interviewers of  Study 1 could be verified: Expectations 3, 4a and 4b, and par-  tially Expectation 2.    The female bilingual interviewer contacted 120 respondents  equally divided among the same four categories (30 respon-  dents in each cell) and in the same way as was done in Study 1.  In this matched-guise study, not only respondents’ accents were  determined by two sociolinguists, but also the accent of the  bilingual interviewer, because there was a chance that she  would accidentally speak standard Dutch and indicate it as a  Limburgian accent, or the other way around.       Table 6 shows the results of a 2 × 2 × 2 analysis of variance  with the cooperativeness scores as the dependent variable and  the factors: interviewer accent (standard Dutch, Limburgian),  respondent gender (male, female), and respondent accent  (standard Dutch, Limburgian). Table 7 presents the means and  standard deviations for each of the respondent groups.  The part of Expectation 2 that could be tested, does the fe-  male interviewer have more success with female respondents  than with male respondents, was not confirmed. The female  interviewer had as much success with female respondents (M =  2.15, SD = 1.31) as with male respondents (M = 2.72, SD =  1.70).  Expectation 3 was not confirmed; when the interviewer used  standard Dutch, she did not have significantly more success (M  = 2.47, SD = 1.56) than when she used Limburgian (M = 2.40,  SD = 1.53). There was also no significant interaction between  the accent of the interviewer and the accent of the respondent;  for that reason, expectations 4a and 4b were not confirmed.  Agreement between interviewer and respondent in accent (M =  2.50, SD = 1.57 for Standard Dutch, and M = 2.60, SD = 1.61  for Limburgian) did not lead to more success than disagreement     Table 6.  Female interviewer’s accent and respondents’ gender and accent: re- sults of analysis of variance.     F  (df = 1,112)p  Expectations    Gender respondent (Expectation 2, partially) 4.08 .053  Accent interviewer (Expectation 3) .07 .81  Accent interviewer*accent respondent   (Expectation 4) .69 .41  Possible interactions    Accent interviewer*gender respondent 1.71 .19  Accent interviewer*gender respondent*accent  respondent .23 .64  Table 7.  Mean cooperativeness scores (M) and standard deviations (SD) for all  respondent groups, per interviewer group (n = 120, 1 = no cooperation,  5 = full cooperation).   Respondents   St. Dutch  Male  St. Dutch  Female Limb. Male Limb.  Female Total  Accent  emale  interviewer n M SD n M SD n M  SD n M SD n M SD St. Dutch 153,00 1.70152.00 1.31 15 2.87  1.85 15 2.00 1.14602.47 1.56 Limb. 152.20 1.57152.20 1.38 15 2.80  1.74 15 2.40 1.50602.40 1.53 Total 302.60 1.65302.10 1.32 30 2.83  1.76 30 2.20 1.32 120 2.43 1.54   did (M = 2.20, SD = 1.45 for a Limburgian interviewer and a  Standard Dutch respondent, and M = 2.43, SD = 1.57 for a  Standard Dutch interviewer and a Limburgian respondent).   None of the expectations based on either the authority prin- ciple (Expectation 3) or the liking principle (Expectations 2 and  4a and b) were confirmed. Table 7 shows that again there were  no interactions between the interviewer’s accent and the re- spondent’s gender and accent. The effect of the interviewer’s  accent on cooperativeness did not depend on the gender and/or  the accent of the respondent.    Conclusion and Discussion  We reported above on two studies into the possible impact of  interviewers’ and respondents’ gender and accent on willing- ness to cooperate in a telephone survey. In Study 1, twelve  standard-speaking interviewers (6 men, 6 women) and twelve  interviewers with a regional accent (6 men, 6 women) called  1925 male and female respondents (speaking either the standard  or the regional variety), to request them to take part in a tele- phone survey on the Dutch language. There were no significant  differences in cooperation between the groups: interviewers’  and respondents’ gender and accent did not affect cooperative- ness. For that reason, none of the expectations were confirmed.  The results for the female interviewers were verified and  confirmed in Study 2, a matched-guise study, featuring a fe- male interviewer who mastered both the standard accent and  the regional variety. She called standard-speaking and regional-  speaking male and female respondents, using the two varieties  alternately. She attained similar cooperativeness scores using  both varieties, and when calling either men or women. In this  study, there were also no significant differences between the  groups; therefore, none of the expectations were confirmed here  either. The number of participants per call was purposely based  on a large effect size. The reason for this choice was that de-  tecting medium or small effects on willingness to cooperate in a  telephone survey would have been less relevant.   In both studies, none of the expectations based on the au-  thority principle of the compliance theory were confirmed.  Male interviewers were not more successful than female inter- viewers (Expectation 1); interviewers who spoke standard  Dutch were not more successful than interviewers who had a  Limburgian accent (Expectation 3); nor were the expectations  based on the liking principle confirmed. Cooperativeness was  not higher in same-sex interviews than in interviews with a  person of the opposite sex (Expectation 2), and the coopera- Copyright © 2012 SciRes.  166   
 M.-J. PALMEN  ET  AL.  tiveness was not higher when interviewers called a person who  spoke the same language variety than when they called a person  who spoke another variety (Expectation 4).    Our results for the effect of gender based on the authority  principle, Expectation 1, are in line with those of Groves &  Fultz (1985), Gerritsen & Van Bezooijen (1996), and Pickery &  Loosveldt (1998), but they do not corroborate the results of  Baruffol et al. (2001), and Hansen (2007). They found that  male interviewers had higher response rates than female inter- viewers in telephone surveys. The difference between their  results and ours could result from the fact that the interviewers  in their studies had the ability to negotiate with the respondent  about taking part in the survey, whereas the interviewers in our  study had less opportunity because we instructed them not to  try any further in the case of firm, radical refusals. According to  Hansen (2007), the higher response rates of male interviewers  in their study may be a result of men being more impertinent  than women; male interviewers may, therefore, be more likely  than female interviewers to talk respondents into taking part in  an interview, in spite of an initial refusal.  Our results for the effect of gender based on the liking prin- ciple, Expectation 2, are not in line with Durrant et al. (2010)  who found in face-to-face interviews a tendency for female  respondents to cooperate more often than male respondents if  the interviewer is a woman. They argue that their result could  be due to either the liking principle or to the potential fear of  women for a male stranger. Since the liking principle explana- tion raises the question of why this principle does not work for  male interviewers with male respondents, the second explana- tion is more plausible. Moreover, the fear of a male stranger  explanation can account for the difference between our results  and those of Durant et al. (2010). Their results were based on  the response rate in face-to-face interviews, whereas ours were  based on telephone surveys. It is likely that the fear of a male  stranger is larger in face-to-face communication than in tele- phone communication.  Our results for the effect of the language variety used by the  interviewer, Expectation 3, are also not in line with other stud- ies. Oksenberg & Cannell (1988) found, consistent with Ex- pectation 3, that interviewers speaking the standard American  English had more success in the US than interviewers with a  regional accent. However, Walrave (1996) and Dehue (1997)  observed precisely the opposite in the southern part of the  Dutch-speaking language area, Flanders, where a dialect is  spoken that is viewed by its speakers as enjoyable and pleasant.  Walrave observed that a slight non-standard accent inspires  confidence on the part of the respondent and leads to a higher  response than the standard language, and Dehue states that  when a call led to a successful interview, more likely than not,  the interviewer had a regional accent. In our study, however,  both the standard Dutch interviewers and the Limburgian inter- viewers had a similar effect on cooperativeness with all re- spondents; however, we do not know whether our results can be  extrapolated to other varieties of Dutch. Limburgian was cho- sen because it is known as an enjoyable and pleasant accent,  and for that reason it seemed an excellent variety to test the  liking principle in telephone surveys. It could be, though, that  studies of other varieties would have been more in line with the  observations of Walrave (1996) & Dehue (1997), or with the  results of Oksenberg & Cannell (1988). This calls for further  research..  Our result that agreement between interviewer and respon- dent in language variety did not increase cooperativeness (ex- pectation 4a and b) is not consistent with what we expected on  the basis of the study of Mai & Hoffmann (2011) who found  that agreement between salesperson and consumer in language  variety increased purchase intention. Durrant et al. (2010) also  found that agreement between interviewer and respondent in a  number of attributes increases response rate, although they did  not study agreement in language variety. The difference be- tween their results and ours could again be due to the fact that  their results are based on face-to-face communication and ours  on telephone communication. It is plausible that the effect of  agreement is much higher in the former than in the latter. This  definitely also calls for further research.    We found no effect of the sociolinguistically important fac- tors of gender and accent on cooperation in the initial contact  between interviewer and respondent in telephone surveys.  Apart from the reasons mentioned above, it is also possible that  this is due to the fact there is a large difference between labo- ratory and real life. The many sociolinguistic studies on which  we based our expectations were carried out under circum- stances that were more or less controlled, and thus artificial. An  example of this kind of research is the study by Brouwer (1989)  in which subjects evaluated speakers’ accents and personality  on the basis of a tape recording of speech fragments read by  men and women speaking standard Dutch or Dutch with an  Amsterdam accent. Moreover, in these sociolinguistic studies,  the emphasis was generally on attitudes, and not on the effect  on actual behaviour. If people with a regional accent prefer a  speaker of their own variety (attitude), it does not necessarily  imply they are more likely to participate in a telephone survey  (behaviour) when requested to do so by an interviewer using  their regional variety than they would by an interviewer with an  accent different from theirs.    If the elements of daily life are eliminated, factors such as  gender and accent might play a role; however, in everyday  communicative situations, their influence seems to be overruled  by other factors. In a society known to be “sick of surveys”, the  respondents’ attitude towards telephone surveys could be a first  important interfering factor. Most respondents probably have a  pattern of response with which they react to requests to take  part in surveys, and voice characteristics of the interviewer do  not easily alter this basic attitude. Secondly, situational factors  will play a role in the decision to be cooperative or not: Does  the topic interest me? How do I experience surveys, generally?  And time factors, of course, also play an important part in the  decision to take part in a survey. Vercruyssen et al. (2011)  found that respondents with less free time in Flanders more  often decline to take part in a survey. Van Ingen et al. (2009),  however, show that there is no direct proportional relationship  between business and reluctance to take part in a survey; they  found that people who are busy in work, sport, and volunteer  projects are often more inclined to take part in a survey than  people who don’t have such activities. It is not plausible that an  interviewer’s characteristics will have an impact on these situ- ational factors. If the interviewer does not appear sympathetic,  this will probably have a negative effect, but a sympathetic  impression (for example, because of agreement in gender and/  or accent) will not automatically lead to cooperation. Our re-  sults seem to indicate that various factors that were carefully  eliminated in laboratory experiments are decisive in everyday  life. For communication research, this implies that data col-  lected in authentic communicative situations constitute a valu-  Copyright © 2012 SciRes. 167   
 M.-J. PALMEN  ET  AL.  able and necessary complement to data collected in laboratory  situations.  As the survey results are decisive, measures aimed at in-  creasing cooperativeness should focus on the way the fieldwork  is organized. It is important for research bureaus to know they  probably do not have to exclude potential employees with a  regional accent, or female interviewers (which could lead to a  considerable increase in the number of potential interviewers),  and that it is probably not necessary to make efforts to match  certain characteristics of interviewers and respondents.  REFERENCES  Barret, M., & Davidson, M. J. (2006). Gender and Communication at  Work. Aldeshot: Ashgate Publishing.  Baruffol, E. P. Verger, & Rotily, M. (2001). Using the telephone for  mental health surveys. An analysis of the impact of call rank, non-  response, and interviewer effect. Population, 56, 987-1010.  doi:10.2307/1534750  Bethlehem, J. G. (2002). Weighting nonresponse adjustment based on  auxiliary information. In R. M. Groves, D. A. Dillman, J. L. Eltinge,  & J. A. Roderick (Eds.), Survey Nonresponse (pp. 275-288). New  York: Wiley.  Biemans, M. (2000). Gender variation in voice quality. Utrecht: Lot.    Brouwer, D. (1989). Gender variation in Dutch: A sociolinguistic study  of Amsterdam speech. Dordrecht: Foris.  Byrne, D. (1971). The attraction paradigm. New York: Academic  Press.  Cacioppo, J., & Petty, R. (1982). Language variables, attitudes, and  persuasion. In E. B. Ryan, & H. Giles (Eds.), Attitudes towards lan- guage variation (pp. 189-207). London: Edward Arnold.    Cohen, J. (1992). Quantitative methods in psychology: A power primer.  Psychological Bulletin, 112, 155-159.    doi:10.1037/0033-2909.112.1.155  Couper, M., & de Leeuw, E. (2003). Nonresponse in cross-cultural and  cross-national surveys. In J. Harkness, F. van de Vijver, & P. Mohler  (Eds.), Cross-cultural survey methods (pp. 157-178). Hoboken, NJ:  Wiley.  Coupland, N., & Giles, H. (1988). Introduction: The communicative  context of accommodation. Language and communication, 8, 75-  182. doi:10.1016/0271-5309(88)90015-8  Curtin, R., Presser, S., & Singer, E. (2000). The effects of response  rates changes on the index of consumer sentiment. Public Opinion  Quarterly, 64, 413-428. doi:10.1086/318638  Curtin, R., Presser, S., & Singer, E. (2005). Changes in telephone sur-  vey nonresponse over the past quarter century. Public Opinion Quar-  terly, 69, 87-98. doi:10.1093/poq/nfi002  Daan, J., & Blok, D. (1969). From the urban agglomeration of Western  Holland to the border. Explanation of the map dialects and onomas-  tics, the atlas of the Netherlands. Amsterdam: Noord-Hollandsche  Uitgevers Maatschappij.  De Heer, W. (1999). International response trends. Journal of Official  Statistics, 15, 129-142.  Dehue, F. (1997). Results observation introduction behaviour com-  muters. Heerlen: Centraal Bureau voor de Statistiek.  De Leeuw, E. D., & de Heer, W. (2002). Trends in household survey  nonresponse. In R. M. Groves, D. A. Dillman, J. L. Eltinge, & J. A.  Roderick (Eds.), Survey nonresponse (pp. 41-54). New York: Wiley.  De Leeuw, E. D., Callegaro, M., Hox, J., Korendijk, E., & Lensveldt-  Mulders, G. (2007). The influence of advance letters on response in  telephone surveys. Public Opinion Quarter ly, 71, 413-443.  doi:10.1093/poq/nfm014  Dijkstra, W., & Smit, J. H. (2002). Persuading reluctant recipients in  telephone surveys. In R. M. Groves, D. A. Dillman, J. L. Eltinge, and  R. J. A. Little (Eds.), Survey nonresponse. (pp. 121-134), New York:  Wiley.  Durrant, G. B., Groves, R. M., Staetsky, L., & Steele, F. (2010). Effects  of interviewer attitudes and behaviors on refusal in household sur-  veys. Public Opinion Quarterly, 74, 1-36. doi:10.1093/poq/nfp098  Eimers, T., & Thomas, E. (2000). Instruction callcenter personnel de-  mands new approach. Nijmegen: ITS.  Feskens, R., Hox, J., Lensveldt-Mulders, G., & Schmeets, H. (2007).  Nonresponse among ethnic minorities: A multivariate analysis. Jour-  nal of Official Statistics, 2, 387-408.  Fienberg, S. E. (2007). The analysis of cross-classified categorical data.  New York: Springer. doi:10.1007/978-0-387-72825-4  Gerritsen, M., & van Bezooijen, R. (1996). How can I increase the  effect of my telemarketing business? Ad Rem, 4, 9-12.  Gerritsen, M., & Palmen, M.-J. (2002). The effect of prenotification  techniques on the refusal rate in telephone surveys. A real-life study  in the light of the compliance and elaboration likelihood theory.  Document Design, 3, 16-29. doi:10.1075/dd.3.1.04ger  Giles, H. (1973). Communicative effectiveness as a function of ac-  cented speech. Speech Monographs, 40, 330-331.  doi:10.1080/03637757309375813  Grondelaers, S. A., & van Hout, R. W. N. M. (2010). Is standard Dutch  with a regional accent standard or not? Evidence from native speak- ers’ attitudes. Language Variation and Change, 22, 221-239.  doi:10.1017/S0954394510000086  Grondelaers, S. A., van Hout, R. W. N. M., & Steegs, M. (2010). Eva-  luating regional accent variation in standard Dutch. Journal of Lan-  guage and Social Psyc ho lo gy, 29, 101-116.  doi:10.1177/0261927X09351681  Groves, R. M. (2006). Nonresponse rates and nonresponse bias in  household surveys. Public Opinion Quarterly, 70, 646-675.    doi:10.1093/poq/nfl033  Groves, R. M. (2011). Three eras of survey research. Public Opinion  Quarterly, 75, 661-671. doi:10.1093/poq/nfr057  Groves, R. M., Dilman, D. A., Eltinge, J. L., & Little, R. J. A. (2002).  Survey nonresponse. New York: John Wiley and Sons, Inc.    Groves, R. M., Cialdini, B., & Couper, M. (1992). Understanding the  decision to participate in a survey. Public Opinion Quarterly, 56,  475-495. doi:10.1086/269338  Groves, R. M., & Couper, M. (1996a). Nonresponse in household in-  terview surveys. New York: Wiley-Interscience.  Groves, R. M., & Couper, M. (1996b). Contact-level influences on  cooperation in face-to-face surveys. Journal of Official Statistics, 12,  63-83.  Groves, R. M., & Fultz, N. (1985). Gender effects among telephone  interviewers in a survey of economic attitudes. Sociological Methods  and Research, 14, 31-52. doi:10.1177/0049124185014001002  Groves, R. M., & McGonagle, K. A. (2001). A theory-guided inter-  viewer training protocol regarding survey participation. Journal of  Official Statistics, 17, 249-265.  Groves, R. M., O’Hare, B. C., Gould-Smith, D., Benkí, J., & Maher, P.  (2008). Telephone interviewer voice characteristics and the survey  participation decision. In J. M. Lepkowski, C. Tucker, J. M. Brick, E.  de Leeuw, L. Japec, P. J. Lavrakas, M. W. Link, & R. L. Sangster  (Eds.),  Advances in telephone survey methodology (pp. 385-400).  New York: John Wiley and Sons, Inc.    Groves, R. M., & Peytcheva, E. (2008). The impact of nonresponse  rates on nonresponse bias. Public Opinion Quarterly, 72, 167-189.  doi:10.1093/poq/nfn011  Hansen, K. M. (2007). The effect of incentives, interview length, and  interviewer characteristics on response rates in a CAT study. Inter-  national Journal of Public Opinion Research, 19, 112-121.  doi:10.1093/ijpor/edl022   Heijmer, T., & Vonk, R. (2002). Effects of a regional accent on the  evaluation of the speaker. Nederlands Tijdschrift voor de Psy-  chologie, 57, 108-113.  Holmes, J., & Meyerhoff, M. (2005). The handbook of language and  gender. Oxford: Blackwell Publishing.  Hoppenbrouwers, C. (2001). The classification of Dutch dialects. Dia- lects of 156 towns and villages classified according the FFM. Assen:  Koninklijke Van Gorcum.  Houtkoop-Steenstra, H., & van den Bergh, H. (2000). Effects of intro-  ductions in large-scale telephone survey interviews. Sociological Me-  thods and Research, 28, 281-300.  doi:10.1177/0049124100028003002  Hox, J. J., & de Leeuw, E. D. (2002). The influence of interviewers’  Copyright © 2012 SciRes.  168   
 M.-J. PALMEN  ET  AL.  Copyright © 2012 SciRes. 169  attitude and behavior on household survey nonresponse: An interna- tional comparison. In R. M. Groves, D. A. Dillman, J. L. Eltinge, & J.  A. Roderick (Eds.), Survey nonresponse (pp. 103-120). New York:  Wiley.  Keeter, S., Miller, C., Kohut, A., Groves, R. M., & Presser, S. (2000).  Consequences of reducing nonresponse in a national telephone sur-  vey. Public Opinion Quarterly, 64, 125-148.  doi:10.1086/317759  Keeter, S., Kennedy, C., Dimock, M., Best, J., & Craighill, P. (2006).  Gauging the impact of growing nonresponse on estimates from a na-  tional RDD telephone survey. Public Opinion Quarterly, 70, 759-  779. doi:10.1093/poq/nfl035  Kraaykamp, G. (2005). Dialect and social inequality: An empirical stu-  dy of the social-economic consequences of speaking a dialect in  one’s youth. Pedag og is ch e S tudiën, 85, 390-403.    Lambert, W. (1967). A social psychology of bilingualism. Journal of  Social Issues, 23, 91-109. doi:10.1111/j.1540-4560.1967.tb00578.x  Mai, R., & Hoffmann, S. (2011). Four positive effects of a sales-  person’s regional dialect in services selling. Journal of Service Re-  search, 14, 460-474. doi:10.1177/1094670511414551  Maynard, D. W., & Schaeffer, N. C. (1997). Keeping the gate. Decli-  nations of the request to participate in a telephone survey interview.  Sociological Methods and Research, 26, 34-79.  doi:10.1177/0049124197026001002  Merkle, D., & Edelman, M. (2002). Nonresponse in exit polls: A com-  prehensive analysis. In R. M. Groves, D. A. Dillman, J. L. Eltinge, &  J. A. Roderick (Eds.), Survey nonresponse (pp. 243-258). New York:  Wiley.  Merkle, D., Edelman, M., Dykeman, K., & Brogan, C. (1998). An ex-  perimental study of ways to increase exit poll response rates and re-  duce survey error. The Annual Meeting of the American Association  for Public Opinion Research, St. Louis.  Milroy, J., & Milroy, L. (1999). Authority in language: Investigating  standard English. London: Routledge.  Oksenberg, L., Coleman, L., & Cannel, C. F. (1986). Interviewers’  voices and refusal rates in telephone surveys. Public Opinion Quar-  terly, 50, 97-111. doi:10.1086/268962  Oksenberg, L., & Cannell, C. F. (1988). Effects of vocal characteristics  on nonresponse. In R. Groves, P. P. Biemer, L. E. Lyberg, J. T. Mas-  sey, W. L. Nicholls, & J. Waksberg (Eds.), Telephone survey meth-  odology (pp. 257-273). New York: Wiley.    Palmen, M.-J. (2001). Response in telephone surveys. A study of the  effect of sociolinguistic factors. Nijmegen: Nijmegen University Press.    Pickery, J., & Loosveldt, G. (1998). The impact of respondent and in-  terviewer characteristics on the number of “no opinion” answers: A  multilevel model for count data. Quality and  Qu an tit y,  3 2, 31-45.  doi:10.1023/A:1004268427793  Pondman, L. M. (1998). The influence of the interviewer on the refusal  rate in telephone surveys. Amsterdam: Print Partners Ipskamp.  Singer, E. (2006). Introduction. Nonresponse bias in household surveys.  Public Opinion Quarterly, 70, 637-645. doi:10.1093/poq/nfl034  Singer, E., van Hoewyk, J., Gebler, N., Ragunathan, T., & McGonacle,  K. (1999). The effect of incentives on response rates in interview-  mediated surveys. Journal of Offici a l Statistics, 15, 217-230.  Smakman, D. (2006). Standard Dutch in the Netherlands. A sociolin-  guistic and phonetic description. Utrecht: Lot.  Snijkers, C., Hox, J., & de Leeuw, E. D. (1999). Interviewers’ tactics  for fighting survey nonresponse. Journal of Official Statistics, 15,  185-198.  Steeh, C. (1981). Trends in nonresponse rates. Public Opinion Quar-  terly, 45, 40-57. doi:10.1086/268633  Steeh, C., Kirgis, N., Cannon, B., & DeWitt, J. (2001). Are they really  as bad as they seem? Nonresponse rates at the end of the 20th cen-  tury. Journal of Official Statistics, 17, 227-247.    Stoop, I. A. L. (2005). The hunt for the last respondent. Nonresponse in  sample surveys. The Hague: Sociaal en Cultureel Planbureau.    Stoop, I., Billiet, J., Koch, A., & Fitzgerald. R. (2010). Improving sur- vey response. Lessons learned from the european social survey.  Chicester: John Wiley and Sons, Ltd. doi:10.1002/9780470688335  Tielen, M. (1992). Male and female speech: An experimental study of  sex-related voice and pronunciation characteristics. Ph.D. Thesis,  Amsterdam: University of Amsterdam.    Van Bezooijen, R. (1995). Sociocultural aspects of pitch differences  between Japanese and Dutch women. Language and Speech, 38, 253-  265.  Van der Vaart, W., Ongena, Y., Hoogendoorn, A., & Dijkstra, W. (2005).  Do interviewers’ voice characteristics influence cooperation rates in  telephone surveys? International Journal of Public Opinion Re-  search, 18, 488-499. doi:10.1093/ijpor/edh117  Van Ingen, E., Stoop, I., & Breedveld, K. (2009). Nonresponse in the  Dutch time use survey: Strategies for response enhancement and bias  reduction. Field Methods, 21, 69-90.   doi:10.1177/1525822X08323099  Vercruyssen, A., van de Putte, B., & Stoop, I. A. L. (2011). Are they  really too busy for survey participation? The evolution of busyness  and busyness claims in Flanders. Journal of Official Statistics, 27,  619-632.  Visscher, G. (1999a). Het CBS is niet meer geloofwaardig. Statistics  Netherlands is not trustworthy anymore. NRC Handelsblad, 21 Janu-  ary 1999.  Visscher, G. (1999b). CBS overschrijdt de grens van het betamelijke  (Statistics Netherlands exceeds bounds of decency). NRC Handels-  blad, 1 February 1999.  Walrave, M. (1996). Telemarketing: Breakdown on the line? Leuven/  Amersfoort: Acco.  Wenemark, M., Persson, A., Brage, H. N., Svensson, T., & Kristenson,  M. (2011). Applying motivation theory to achieve increased response  rates, respondent satisfaction, and data quality. Journal of Official  Statistics, 27, 393-414.     
			 
		 |