# Grade Expectations: Rationality and Overconfidence

^{1}Department of Econometrics and Operations Research, Vrije Universiteit Amsterdam, Amsterdam, Netherlands^{2}Tinbergen Institute, Amsterdam, Netherlands^{3}Department of Applied Economics, National Research University Higher School of Economics, Moscow, Russia

Confidence and overconfidence are essential aspects of human nature, but measuring (over)confidence is not easy. Our approach is to consider students' forecasts of their exam grades. Part of a student's grade expectation is based on the student's previous academic achievements; what remains can be interpreted as (over)confidence. Our results are based on a sample of about 500 second-year undergraduate students enrolled in a statistics course in Moscow. The course contains three exams and each student produces a forecast for each of the three exams. Our models allow us to estimate overconfidence quantitatively. Using these models we find that students' expectations are not rational and that most students are overconfident, in agreement with the general literature. Less obvious is that overconfidence helps: given the same academic achievement students with larger confidence obtain higher exam grades. Female students are less overconfident than male students, their forecasts are more rational, and they are also faster learners in the sense that they adjust their expectations more rapidly.

## Introduction

Most people overestimate their abilities. Svenson (1981), for example, estimates that 93% of US drivers and 69% of Swedish drivers consider their driving skills “above the median.” Overconfidence appears to be one of the most robust findings in experimental psychology (De Bondt and Thaler, 1995). We contribute to this literature by asking how well undergraduate students forecast their grade, given their ability and other control variables.

The first, perhaps, to investigate this issue was Murstein (1965) using a sample of 76 students from a course in educational psychology at Louisiana State University. Persistency of overconfidence was found, especially for the weaker students. Grimes (2002) studied a sample of 253 students enrolled in a principles of macroeconomics course at Mississippi State University, and found a high degree of overconfidence. Nowell and Alston (2007) used data from a survey conducted in 32 courses, representing every class offered by the economics department at Weber State University in Utah during one semester. The sample consisted of 715 students with a 70% response rate. The authors found that male students with a lower grade point average (GPA) have greater overconfidence; that students in upper division classes have less of a tendency to overestimate their grades relative to students taking lower division courses; that gender matters; and that increasing the importance of tests reduces overconfidence. Hossain and Tsigaris (2015) considered students in a second year statistics for business and economics course at Thompson Rivers University in Kamloops, British Columbia (Canada). A total of 169 students were surveyed with a response rate of over 90%. Students were asked to make several forecasts of their final exam grade during the course. The rational expectations hypothesis was rejected. Expectations move closer to the realized grade as students receive new information on their actual performance closer to the exam. Many other papers reject the hypothesis of rational expectations and confirm student grade overconfidence; see Kruger and Dunning (1999), Svanum and Bigatti (2006), Andrews et al. (2007), Burns (2007), Jensen and Moore (2008), Khachikian et al. (2011), Hossain and Tsigaris (2013), Feld et al. (2017), Foster et al. (2017), Serra and DeMarree (2016), and Sturges et al. (2016).

Is overconfidence helpful or harmful for the student? There is no consensus. Overconfidence may induce a student to allocate less time to study, resulting in poor exams grades. On the other hand, Ballard and Johnson (2005) argue that expectations could become self-fulfilling, possibly because the student with higher expectations will work harder and more intensely on the course. They found that expected grades relate positively to a student's performance in class. Johnson and Fowler (2011) argue, along similar lines, that overconfidence may increase ambitions, morale, resolve, persistence, and thus actually increase the probability of success.

Does overconfidence depend on gender? Are women better forecasters? Again, there is no consensus. Guzman (2012) demonstrated that in the housing market gender is a significant factor in price expectations. Women are less optimistic (perhaps more realistic) about housing prices than men. Also, women tend to be better forecasters of unemployment and inflation than men, also when one controls for income, education, race, age, marital status, number of children in the household, et cetera. Lundeberg et al. (1994) concluded from a sample consisting of three psychology courses containing 70 men and 181 women that both men and women tend to be overconfident, but men more so, especially when they are wrong! Nowell and Alston (2007) analyzed a sample of students enrolled in economics and quantitative courses. They concluded that men were 9% more likely to overestimate their grade than women. Jakobsson (2012) also found a gender difference in the prediction error of exam grades from a sample of 98 students in introductory macroeconomics at Karlstad University (Sweden). Others do not find significant differences in prediction accuracy between men and women. Maxwell and Lopus (1994) reported that both men and women tend to overstate their grade point averages, but they did not find a difference by gender. Grimes (2002) and Andrews et al. (2007) also did not find gender differences in the forecast error. To the best of our knowledge, there is only one paper which states the opposite: Sharma and Shakeel (2015) considered students in India and found that the male students seemed to be more modest in the prediction of their exam grades than the female students.

How persistent are overly optimistic expectations? Do students adjust their forecasts? Murstein (1965) found persistency of overconfidence, especially for the weaker students. The vast majority of the strong students showed no significant change in their predictions as their grade experience accumulated. They believed that they deserved high grades and they received high grades. The weaker students did not change their predictions either, although they should have. In a sample of 60 students from a course in research methods in the Department of Psychological Sciences at Texas Tech University, Serra and DeMarree (2016) concluded that students' predictions of their grades were persistently overconfident because their predictions were biased by their desired level of performance. Foster et al. (2017) experimented with 13 consecutive (weekly) exams in an introduction to educational psychology course. They also found that students did not adjust their expectations. Grimes (2002) and Burns (2007) concluded that students' grade expectations became more accurate as they gained experience in the course. Grimes (2002) noted that women appeared to be more successful in bringing their expectations in line with their performances than men.

The main focus of our paper is the expectation of the student about his or her grade, and the papers discussed above shed some light on what others have found. There exists a vast literature on the broader subject of overconfidence, both from a theoretical and an empirical viewpoint. The bias in social comparative judgments is discussed in comprehensive reviews by Chambers and Windschitl (2004) and Windschitl and O'Rourke Stuart (2015). Moore and Schatz (2017) distinguished between overestimation (what we study), overplacement and overprecision, and they distinguished between two possible benefits of overconfidence: intrapersonal and interpersonal. Armor et al. (2008) argued that “if people believe, rightly or wrongly, that unrealistic optimism has some value, then optimistic bias may be usefully understood as being consistent with people's values and beliefs.” Armor and Taylor (1998) review about 300 theoretical papers on the usefulness of optimism.

In our study we do not discuss a theoretical framework of our findings. We simply introduce a model which helps us to measure overconfidence in students' forecasts. Our aim is to contribute to the questions raised above by analyzing students enrolled in a second-year undergraduate course in statistics at ICEF, Moscow, in total 592 students. During the course each student took three exams and at each exam they forecasted their grade. We address the following research questions: Are students' expectations rational? Are they overconfident? If so, is the level of overconfidence the same for male and female students? Is overconfidence helpful? Do students adjust their exam grades during the course when more information becomes available? And, if so, does the speed of adjustment depend on gender? We find that, in general, students are overconfident, especially male students; overconfidence is helpful; students adjust their forecasts with their experience of the course; and female students adjust their beliefs faster than male students.

The paper is organized as follows. The setup is described in section Methods and Participants. Rationality, overconfidence, and persistence are investigated in section Results. Section Discussion and Conclusions offers some discussion and concluding remarks.

## Methods and Participants

### Course Organization and Grading

The International College of Economics and Finance (ICEF) in Moscow was established in 1997 jointly by the London School of Economics and Political Science (LSE) in London and the Higher School of Economics (HSE) in Moscow. The college offers a 4-year bachelor's program, which is considered to be one of the top programs in economics in Russia. Each year about 200 students enter the program, typically immediately after high school. In their first year the students follow, among other subjects, a course called Statistics-1, and in their second year they follow Statistics-2. Both courses are compulsory. Our data are obtained from students following Statistics-2 over a 5-year period, 2011–2015. In total, 964 students took this course during these 5 years.

In Statistics-2 students take three exams every year, at the end of October (exam 1), the end of December (exam 2), and the end of March (exam 3). The exams are written exams, not multiple choice, and each consists of two parts (80 min each) with a 10 min break between the two parts. The level of the exam questions is the same in the two parts. In order to avoid cheating, students are not allowed to leave and come back during each part of the exam. At the end of part 1 and at the end of the exam the examiner collects each student's work. Each part is graded out of 50 points.

In addition, students have weekly homework assignments although these are not compulsory. All handed-in assignments are graded (out of 100). The variable *HW* denotes for each student the sum of all assignments' grades divided by the total number of assignments. For example, if a student hands in 20 of the 25 assignments and scores 100 (the maximum) for each, then *HW* = (20·100)/25 = 80.

After completion of the three course exams, students take two additional exams (some only take one) in early May administered by the University of London, called *STAT*_{1} and *STAT*_{2}. These, like the other exams, are also graded with a maximum score of 100. The final grade *G*_{tot} for the course is then determined as

where *G*_{j} is the grade obtained in exam *j*.

Students fail if *G*_{tot} is smaller than some threshold to be determined by the teacher, but lying between 32 and 37. Student also fail if *G*_{3}, the grade in the third exam, is <25. Exam 3 thus plays a special role in two ways: its weight is higher than for the other two exams and there is a threshold grade of 25.

### Self-Assessment

At the end of the first part of each of the three exams each student was asked to forecast (out of 100) his/her grade for this exam (the two parts together). Students were told that answering this question is voluntary: they can answer or they can skip the question. They were also told that their answers could be used for research purposes anonymously. At the moment when the student writes down the forecast he/she knows the questions and his/her answers in part 1, but the student does not yet know the questions of part 2. To encourage students to fill in their forecast and to actually try their best, a bonus is promised. If the difference between the forecast and the grade is ≤ 3 in absolute value, then one bonus point is added to the grade. For example, if the forecast is 49 and the grade is 52, then the grade for this exam is marked up to 53. This procedure had to be and has been approved by the ICEF administration. As a result of the procedure and the possibility of a bonus, the response rate was extremely high (97%). The idea of giving each student an incentive to express his/her opinion was also used in a recent experiment by Blackwell (2010), where students were asked to assess the difficulty of an assignment by guessing the class average, earning a bonus if their guess was close enough.

Smart (or risk averse) students utilize this bonus in the third exam, where the grade must be ≥25. If the student chooses the forecast *F*_{3} such that 21 ≤ *F*_{3} ≤ 27, then a grade *G*_{3} = 24 would be marked up to 25. Some students actually do this, but they then typically choose *F*_{3} = 24 or 25 and not, say, 21 or 27. The special role of the third exam and the overrepresentation of 24 and 25 in the sample of third exam forecasts have to be taken into account when we do our statistical analysis, and we shall discuss this issue further below.

### The Data

The data consist of the grades *G*_{j} and the forecasts *F*_{j} (*j* = 1, 2, 3) for each of our students, and our interest is focused on the excess expectation in exam *j*:

We have some background knowledge on each student, namely the grades of the first-year calculus (*calc*) and first-year statistics (*stats*) exams, the grade point average at the end of the first year (*gpa*), and whether the student is male or female (*female* = 1 for women and 0 for men). We also know in which year the exam took place (*year* = 2011, 2012, 2013, 2014, 2015).

For the homework assignments we know for each student how many assignments the student handed in (*nhw*_{j}), the sum of the grades per exam period (*shw*_{j}), and the number of handed-out assignments (*nhwmax*_{j}). The index *j* now refers to a period rather than to a point in time: *j* = 1 refers to the period up to the first exam, *j* = 2 to the period between the first and second exams, and *j* = 3 to the period between the second and third exams. The number of handed-out assignments (*nhwmax*_{j}) may vary from year to year, in fact from 4 to 7 in period 1, from 6 to 7 in period 2, and from 9 to 12 in period 3.

From these “raw” data we can compute the ratios

which denote, respectively, the relative number of submitted assignments for each student in period *j* (0 ≤ *rnhw*_{j} ≤ 1), and the average submitted assignment grade for each student in period *j* (0 ≤ *rshw*_{j} ≤ 100), where we set *rshw*_{j} = 0 if *nhw*_{j} = 0. These ratios will be used later in the analysis.

In order to obtain a clean and complete sample, some data screening was necessary. Of the original 964 students we excluded those students who (a) did not take all three exams; or (b) had repeated the first year; or (c) had failed the course last year; or (d) had taken a break between the first and second year. This left us with 840 students.

Of these 840 students, a further 248 were excluded because they did not provide all three forecasts or we didn't have their first-year results. As a consequence, 592 students remain on which we have complete information. The results of the University of London exams are not used in our analysis.

A summary of these “raw” and basic data is provided in Table 1 (Kernel density plots of the basic data are provided as Supplementary Material). There is substantial variation in the exam grades *G* and the forecasts *F* over the years, possibly because the difficulty of the exams varies (although the same instructor taught the course over this period) or the quality of the student population varies (because of changes in admission policies). This suggests that year dummies may be important. On the whole it seems that students are too optimistic about their abilities, because *F*_{j} > *G*_{j} occurs more frequently than *F*_{j} < *G*_{j}. Also, in the first two exams large deviations occur where |*F*_{j} − *G*_{j}| can be larger than 10, in contrast to the third exam where the deviation is much smaller. This suggests that students learn from their past forecast errors.

From these raw data it is not immediately clear what the answers are to our questions. To achieve this we need more sophisticated statistical techniques than simple averages. In the next section we will address each of our research questions in turn and develop the required models as we progress.

## Results

### Rationality

Our first question is whether our students have rational expectations about their exam grades. In the Introduction we mentioned some literature where it is found that students overestimate their abilities, that is, that they are not rational. If we also find this (as we shall) then a second question arises, namely whether male and female students are equally irrational or that perhaps female students behave more rationally than male students.

Our experimental data differ from the data in most papers in three respects: first, we use a 0–100 grade system, while other papers typically use the more discrete F-D-C-B-A (0–4) grade system; second, our students make their forecast after they have already finished half the exam; and third, our students have a real incentive to make their forecast as precise as possible, as they get a bonus for an accurate forecast.

There is, in addition, one other feature of our data, namely the fact that we collected exam results and the associated forecasts during 5 years (2011–2015). We know from the previous section that the exams are not equally difficult in each year, and these discrepancies need to be taken into account. Thus, following Hossain and Tsigaris (2015), we regress the excess expectation *D*_{j, i} = *F*_{j, i} − *G*_{j, i} for student *i* in exam *j* for each of the three exams separately, and include year dummies *year*12_{i} − *year*15_{i}. The regression then reads

where *x*_{j, i} is the vector of all available information at the time of exam *j* of the *i*-th student's previous academic achievements, and the control variables are a vector of time dummies

and the female/male dummy *female*_{i}. The α_{j}, β_{j}, γ_{j}, and δ_{j} are unknown parameters (parameter vectors), and ε_{j, i} is the random error, which we assume to be independently and identically distributed with mean zero.

We define a student to be rational when the conditional expectation E(*D*_{j, i}|*x*_{j, i}) = 0 and this translates to testing the null hypothesis

for each of the three exams *j* = 1, 2, 3. Note that the dimension of β_{j} is not the same for each *j*, because more information is available at the second exam than at the first exam, and even more information is available at the third exam. In fact, the dimension is 5 at the first exam (*calc, stats*, *gpa*, *rnhw*_{1}, *rshw*_{1}), 8 at the second (the previous plus *G*_{1}, *rnhw*_{2}, *rshw*_{2}), and 11 at the third (the previous plus *G*_{2}, *rnhw*_{3}, *rshw*_{3}).

Recall that the third exam is special because students fail the course if *G*_{3} < 25. From the student's point of view it makes good sense to predict 21 ≤ *F*_{3} ≤ 27, because then (and only then) a grade *G*_{3} = 24 will be marked up to 25. This is “rational” behavior, but not according to our definition. In practice, these students choose *G*_{3} = 24 or 25, but almost never 26 or 27. Also, some students are confused and believe that this rule applies to all three exams and not only to the third. To avoid these problems we only include forecasts which satisfy *F*_{j, i} > 25. We recognize that this censoring could (slightly) shift the data so that finding overconfidence becomes more likely. Since this is something that the data will not reveal, we need to assume that such a shift does not take place.

Table 2 contains the regression results for each of the three exams separately. The last row in the table contains the *p*-value of the *F*-test used for testing the hypothesis that β_{j} = 0. The reported *p*-values are < 0.2% thus rationality is firmly rejected in this model. The *p*-value is lowest at the first exam, still very low at the second (where more information is available), and much higher (but still < 0.2%) at the third (where even more information is available). More information thus leads to more rational decisions, which is not as obvious as it may seem, because too much information might turn into confusion and lead to less rational behavior. This, however, does not happen here.

Some preliminary conclusions can be drawn from Table 2. First, it seems that good students (high marks in *calc*, *stats*, and *gpa* in the previous year) are more cautious than not so good students in their predictions, at least for the first exam. If the student does well in the first exam, then he/she becomes less cautious and in fact tends to overpredict the results of both the second and the third exam. Having learnt their lesson, students become more cautious again: doing well in the second exam leads to more rather than to less caution in predicting their mark for the third exam. There is a big impact of gender. Women are more cautious than men in all three exams, although the impact diminishes over time.

Kruger and Dunning (1999) concluded that low-performing students significantly overestimate their performance, while high-performing students are more accurate in their forecasts. Feld et al. (2017) pointed out that this effect could at least partially be explained by measurement errors. They showed that after correction for measurement errors (using instrumental variables) the Dunning-Kruger effect is still observed, but significantly weaker than before the correction.

To further investigate the difference in rationality between women and men, we also estimate the extended model.

for men and women separately. The results are presented in Table 3, where we have again excluded all students with *F*_{j, i} ≤ 25.

By including the *female* dummy (as in Table 2) we distinguish between men and women, but only by allowing the level to change from α_{j} for men to α_{j} + δ_{j} for women. By separating men and women (as in Table 3, now of course without the *female* dummy), we also allow the β_{j}-coefficients to be different.

Our preliminary conclusions still hold in this extended framework. Women are more cautious than men. Good students are more cautious than not-so-good students in their predictions, at least for the first exam. If the student does well in the first exam, then he/she becomes too optimistic in predicting the second exam, but doing well in the second exam does not lead to such optimism. The *p*-values are higher than in Table 2 but still well under 0.1%, except for women in the third exam where the *p*-value is close to 5%. We thus find that our female students became more rational in the third exam, while men continue to exhibit irrational behavior.

It is often thought that women behave more rationally than men, and this is indeed what we find. But there is no consensus in the literature. Ballard and Johnson (2005) reported that gender is a significant determinant of student expectations: women in an introductory microeconomics course expected a grade that was one-fourth of a letter grade (0.25 on a 4.0 scale) lower than the grade expected by the men. However, after controlling for expectations and secondary-schooling experience with economics, the gender effect became small and insignificant. Hossain and Tsigaris (2015) also found that gender makes no difference in this respect.

### Overconfidence

In the previous section we rejected rationality in predicting exam results and we saw that there is a difference between male and female students. Our next step is to try and explain this lack of rationality, and our hypothesis is that students (especially male students) are too confident about their abilities. When a student has more confidence than is justified by his or her grades, we call this student “overconfident”; see i.e., Windschitl and O'Rourke Stuart (2015).

It makes sense that a student who does well in exams gains in confidence. But perhaps the opposite is also true, that is, a confident student-other things being equal-performs better than one lacking in confidence (Ballard and Johnson, 2005; Johnson and Fowler, 2011). In addition to studying overconfidence we also try to answer this somewhat subtler question in the current section.

An overconfident student will produce a forecast which is higher than can be explained by previous academic results. We write the forecast as

which is the same as Equation (3), except that the dependent variable is now the forecast *F*_{j, i} rather than the excess expectation *D*_{j, i} and also that the *female* dummy has been deleted.

The reason for not including the *female* dummy is that we think of the forecast as a combination of two effects: a part based on factual information and a remainder which we identify with overconfidence. This overconfidence will depend on other things, one of which may be gender. We don't observe the remainder (the errors ε_{i,i}), but we can predict it through the residuals

where ${\widehat{\alpha}}_{j},{\widehat{\beta}}_{j}$, and ${\widehat{\gamma}}_{j}$ are the least-squares estimates from Equation (5). These residuals thus capture that part of the student's forecast which cannot be explained rationally, and thus correspond to our idea of (over)confidence, which is why we denote them by *conf*_{j, i}. Note that if we would include the *female* dummy in Equation (5) then *conf* and *female* would be orthogonal to each other, and this is not reasonable.

Overconfidence, thus defined, may include some information which is not available to us, such as private lessons taken before the exam or certain psychological features of the student. Since this information is not available to us we ignore it.

In the first step of the estimation procedure we thus estimate *F*_{j, i} and obtain the residuals *conf*_{j, i}. In the second step we regress the exam grades *G*_{j, i} on the same set of regressors as in Equation (5) and, in addition, on the residuals *conf*_{j, i} and the *female* dummy (and a cross term):

The results of the two-step procedure are presented in Table 4. The left panel (columns *F*_{1}, *F*_{2}, *F*_{3}) gives the results of the first step. We see that first-year calculus and (to a lesser extent) statistics are important for *F*_{1}, but that home assignments are not important for the forecast. For *F*_{2}, the result *G*_{1} of the first exam is important, while first-year calculus remains important as well. For the final exam the forecast *F*_{3} depends much on the results of the earlier two exams *G*_{1} and *G*_{2}. The results of first-year calculus (and statistics) are not important anymore; these are absorbed in the grades *G*_{1} and *G*_{2}, because recent information is more relevant than older information. We note that ${R}_{adj}^{2}$ increases with the exam number, suggesting that with each exam students become more accurate in their forecasts.

The right panel (*G*_{1}, *G*_{2}, *G*_{3}) gives the results of the second step and allows us to test various hypotheses. We see that ϕ is significantly positive at the 1% significance level for all three exams. Its value increases with time/exam number (0.225, 0.282, 0.345), so the impact of overconfidence increases; at least for the men. For the women, the impact decreases (0.396, 0.272, 0.158) when we take the cross term into account. There is evidence in the literature that the more important is the exam the smaller is the overconfidence (Nowell and Alston, 2007). In our case this is true for women but not for men.

The *female* dummy is not significant and its cross term with *conf* is significant (at 5%) only for the first exam. In contrast to the results in the left panel (the forecasts), more of the “factual” regressors *x*_{j} are significant in the right panel (the grades). The impact of the first-year courses (calculus, statistics, GPA) decreases during the second year, as is to be expected. The grades are significant: *G*_{1} is significant in the *G*_{2} regression (and somewhat less in the *G*_{3} regression), and *G*_{2} is significant in the *G*_{3} regression. Homework results, while not significant for the students' forecasts, are significant for the grades, but only the most recent homework results. The reason, perhaps, is that students understand that homework results are not representative, because there is much collaboration among students and in fact some cheating, so they don't take it into account when forming their forecast. But the plain fact that a student submits the homework (whether own work or not) apparently helps to get a better grade. This finding agrees with Weems (1998), but not with Geide-Stevenson (2009).

Thus we conclude that (a) overconfidence helps in getting a better grade; (b) the impact of information deteriorates quickly over time; (c) homework results are important for the grades, but unimportant for the forecasts; and (d) gender is *not* significant in exam grades.

We next ask: does overconfidence depend on gender? To answer this question we consider the regression

where we note that *conf* is orthogonal to the *year* dummy and the available information in *x*, because *conf* is the vector of residuals from Equation (5). Regression results are presented in the Table 5.

Table 5 shows that, given the same objective factors, male students tend to be more optimistic in forecasting their exam grades than female students. The difference ranges from approximately 2.7–5.0 grade points, and this difference seems to decrease over time (within one exam year). Transforming these grade points to a 0–4 scale we divide by 25 and obtain 0.11 and 0.20, which is of the same order as in Ballard and Johnson (2005).

### Persistence

In the previous section we predicted and studied overconfidence as measured by the residuals *conf*_{j, i} for each student *i* and exam *j*. We found that this overconfidence tends to become smaller as the year progresses. We now address this issue in more depth. That is, we ask whether overconfidence decreases, which would mean that students adjust their (over)confidence.

To answer this question we estimate the dynamic regressions

If |θ_{2}| < 1 then learning takes place between exams 1 and 2. Similarly, if |θ_{3}| < 1 then learning takes place between exams 2 and 3. We run these regressions separately for men and women, because we have seen that overconfidence is not the same for men and women.

The results are presented in Tables 6, 7. It is clear that adjustment occurs, since |θ_{j}| is significantly smaller than one. For the adjustment from exam 1 to exam 2 we find θ_{2} = 0.19 for women and θ_{2} = 0.35 for men. The difference between men and women is statistically significant. For the adjustment from exam 2 to exam 3 we find θ_{3} = 0.43 for women and θ_{3} = 0.35 for men, and this difference is not statistically significant.

Thus we conclude that (a) confidence adjustment occurs for both male and female students; (b) the adjustment from exam 1 to exam 2 is stronger than the adjustment from exam 2 to exam 3; (c) female students are faster learners, certainly in the step from exam 1 to exam 2; and (d) overconfidence persists (since the values of θ_{j} are all positive) and this persistency is stronger for men than for women.

We can go one step further. In the above regressions we estimated the average values of the adjustment coefficient for male and female students. But each student is different and the adjustment coefficient may vary from student to student. In order to estimate the individual values of the adjustment coefficient we model θ_{j} as a function of the individual characteristics of a student:

Inserting Equation (10) in Equation (9) then gives.

Instead of estimates ${\widehat{\theta}}_{j}$ we now obtain distributions (over *i*) of estimates.

Kernel density plots for the distribution of the adjustment coefficient are presented in Figure 1 for the adjustment from exam 1 to exam 2) and Figure 2 for the adjustment from exam 2 to exam 3). We see from the first figure that the density plot for women is shifted toward zero, again demonstrating that female students are faster learners (lower persistence) than male students. There is no significant difference between the density plots for the adjustment from exam 2 to exam 3.

The figures provide further (and more detailed) confirmation of our previous conclusions, namely that (a) adjustment takes place; (b) women are faster learners that men; and (c) there is a persistency of overconfidence from one exam to the next, which is stronger for men than for women.

## Discussion and Conclusions

In this paper we studied second-year undergraduate students in a statistics course over a period of 5 years, comparing their grades with their forecasts. As expected, we find that the students' grade expectations are not rational and that most students are overconfident, which is in agreement with the general literature. Our study had the advantage of a relatively large number of students and a high response rate, and thus contributes to various issues (many of them unresolved) in the general area of rationality, overconfidence, and persistence. The following conclusions emerge.

First, overconfidence decreases during the course and is smallest at the third exam, which shows that students adjust their expectations as information accrues (Grimes, 2002; Burns, 2007), in particular when the third exam has a higher weight in the total course grade (Nowell and Alston, 2007). Some studies did not find this adjustment (Murstein, 1965; Serra and DeMarree, 2016; Foster et al., 2017), others did (Grimes, 2002; Burns, 2007). One of the reasons for the discrepancy in the literature may be attributed to the content/essence/nature of the course. Foster et al. (2017) studied the results of 13 consecutive exams in educational psychology, where the content for each exam covers a separate topic. Our course is quite different in that the content is cumulative: the next exam uses concepts from previous parts of the course.

Second, female students have a lower level of overconfidence than male students, thus exhibiting more rational behavior. This is of interest because the literature is not in agreement on this issue. Our results are similar to what Guzman (2012) found in financial forecasts and Lundeberg et al. (1994) and Jakobsson (2012) for grade forecasts. But others (e.g., Maxwell and Lopus, 1994; Grimes, 2002; Andrews et al., 2007) did not find such differences.

Third, female students are not only better forecasters, they are also faster learners than male students, showing a faster adjustment of their grade expectations. We did not find difference in *grades* between female and male students. The role of gender in the fields of science, technology, engineering, and math (STEM) is widely discussed in the literature (see e.g., Schmader, 2002, 2010). Schmader (2002) concluded that when gender identity is not linked to test performance, women perform equally to men, and this finding is confirmed in our study.

Fourth, overconfidence has a positive effect on exam grades. Some studies suggest that overconfident students are less successful at exams since they allocate less time and efforts to study. This may be the case for some students, but we find that for most students overconfidence is advantageous, possibly because it increases ambition, morale, resolve, persistence, and hence the probability of success (Ballard and Johnson, 2005; Johnson and Fowler, 2011).

Finally, a suggestion to teachers based on our findings. Don't wait too long in setting your first test. This will help students to adjust their expectations at an early stage, and this in turn will be of use to them in their allocation of time and effort for the course.

## Author Contributions

All authors listed have made a substantial, direct and intellectual contribution to the work, and approved it for publication.

## Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

## Acknowledgments

We are grateful to two referees and to seminar participants at the Department of Applied Economics, Higher School of Economics, Moscow for constructive comments.

## Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2017.02346/full#supplementary-material

## References

Andrews, K., Swanson, J., and Kugler, P. (2007). Grade expectations. *J. Econ. Econ. Educ. Res.* 8, 3–18.

Armor, D. A., and Taylor, S. E. (1998). Situated optimism: specific outcome expectancies and self-regulation. *Adv. Exp. Soc. Psychol.* 30, 309–379. doi: 10.1016/S0065-2601(08)60386-X

Armor, D. A., Massey, C., and Sackett, A. M. (2008). Prescribed optimism: is it right to be wrong about the future? *Psychol. Sci.* 19, 329–331. doi: 10.1111/j.1467-9280.2008.02089.x

Ballard, C., and Johnson, M. (2005). Gender, expectations, and grades in introductory microeconomics at a US university. *Fem. Econ.* 11, 95–122. doi: 10.1080/1354570042000332560

Blackwell, C. (2010). Rational expectations in the classroom: a learning activity. *J. Econ. Educ.* 10, 1–6.

Burns, D. J. (2007). An examination of the accuracy of students' expected grades. *Acad. Educ. Leadership J.* 11, 45–58.

Chambers, J. R., and Windschitl, P. D. (2004). Biases in social comparative judgments: the role of nonmotivated factors in above-average and comparative-optimism effects. *Psychol. Bull.* 130, 813–838. doi: 10.1037/0033-2909.130.5.813

De Bondt, F. M., and Thaler, R. H. (1995). “Financial decision making in markets and firms: a behavioral perspective,” in *Handbooks in Operations Research and Management Science*, Vol. 9, eds R. A. Jarrow, V. Maksimovic, and W. T. Ziemba (Amsterdam: Elsevier Science), 385–410. doi: 10.1016/S0927-0507(05)80057-X

Feld, J., Sauermann, J., and de Grip, A. (2017). Estimating the relationship between skill and overconfidence. *J. Behav. Exp. Econ.* 68, 18–24. doi: 10.1016/j.socec.2017.03.002

Foster, N. L., Was, C. A., Dunlosky, J., and Isaacson, R. M. (2017). Even after thirteen class exams, students are still overconfident: the role of memory for past exam performance in student predictions. *Metacogn. Learn.* 12, 1–19. doi: 10.1007/s11409-016-9158-6

Geide-Stevenson, D. (2009). Does collecting and grading homework assignments impact student achievement in an introductory economics course. *J. Econ. Econ. Educ. Res.* 10, 3–14.

Grimes, P. W. (2002). The overconfident principles of economics student: an examination of metacognitive skill. *J. Econ. Educ.* 33, 15–30. doi: 10.1080/00220480209596121

Guzman, G. (2012). Sexonomics: gender, wealth, and expectations in financial markets. *J. Econ. Soc. Meas.* 37, 1–60. doi: 10.3233/JEM-2012-0357

Hossain, B., and Tsigaris, P. (2013). Can business students forecast their own grade? *Am. J. Bus. Educ.* 6, 85–91. doi: 10.19030/ajbe.v6i1.7486

Hossain, B., and Tsigaris, P. (2015). Are grade expectations rational? a classroom experiment. *Educ. Econ*. 23, 199–212. doi: 10.1080/09645292.2012.735073

Jakobsson, N. (2012). Gender and confidence: are women underconfident? *Appl. Econ. Lett.* 19, 1057–1059. doi: 10.1080/13504851.2011.613745

Jensen, P. A., and Moore, R. (2008). Students' behaviors, grades & perceptions in an introductory biology course. *Am. Biol. Teach.* 70, 483–487. doi: 10.1662/0002-7685(2008)70[483:SBGPIA]2.0.CO;2

Johnson, D. D., and Fowler, J. H. (2011). The evolution of overconfidence. *Nature* 477, 317–320. doi: 10.1038/nature10384

Khachikian, C. S., Guillaume, D. W., and Pham, T. K. (2011). Changes in student effort and grade expectation in the course of a term. *Eur. J. Eng. Educ.* 36, 595–605. doi: 10.1080/03043797.2011.637189

Kruger, J., and Dunning, D. (1999). Unskilled and unaware of it: how difficulties in recognizing one's own incompetence lead to inflated self-assessments. *J. Pers. Soc. Psychol.* 77, 1121–1134. doi: 10.1037/0022-3514.77.6.1121

Lundeberg, M. A., Fox, P. W., and Punccohar, J. (1994). Highly confident but wrong: gender differences and similarities in confidence judgments. *J. Educ. Psychol.* 86, 114–121. doi: 10.1037/0022-0663.86.1.114

Maxwell, N. L., and Lopus, J. S. (1994). The lake wobegon effect in student self-reported data. *Am. Econ. Rev.* 84, 201–205.

Moore, D. A., and Schatz, D. (2017). The three faces of overconfidence. *Soc. Pers. Psychol. Compass* 11:e12331. doi: 10.1111/spc3.12331

Murstein, B. I. (1965). The relationship of grade expectations and grades believed to be deserved to actual grades received. *J. Exp. Educ.* 33, 357–362. doi: 10.1080/00220973.1965.11010894

Nowell, C., and Alston, R. M. (2007). Thought I got an A! overconfidence across the economics curriculum. *J. Econ. Educ.* 38, 131–142. doi: 10.3200/JECE.38.2.131-142

Schmader, T. (2002). Gender identification moderates stereotype threat effects on women's math performance. *J. Exp. Soc. Psychol.* 38, 194–201. doi: 10.1006/jesp.2001.1500

Schmader, T. (2010). Stereotype threat deconstructed. *Curr. Dir. Psychol. Sci.* 19, 14–18. doi: 10.1177/0963721409359292

Serra, M. J., and DeMarree, K. G. (2016). Unskilled and unaware in the classroom: college students' desired grades predict their biased grade predictions. *Mem. Cogn.* 44, 1127–1137. doi: 10.3758/s13421-016-0624-9

Sharma, V., and Shakeel, M. (2015). Illusion versus reality: an empirical study of overconfidence and self attribution bias in business management students. *J. Educ. Bus.* 90, 199–207. doi: 10.1080/08832323.2015.1014458

Sturges, D., Maurer, T. W., Allen, D., Gatch, D. B., and Shankar, P. (2016). Academic performance in human anatomy and physiology classes: a 2-yr study of academic motivation and grade expectation. *Adv. Physiol. Educ.* 40, 26–31. doi: 10.1152/advan.00091.2015

Svanum, S., and Bigatti, S. (2006). Grade expectations: informed or uninformed optimism, or both? *Teach. Psychol.* 33, 14–18. doi: 10.1207/s15328023top3301_4

Svenson, O. (1981). Are we all less risky and more skillful than our fellow drivers? *Acta Psychol.* 47, 143–148.

Weems, G. (1998). The impact of homework collection in intermediate algebra. *Res. Teach. Dev. Educ.* 15, 21–26.

Keywords: rational expectations, classroom experiment, overconfidence, education, behavioral education, gender difference, persistence

JEL Classification: A22, I21, C91, D01, D90

Citation: Magnus JR and Peresetsky AA (2018) Grade Expectations: Rationality and Overconfidence. *Front. Psychol*. 8:2346. doi: 10.3389/fpsyg.2017.02346

Received: 19 August 2017; Accepted: 22 December 2017;

Published: 12 January 2018.

Edited by:

Salvador Chacón-Moscoso, Universidad de Sevilla, SpainReviewed by:

Fernando Marmolejo-Ramos, University of Adelaide, AustraliaMichael Roy, Elizabethtown College, United States

Copyright © 2018 Magnus and Peresetsky. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Anatoly A. Peresetsky, apereset@gmail.com; aperesetsky@hse.ru