首页    期刊浏览 2025年02月25日 星期二
登录注册

文章基本信息

  • 标题:Which journal rankings best explain academic salaries? evidence from the University of California.
  • 作者:Gibson, John ; Anderson, David L. ; Tressler, John
  • 期刊名称:Economic Inquiry
  • 印刷版ISSN:0095-2583
  • 出版年度:2014
  • 期号:October
  • 语种:English
  • 出版社:Western Economic Association International
  • 摘要:The ranking of an academic journal is important to authors, universities, journal publishers, and research funders. Rankings are gaining prominence as countries adopt regular research assessment exercises that especially reward publication in high-impact journals. (1) Moreover, journal rankings are used as a proxy for the value of published research in many other assessments of departments and individuals (e.g., Combes and Linnemer 2003; Coupe 2003; Kalaitzidakis, Mamuneas, and Stengos 2003; Macri and Sinha 2006). (2) Yet despite this role, even in a rankings-oriented subject like economics there is no consensus on which set of journal rankings is best or most appropriate for a given setting. In order to provide some guidance, in this article, we use academic labor market data to sift between different schemes for ranking economics journals.
  • 关键词:Open access journals

Which journal rankings best explain academic salaries? evidence from the University of California.


Gibson, John ; Anderson, David L. ; Tressler, John 等


I. INTRODUCTION

The ranking of an academic journal is important to authors, universities, journal publishers, and research funders. Rankings are gaining prominence as countries adopt regular research assessment exercises that especially reward publication in high-impact journals. (1) Moreover, journal rankings are used as a proxy for the value of published research in many other assessments of departments and individuals (e.g., Combes and Linnemer 2003; Coupe 2003; Kalaitzidakis, Mamuneas, and Stengos 2003; Macri and Sinha 2006). (2) Yet despite this role, even in a rankings-oriented subject like economics there is no consensus on which set of journal rankings is best or most appropriate for a given setting. In order to provide some guidance, in this article, we use academic labor market data to sift between different schemes for ranking economics journals.

The basis of our approach is that for research-intensive universities, hiring, promotion, and salary advancement are likely directly related to perceived research impacts. When faculty debate merits of publication in one journal or another during labor market deliberations, they are choosing over actions with costly consequences. In contrast, some other ways of ranking journals, such as citation-based metrics, may be prone to strategic manipulation, as we discuss below. In our specific example, we relate salaries of economists in the University of California system to their lifetime publications in 700 different academic journals. These data enable us to test amongst various sets of competing journal rankings, looking in particular at how aggressively each down-weights lower-ranked journals. We also examine how the academic labor market discounts older articles, co-authored articles, and the structure of research portfolios.

As far as we are aware, this is the first article to use labor market data to examine journal rankings. Most studies in this area instead consider variants of citation-based metrics. (3) The most common citation-based metric is the Two Year Impact Factor: the total citations by all journals in a database in a particular year to papers published in the journal considered over the previous 2 years, divided by the 2-year total of papers in that journal. Since the path-breaking work by Liebowitz and Palmer (1984), journal rankings in economics generally weight citations by an assessment of the quality of the source journal to provide adjusted impact factors. However, this growing reliance on impact factors may encourage strategic behavior. For example, editors may coerce authors to add citations to their journals in order to inflate impact factors. Wilhite and Fong (2012) find 175 coercing journals in their survey of researchers in economics, sociology, psychology, and business. (4) Collusion between journals is also possible, where authors of articles in one or more journals deliberately cite papers from another journal. (5) Similarly, individual authors could form cross-citation clubs to raise their personal citation counts, which also inflate the impact factors of the journals they publish in. Hence, what is required for ranking schemes is a robust measure of journal quality which is not easily manipulated.

There is a substantial empirical literature on the academic labor market for economists, but this research has not focused on uncovering measures of journal quality. (6) Instead, academic earning equations are used to consider the negative impact of seniority (Bratsberg, Ragan, and Warren 2003; Moore, Newman, and Turnbull 1998; 2010; Ransom 1993), the returns to co-authorship (Sauer 1998; Moore, Newman, and Turnbull 2001; Hilmerand Hilmer 2005; Hilmer, Hilmer, and Ransom 2012), and the returns to the quantity versus the quality of research, with citations typically a proxy for quality and article counts a proxy for quantity (Hamermesh, Johnson, and Weisbrod 1982; Hilmer, Hilmer, and Ransom 2012). (7) Quantity and quality are revisited by Hamermesh and Pfann (2012), who find that both matter to salary whereas citations (quality) are an important determinant of reputation (using Nobel prizes, Clark medals, Econometric Society fellowships and departmental reputation as proxies) but the quantum of publications is not.

The closest study in the literature to this article is Ellison (2013), who examines a particular academic labor market outcome--which young economists (a PhD since 1988) gained tenure at top 25 U.S. economics departments by 2006/2007. There is a hierarchy of departments, so where each economist gains tenure is a proxy for the labor market's assessment of quality. Tenure decisions are actions with costly consequences, so should be highly informative. Ellison therefore uses this information to assess the widely used h-index of Hirsch (2005), which is the largest number h such that the author has written h papers that have each been cited at least h times. Ellison compares the original h-index with a more generalized form, finding the academic labor market in economics to support the case for a version that emphasizes a smaller number of more highly cited papers. Of relevance to the current study is the explicit emphasis placed by Ellison on labor market information:

... I propose that a useful criterion for assessing competing indexes (and for assessing whether the indexes have any value at all) is to examine whether they are consistent with labor market outcomes.

(Ellison 2013, 63)

Although Ellison (2013) does not examine journal rankings, the same logic of seeking congruence with labor market data can be applied to sifting between the various journal ranking schemes, using information from salary decisions that have costly consequences. We also focus on journal rankings rather than citations because the time lag from publication to receiving citations is impractical for research assessment exercises that often examine just the past 6 years (Tressler and Anderson 2012). For this reason, journal quality measures remain widely used as a proxy for the value of published research, not just in tea room conversations but also when universities make forward-looking decisions on hiring, tenure, promotion, and salary increases.

II. JOURNAL ASSESSMENT MEASURES

This section outlines key characteristics of some representative journal assessment measures that have been used in the economics literature. A wide range of these journal ranking schemes have been proposed by economists, with no consensus on which is the best or which should be used in particular situations. Indeed, according to Engemann and Wall (2009, 128)
   There is no such thing as the correct ranking of economics
   journals. Instead, there is a universe of rankings,
   each the result of a set of subjective decisions by
   its constructor ...


These subjective decisions include which journals to cover, over which years, and whether to rely on citations or expert opinion. If rankings are citation-based further decisions are needed about whether to adjust for the citing sources and for self-citations by authors and journals. Approaches that rely on expert opinion are also subject to potential bias, not only from nonresponse by those surveyed but also at any earlier stage by choosing whose opinion is sought, over what universe of journals. For example, perceptions of older and younger academics may differ on the ranking for new journals, such as the American Economic Journals.

Laband and Piette (1994) ("LP94") apply the pioneering methodology of Liebowitz and Palmer (1984), where adjusted impact factors are determined by using the sum of citations to each journal in an iterative process. The adjusted impact factors are used to weight the citation sources and to provide journal weights. This approach is sometimes referred to as an Eigenfactor approach. The ISI Journal Citation Reports provide the database of articles published in 1990 and their citations to papers published from 1985 to 1989. Only 130 economics journals are given weights in the Laband and Piette scheme, which is the least permissive of any scheme considered here.

Kodrzycki and Yu (2006) also apply the Liebowitz and Palmer methodology, but adjust for the citing intensity of subdisciplines. While also using the ISI Journal Citation Reports, they develop a list of journals commonly used by economists, rather than relying primarily on the ISI list of economics journals. Unlike Liebowitz and Palmer, and Laband and Piette, they consider citations from all social science journals, to provide one set of journal assessment measures ("K&Y_all"), and citations from those journals they classified as economics journals for a second set of measures ("K&Y_econ"). (8) Citations are from 2003 to articles published over the 8 years from 1996 to 2003. The rankings are provided for 181 economics journals. A more recent application of the Liebowitz and Palmer methodology by Kalaitzidakis, Mamuneas, and Stengos (2011) ("KMS") uses the average of the citations in each of the years 2003 to 2008 to articles published in the preceding 10 years. (9) The ISI Journal Citation Reports database is used and only journals classified as economics journals are considered, which gives nonzero rankings to 209 journals.

While iterative (or Eigenfactor) approaches are widely used, some popular journal rankings are based on the simpler direct count of citations. For example, Coupe (2003) uses an average of 2-year impact factors from 1994 to 2000 based on the ISI Journal Citation Reports for 273 economics and related journals. (10) Coupe uses this average, along with three other journal weighting schemes, to provide rankings of departments and individual economists. (11) The Research Papers in Economics (RePEc) Website provides a number of different journal impact assessments using the RePEc citation database. The assessment measure used in this article is the basic impact factor based on direct citations, which is corrected for self-cites to journals, but not self-cites to an author's own papers. This is one of the most permissive schemes, in that impact factors are provided for 984 journals. (12)

[FIGURE 1 OMITTED]

The other permissive ranking, in the sense of trying to not exclude any economics journal, is from Combes and Linnemer (2010). These authors use a number of approaches in order to cover all EconLit journals. They start with 304 journals drawn from EconLit and the ISI Journal Citation Reports, then combine several citation indices, adjusted for the field of specialization, and use a Google Scholar h-index to regress their index on Google Scholar citation data to extrapolate it to all EconLit journals. Assessment measures are proposed based on assumptions about the desired degree of down-weighting of lower-ranked journals, which Combes and Linnemer call "convexity." They propose a "medium" convexity measure, "CLm," and a "high" convexity measure, "CLh," both of which are available for the same set of 1,168 journals. This convexity can be thought of as the elasticity of the weight given to a particular journal with respect to its ranking, as we show below.

The difference between the "medium" and the "high" convexity measures is illustrated in Figure 1, for the top 102 journals in the Combes-Linnemer scheme. These comprise their "AAA," "AA," and "A" groups, with the bottom-ranked journal in the "A" group being Applied Economics. The "medium" and "high" convexity indexes are similar for the top four journals but then a large gap opens up, with the fifth ranked journal (Review of Economic Studies) being equivalent to either 0.81 of the top journal (Quarterly Journal of Economics [QJE]) under medium convexity or else just 0.66 under high convexity. The relative penalty for lower-ranked journals then grows, with the 20th journal being either 0.54 or 0.29 of the QJE and the 50th ranked being just 0.30 or 0.09 of the QJE. To provide a sense of the type of journal at various ranks, Figure 1 highlights the position of four general interest journals of varying quality and their CLm and CLh indexes: the Economic Journal (64.5, 41.6), Economics Letters (30.4, 9.2), Economic Inquiry (24.2, 5.9), and the Southern Economic Journal (19.0, 3.6). These are large gaps in the assessment of relative impact and so it should be possible to detect which degree of convexity is most congruent with labor market data.

The final approach to ranking journals we consider gives an alternative to using citations. Instead, Mason, Steagall, and Fabritius (1997) ("MSF") derive reputational weights from a survey of chairs of economics department in the United States. All 965 departments listed in the December 1989 American Economic Review (AER) were surveyed in 1992 and 1993 with a response rate of 22.4% yielding 216 usable replies. Respondents ranked journals on a 0 to 4 scale (allowing noninteger scores). This ranking has a relatively low degree of convexity, in the sense of not heavily penalizing lower-ranked journals. (13) But these reputational weights are only available for 142 journals so MSF is the second least permissive of the schemes that we consider.

It is clear from this discussion that journal assessment schemes differ in three main ways: the ranking of journals, the degree of coverage or nonzero weights, and the rate at which the weights decline for lower-ranked journals. We illustrate these aspects by applying each of the nine schemes considered to the career publications by the University of California economists in our sample (described below). Applying these schemes to the output of actual economists helps show the important impact of coverage assumptions. In fact, the two least permissive schemes (MSF and LP94) would exclude over one-third of the academic articles published by economists working in economics departments in the largest research-intensive public university system in the United States (Table 1). It is unclear whether the academic labor market also places zero value on publication in these excluded journals. Even the most permissive schemes that attempt to cover the universe of economics journals (RePEc and Combes-Linnemer) would miss over one-tenth of the articles published by these economists. (14)

To show the rate at which the weights decline for lower-ranked journals under each of the nine journal ranking schemes we estimate: log(relative weight) = [alpha] + (Slog(rank). Relative weight of each journal under each ranking scheme is calculated by dividing the journal's actual weight by the weight given to the highest-ranked journal in the scheme (typically the QJE). The regressions are estimated on the nonzero weighted journals published in by the academics in our sample and the coefficient of interest is the elasticity of the relative weight with respect to the rank of the journal. The rank elasticities in Table 1 range from -0.22, for the least aggressive scheme (MSF), to -1.92 for the scheme (LP94) that reduces weights the fastest when moving down the ranks. Although the CLh weights are proposed by Combes and Linnemer as having high convexity (i.e., a rapid decline in the weight given to lower-ranked journals, as shown in Figure 1), it in fact has only a moderate elasticity of the weight with respect to rank. Four other journal ranking schemes have a more elastic response of journal weights to rank (K&Y_all, K&Y_econ, KMS, and LP94), and these same four schemes also exclude at least one-quarter of the articles published to date by the economists in the sample used below. Thus, these four schemes can be considered to be especially aggressive in their focus on perceived journal quality.

The final column of Table 1 reports the average lifetime weighted journal output of the economists in our sample, in terms of AER-sized pages (using 1/n for co-authored papers). This is the key variable which will be used to explain academic salaries, and it varies from 144 pages under the MSF journal rankings to just 36 pages with the LP94 weights. This fourfold difference should be large enough to allow the salary data to discriminate between the various schemes, which are ranked in Table 1 from least aggressive to most aggressive, in terms of the combined impact of the assumptions about noncoverage and convexity. III.

III. DATA

Our approach of using labor market data to uncover the implied quality of academic journals can be applied to any group of academics, in any discipline, in any country. We decided to focus on economists within economics departments in the nine campuses of the University of California system for three reasons. First, the public disclosure database of California state worker salaries (http://www.sacbee.com/statepay/) is unusually detailed, as we describe below. Second, this gives us a well-defined target sample frame that is likely to be of inherent interest--the largest research-intensive public university system in the United States. While the salary returns to various journals may differ in private universities or in other public systems, for the first study of journal rankings using salary data it makes sense to start with the largest system (i.e., there should be good external validity in studying the University of California). Third, while all University of California campuses are research-intensive, they span the perceived quality range from excellent (e.g., Berkeley), to very good (e.g., Davis) to those that are less highly ranked (e.g., Riverside) and emerging (e.g., Merced). This range of quality allows us to test if some journal ranking schemes do a better job of matching labor market outcomes in the best departments. (15)

On the other hand, some features of compensation in the University of California system may differ from other universities, potentially limiting the generality of our results so we briefly discuss these here. The University of California takes a system-wide approach to academic personnel policy and practice. Thus in setting academic salaries, individual campuses operate within the constraints of a multi-institution system using a similar approach to appointments, re-appointments, merit reviews, and promotion. This approach centers on review procedures in which academics are judged by their peers, with evaluations based on research and other factors. After department discussions and voting, recommendations are made to deans or equivalent and, with a few exceptions, final decisions are made by campus chancellors. Thus, overall, these procedures are not significantly more centralized than those that would apply in most public and private institutions, with the key inputs being from the department chairperson and departmental assessments of research contributions.

There are two particular aspects of salary setting in the University of California system that should be noted. The first is the process of merit reviews or regular consideration for advancement on the salary scale within ranks, normally undertaken every 2 years. While merit reviews follow the same general process as appointments, promotions, and tenure decisions, the regular nature of these reviews may mean that quality evaluations are less rigorous with salary increases given for publishing in any of a wide range of outlets. Hence, salary variations resulting from merit increases may provide a less reliable signal of research quality than they would in universities without this type of pro-forma review. But potentially offsetting this factor, reviews undertaken by the University of California have suggested that remuneration has lagged behind benchmark institutions. (16) Thus it is likely that external factors, such as competing offers from other institutions and the need to be competitive in making new appointments, may have greater influence on salary decisions than signals from the merit review process.

In order for labor market data to provide a valid signal of perceived journal quality, the sample has to be relatively homogeneous in terms of the weight placed on research performance in salary determination. So we excluded anyone with significant nondepartment administrative responsibilities (e.g., Deans) and those with primarily teaching roles (including affiliated faculty, adjuncts and those obviously on part-time contracts). While economists infiltrate many other departments we did not consider those for our sample since the returns to publishing in particular economics journals appear to differ between economics-related departments within the same universities (Hilmer, Hilmer, and Lusk 2012). The heterogeneity introduced by expanding the sample would increase the risk of omitting from the salary equation factors that are correlated with some of the journal rankings. Instead, a homogeneous sample and a well-specified model are likely to provide a better test-bed for comparing the various journal rankings. Moreover, while our sample of 223 is smaller than in many other studies using academic earnings equations, it has the advantage of being for a well-defined population of interest rather than simply a hodge-podge of universities with publicly available salary data.

The salary data are unusually detailed, with the base salary reported for 2007, 2009, and 2010 and the total salary reported for those 3 years and also for 2008. Total salary is more temporally volatile than base salary, with squared deviations of annual salary around the multiyear mean for an individual being 32% higher, on average, when using total salary rather than base salary. (17) When we calculate the ratio of total salary to base salary for each individual, it ranged from 0.8 to 2.7, suggesting that the total salary received in any year may not be a good guide to the long-run "permanent" salary. Moreover, while the total-to-base ratio averages 1.10 across all 3 years, it fell from 1.14 in 2007 to 1.07 in 2010, presumably because the worsening financial position of the State of California meant that cuts were being made in extraordinary salary payments. For these reasons we use the base salary rather than the total salary.

Another helpful feature of the salary data provided by the Sacramento Bee Website is that details are provided on the nature of the employment contract, in terms of the pay period. Almost all academics in economics departments at the University of California are on academic year rather than financial year contracts (in contrast to, say, those in agricultural economics departments). In a few cases, especially at UC Berkeley, some economists are on law school scales, so we include a dummy variable in our regressions for individuals whose reported salary is not for a standard scale and 9-month academic year. (18) In contrast, some previous studies of faculty salaries have had to drop individuals for whom it was unclear if their reported salaries were on a 9-month academic year basis (Hamermesh and Pfann 2012).

In addition to salaries we gathered data on gender, the year of first appointment and of hire at the current department, the year of obtaining a PhD (and the department and university), and whether the current appointment was named (such as an endowed or distinguished chair). These details were available from departmental web pages and online curriculum vitae (CVs) for most academics and otherwise we obtained them from dissertation databases and from changes in affiliations on journal articles to date movements. The online CVs also provided the initial information on publications, which were supplemented with searches of EconLit, RePEc, and the Web of Science. Measuring the research outputs of academics with common names can be difficult, but with so many of the sample having their CVs online it helped cross-validate the database search results. We restrict attention to articles that were actually published (with pagination) by the end of 2010. Since our focus is on journal articles, we did not include book reviews, book chapters, editorial introductions, or conference proceedings. (19) The one exception is AER Papers and Proceedings (even though many CVs listed this in the "nonrefereed" section) because seven of

the journal assessment schemes make no distinction between the May issue of the AER and other issues, while LP94 weight the May issue at one-quarter of ordinary issues. Only KMS give Papers and Proceedings a weight of zero. In total, our procedures recorded 5,721 articles in 700 different journals that the 223 economists in our sample had published in over their careers.

Table A1 presents definitions and summary statistics for the variables used in the academic earnings equations. The dependent variable is (log) base salary in 2010, with a mean for the underlying salary data of $156,700 and a range from $78,000 to $310,000 (the maximum total salary is $458,900). (20) The average economist in the sample had spent 12 years at the current university and 18 years at all appointments. One-sixth of the sample is female. Three indicator variables for atypical salary levels (and potentially influential data points) are included: whether the academic is on a nonstandard contract, whether they have a named position (which may fund additional salary), and whether they are a Nobel Prize winner (only one individual). Finally, three indicators of PhD quality are also included: the rank of the PhD-granting department in either the 1995 National Research Council rankings or the Amir and Knauff (2008) rankings, and an indicator for those economists whose PhD was not from an economics department.

IV. RESULTS

The first step in our analysis is to obtain well-specified academic earnings equations to then use as the testing ground for comparing each journal ranking scheme. We then compare nine salary equations, where each uses a different journal ranking scheme to aggregate lifetime publications into a scalar measure for each academic. We use model comparisons statistics and formal non-nested tests to establish which of these models is closer to the truth, in order to see which of the journal ranking schemes is most congruent with the salary data. We then consider the robustness of the results to assumptions about discounting for the age of articles and for co-authorship.

In Table 2, we report the results of various specifications which suggest the following:

indicators of quality for the PhD-granting department are not relevant to salary for this sample (columns 1 to 3); all three of the indicators that we use for atypical salary levels (and potentially influential data points) are statistically significant (columns 4 to 6); the effects of seniority and experience on salary are best modeled as quadratics; there is weak evidence of a premium for males; and, location fixed effects are highly significant. These location effects not only capture cost of living differences but also departmental (and campus) reputational effects, differences in average teaching loads and in the quality of PhD students, and other amenities. (21) Based on these observations, the equation in column 7 of Table 2 is used as the base specification, to which we will then add an output variable measuring lifetime publications in journals, as weighted under each of the nine assessment schemes. Even without the output variable, the base specification explains 72% of the variation in log salary, which is higher than the predictive power of academic earnings equations in other studies. Hence we expect that this sample and regression specification will provide a good test bed for comparing the various journal ranking schemes.

To create the output variable, the number of pages of each of the 5,721 journal articles published by our sample members are multiplied by the assessment weight of the journal (which varies between the ranking schemes). We also adjust for the number of authors of each article (using the "1/n rule") and standardize pages to the size of a typical page in the AER. (22) Thus for each article published by each individual academic the measured output is:

Article Pages X Size Correction

x (1 /number of authors)

X Journal Assessment Weight

and to calculate the lifetime output measure we sum over articles published from the year of the first article until the end of 2010. The full results for the nine different earnings equations, where each in turn uses a different set of journal assessment weights to summarize lifetime output, are reported in Table SI. For these models, the [R.sup.2] ranges from 0.76 to 0.78, so the incremental [R.sup.2] from including the lifetime output measure is 0.04 to 0.06. However, since lifetime output is correlated with experience and seniority, another way to measure the explanatory power of this variable would be to include it first. If we run a simple regression of log salary on each of the lifetime output variables, the [R.sup.2] values would range from 0.46 (using MSF or Coupe weights) to 0.52 (using CLm or CLh weights).

The coefficients on the output measures and a series of model comparison statistics are reported in Table 3. The nine different earnings equations are non-nested, in the sense that it is not possible to impose a set of linear restrictions to derive one model from the other. A standard procedure for model comparison in this case is to use information criteria, with the Akaike's Information Criteria (AIC) and Schwarz's Bayesian Information Criteria (BIC) typically used. We can also compare [R.sup.2] since all the equations have the same number of explanatory variables. The maximized [R.sup.2] and log-likelihood, and the minimized loss of information, is for the earnings equation that uses CLm--the medium convexity weights of Combes and Linnemer (2010). Even though the MSF weights are the second least permissive, in excluding one-third of the articles published to date by this sample, they provide the second best congruence with the labor market data. The greatest loss of information comes from using the KMS weights, which are the third least permissive and have the second largest elasticity of journal weights with respect to the journal ranking (see Table 1).

Our focus is on which journal ranking scheme best fits the salary data, but it is also worth interpreting the magnitudes of some of the regression coefficients. The best fitting model uses CLm weights to aggregate over all journal articles for each academic (effectively calculating the number of standardized pages of (T/E-equivalent quality). For example, Economic Inquiry has a CLm weight of 0.242 and a page size correction factor of 0.9, so a 20-page sole-authored article in Economic Inquiry converts into 4.4 standardized pages of (L/C-equivalent quality (20 x 0.9 x 0.242). The semi-elasticity in the first column of Table 3 indicates that such an article would be associated with an annual salary that was 0.6% higher, which is an increase of $940 at the mean. (23) With a 40-year career and a 5% discount rate, for the average economist in the sample (who is 18 years into their career) such an article would have a net present value of $ 12,400.

Amongst the other variables previously identified from the specification search in Table 2, all are statistically significant at conventional levels except for Male (which has /-statistics between 1.0 and 1.5). The location fixed effects are smaller than in Table 2, suggesting that some of the apparent salary premium at UC Berkeley was productivity-related, but all remain statistically significant. (24) The quadratics suggest that for the average economist, salary is maximized after 30 years of labor market experience and minimized after 27 years of seniority at the current university.

A. Formal Non-Nested Tests

Non-nested tests can help formally discriminate between the competing models in Table 3. These test the validity of one linear model, [H.sub.0] as opposed to its non-nested alternative [H.sub.1]:

[H.sub.0] : y = x[beta] + [[epsilon].sub.0]

[H.sub.1] : y = Z[gamma] + [[epsilon].sub.1]

where X and Z are matrices of explanatory variables, and neither is a linear combination of the other, [beta] and [gamma] are corresponding parameter vectors, and [[epsilon].sub.0] and [[epsilon].sub.1] are random errors. Forming a "compound" model with each competing measure of lifetime output included at the same time is not advisable because of possible multicollinearity. Moreover, this artificial nesting approach does not distinguish between [H.sub.0] and [H.sub.1]; instead, it distinguishes between each competing model and a hybrid (Greene 2012). This can be seen by writing the compound model as:

y = [bar.X][bar.[beta]] + [bar.Z][bar.[gamma]] + W[delta] + [epsilon]

where [bar.X] holds the set of variables in X not in Z, [bar.Z] holds the set of variables in Z not in X, and W has the variables common to the two models. While the test of [bar.[gamma]] = 0 might seem to reject [H.sub.1] and [bar.[beta]] = 0 might reject [H.sub.0], since [delta] remains a mixture of parts of [beta] and [gamma] it is not established by the F-test on the compound model that either of these parts is zero (Greene 2012).

Instead, we use Vuong's (1989) likelihood ratio test that does not presume that either competing model is "true," and instead determines which competitor has verisimilitude (i.e., is closer to the truth). This approach relies on the Kullback-Leibler Information Criterion (KLIC), which, intuitively, is the log-likelihood function under the hypothesis of the true model minus the log-likelihood function for the (potentially mis-specified) model under the assumption of the true model. One model is "better" than another if it is closer to the "truth" under the KLIC (Greene 2012, 535). Vuong's test is directional, with large positive values favoring the null model while large negative values favor the alternative (and values between -1.96 and +1.96 are inconclusive, for 95% significance). We corroborate results for a subset of the bilateral comparisons using Pesaran's (1974) version of a Cox likelihood ratio test, where the null model is rejected against the alternative if there are large negative values of the test statistic. The test is then reversed to see if the alternative is rejected against the null.

The pairwise comparisons of each model, using Vuong's test to see which is closer to the truth, are reported in Table 4. For ease of interpretation, the models are ordered with those using the most inclusive ranking schemes listed first. The format of the table is that each cell contains a bilateral z-statistic test result, with significant positive values counting as evidence in favor of the model in the column against the model in the row and negative values counting as evidence for the row model against the column model. The model that uses CLm, the medium convexity weights of Combes and Linnemer, is favored against all of the competing models except for the one using MSF weights, for which the comparison is inconclusive (z = -1.40). The comparison of models that use the CLm and CLh weights to calculate lifetime output yields a significant rejection of the high convexity weights (z = 2.27), which is notable since there is no difference in the set of journals or the rankings and just a difference in the weights. The only other significant results in the table are that the model using KMS weights is rejected against the model using CLh weights, and also weakly rejected against the model using MSF weights (at p = 0.052). (25)

The results in Table 4 suggest that more inclusive journal weighting schemes are most congruent with salaries of University of California economists. To see how robust this finding is, Cox-Pesaran tests were carried out to compare the models using CLm and CLh weights, and to compare the models using MSF, KMS, and LP94 weights, since these capture the extremes in terms of least and most aggressive down-weighting for lower-ranked journals. The model using CLh weights is rejected against the one using CLm weights (p = 0.00) while there is no reverse rejection (p = 0.30). Similarly, the model using LP94 weights is rejected against the model using CLm weights (p = 0.00) but not the reverse (p = 0.26). When the least aggressive MSF weights are used, the model using KMS weights (the second most aggressive) is rejected against it (p = 0.00) while there is no reverse rejection (p = 0.23), and the models with MSF weights and LP94 weights reject against each other. Thus the congruence of the less-aggressive journal weighting schemes with the salary data appears to be a robust finding that does not depend on using just one type of non-nested test.

B. Should Older Articles Be Discounted?

The results reported thus far treat an article published in, say, 1978 the same as one from 2008; adjustments are made for length, coauthors, page sizes, and journal quality, but not for vintage. To test if this assumption of no age discounting is appropriate, we calculated for each article published by each academic in year t:

Article Pages x (1 /[(Age).sup.[delta]])

X SizeCorrection

X (1/number of authors)

x Journal Assessment Weight

where Age = 2011-1, and the age discount factor, 8 varied from 0 to 2, in increments of 0.1. In other words, we allowed for no age discounting ([delta] = 0), for inverse age discounting ([delta]=1) where a 20-year-old article has 1/20th the impact on current salary of a 1-year-old article, and for a variety of more extreme and intermediate cases. The best-fitting model, using CLm weights, was estimated for each of these 21 values of [delta] and the maximized log-likelihoods are compared in Figure 2. There appears to be weak age discounting, with the log-likelihood maximized at [delta] = 0.4, which is four-points above the value at [delta] = 0. The maximized likelihood declines steeply at higher discount rates. (26)

[FIGURE 2 OMITTED]

To check if our finding of greater congruence between salary and the less aggressive journal weighting schemes is robust to different assumptions about the age discounting of articles, the academic earnings equations were re-estimated. Since there is no rule-of-thumb for (1/[(Age).sup.0.4]) we discounted according to the inverse of the square root of age, (1/[(Age).sup.0.5]) noting that there was only a half-point difference in the maximized log-likelihoods at 8 = 0.4 and 0.5. The full set of estimation results are reported in Table S2 and the results of the Vuong non-nested tests are in Table 5.

The salary data continue to favor the less aggressive journal weighting schemes when age-discounting of articles is allowed for. The model using the least aggressive MSF weights previously weakly rejected against two models but it now rejects against four (KMS, Coupe, and both

Kodrzycki and Yu schemes). Similarly the model with CLm weights now rejects against six models (and against the model using Coupe weights at p = 0.118). Moreover, the models using six of the assessment schemes now reject against the model using the KMS weights (which are the second most aggressive), whereas previously only three models rejected against this scheme. The final change caused by allowing age discounting is that the model using the simple RePEc impact factors now rejects against three others (both Kodrzycki and Yu schemes and KMS) whereas previously that model rejected against no others.

C. Do Results Depend on Co-Authored Articles Being Fully Pro-Rated?

The results reported thus far rely on applying the "1/n rule" to fully pro-rate co-authored articles. Liebowitz (2013) argues that full proration is required for researchers to choose efficient sized teams, and early empirical results from regressing salaries on co-authored articles were consistent with this (Sauer 1998). (27) But more recent salary studies reject proration and support the hypothesis that each author of a co-authored paper receives the same value as for a sole-authored paper (Hilmer, Hilmer, and Ransom 2012). Similarly, in a small survey of department chairs, Liebowitz (2013) reports that each author of a paper with two authors receives an average of 89% of the value of a sole-authored paper. In a similar survey 30 years earlier the average was 70%, so the trend within economics appears to be away from full proration.

To see what the data indicate about proration, and to check robustness of our key finding that salary data are most supportive of the more inclusive journal weighting schemes, we calculated for each article published by each academic in year t:

Article Pages x (1/[(Age).sup.[delta]])

X Size Correction

x (1/number of authors).sup.[alpha]]

X Journal Assessment Weight

where the co-authorship discount factor, a varied from 0 (no proration) to 1 (full proration), and age discount factors of 0 and [delta] = 0.5 are used. The best-fitting model, using CLm weights, was re-estimated with a incrementing by 0.05 each time, with Figure 3 displaying the maximized log-likelihoods. (28) If age discounting is ignored, the salary data are consistent with [alpha] = 0.25 (although the log likelihood is fairly flat over 0 [less than or equal to] [alpha] [less than or equal to] 0.4) and significantly reject full proration. With [alpha] = 0.25 a three-authored paper gives each author three-quarters of the value of a sole-authored paper, while the "1/n rule" would give each author only one-third of the value. However, the significant negative correlation between the age of articles and the number of coauthors (r = -0.23) makes results on co-author discounting sensitive to how age discounting is treated. Specifically, if articles are discounted by the square root of their age ([delta] = 0.5) the salary data are more consistent with full proration, with the log-likelihood maximizing at [alpha] = 0.7 and no significant difference from [alpha] = 1. A comparison of Figures 2 and 3 suggests that age discounting makes a bigger difference to the fit of an academic salary model than co-authorship discounting; also, using age-discounted articles raises log-likelihood by an average of 1.4 points.

Although the results do not provide clear evidence for or against use of the "1/n rule," in terms of the main finding of this article this ambiguity does not matter. Irrespective of whether we use age discounted articles and [alpha] = 0.7 for the degree of proration, or no age discount and [alpha] = 0.25, the results of the non-nested testing are very similar to what is reported in Tables 4 and 5. (29) The salary model with CLm weights is favored over all competing models (at p < 0.05), except for the model using MSF weights where the comparison is inconclusive (z = -1.13), and the model using KMS weights is also rejected against models using CLh and MSF weights. This is the same pattern as shown in Table 4, so we view our finding that the labor market data are most consistent with the more inclusive journal weighting schemes that do not aggressively down-weight lower-ranked journals as a robust one.

[FIGURE 3 OMITTED]

V. EXTENSIONS

Our main research question of what the academic labor market says about various journal ranking schemes has been answered. Using salaries from economics departments in the University of California as our labor market indicator, we find support for those journal ranking schemes that are more inclusive and which do not aggressively penalize publication in lower-ranked journals. However, our salary regressions can also answer several related questions, which we consider in this section.

A. Are Results Different for the Best Departments?

Combes and Linnemer (2010) suggest their high convexity index is useful to compare the best departments and their medium convexity index is suitable for middle-ranked departments. We therefore examine whether the finding that salaries are best explained by journal ranking schemes that have the lowest rate of decline in the journal weights when moving down the ranks is also found if we restrict attention to the top four economics departments in the University of California system: Berkeley, San Diego, Davis, and Los Angeles. In keeping with the results on age discounting, the published output to date of each academic is calculated using the inverse of the square root of the age of each article. The full estimation results are reported in Table S3 and the results of the Vuong non-nested tests for the subsample of academics in the top four departments are in Table 6.

There is no evidence that salaries of academics in the top four departments are more congruent with the journal ranking schemes that more heavily penalize the lower-ranked journals. In fact, the Vuong tests suggest that the salary regression that uses the MSF rankings, which have the smallest elasticity of the journal weight to the journal rank, now rejects against six of the other regression models (and is inconclusive against the models with CLm and LP94 weights). Moreover, the Cox-Pesaran test shows the model using the LP94 weights, which most heavily penalize lower-ranked journals, is rejected against the model using MSF weights (p = .00) but the reverse rejection is only weakly significant (p = .07). In terms of the specific claim of Combes and Linnemer that their high convexity weights are most suitable for comparing the best departments, the salary evidence suggests the opposite. Under both the Vuong test (z = 3.00) and the Cox-Pesaran test (p = .00) the model using CLh weights is significantly rejected against the model using CLm weights. In other words, the pattern of salaries in the top four economics departments in the UC system are no more supportive of journal ranking schemes that heavily penalize the lower-ranked journals than is the pattern of economist salaries in the UC system as a whole. A similar result could be found by interacting a dummy variable for being in the top four departments with the journal quality measures, and regardless of whether age discounted or not, none of these interaction variables are statistically significant. (30)

B. What Weight Should Be Placed on AER Papers and Proceedings?

The journal assessment schemes that we use differ in their treatment of the Papers and Proceedings May issue of the AER. While seven of the schemes do not discriminate, KMS places zero weight on articles in the May issue and LP94 gives them about one-quarter of the weight of ordinary issues. The Papers and Proceedings issue is a common outlet for the University of California economists in our sample, with 160 articles published there (and 242 in the other issues of the AER). We therefore use our academic earnings equations to see what the data indicate about the appropriate weight to place on Papers and Proceedings compared with ordinary issues of the AER. Our best-fitting model, using CLm weights with articles discounted according to the inverse of the square-root of their age, was re-estimated 101 times, incrementally decreasing the weight for articles in the Papers and Proceedings issue from 100% of an ordinary AER down to zero. The maximized log-likelihoods from this search procedure are illustrated in Figure 4.

The salary data suggest that treating one page in the Papers and Proceedings issue as equivalent to 58% of a page in the other issues of the AER is most appropriate. Thus journal assessment schemes that do not discriminate between the May issue and other issues (likely because they rely on ISI data that do not distinguish journal issue numbers when counting citations) would seem to overstate the impact of Papers and Proceedings. On the other hand, the two assessment schemes that do discriminate may have down-weighted Papers and Proceedings too heavily. Nevertheless we caution that the difference in the maximized log-likelihoods is very small, across any of the values of the relative weight term for Papers and Proceedings. These small differences suggest that even for journals in which our sample have published a large number of articles, deriving journal-specific impact factors from the salary data--what might be dubbed "Market Impact Factors"--may be difficult, despite our statistical power to discriminate between different schemes for weighting the entire spectrum of journals.

C. Does the Portfolio Matter?

Our focus thus far has been on the total published output (weighted by journal quality, age of the article and number of co-authors) without considering the portfolio of articles. This assumes that it is possible to substitute more articles in lower-ranked journals for fewer, higher-ranked articles in order to achieve the same salary. But instead it may be that failure to publish at least one article in the top journals influences tenure, hiring, or promotion decisions at good departments (Card and DellaVigna 2013). In this case, failure to publish in such journals would be expected to impose a penalty even if overall research output would suggest the same salary. To consider this issue we study the influence of not having published in the top three journals; the AER (not including the Papers and Proceedings issue), JPE, and QJE, and also in the top five by adding Econometrica and the Review of Economic Studies. (31)

As shown in Table 7 a significant faction of University of California economists do not (yet) have publications in these journals; 44.4% of all faculty and 28.8% of full professors do not have at least one article in the top three journals. Setting the bar as ever publishing in the top five journals, there are still 32.7% of all faculty and 17.3% of full professors who do not meet this standard. (32) To estimate the effects on salary we re-estimated the salary regressions reported in Table S2, adding a dummy variable for the economists with no articles in the top three journals or for those with no articles in the top five. As shown in the table there is no statistically significant penalty for not having published in any of the top three journals conditional on overall research output. (33) But when we consider the economists not having any articles in any of the top five journals, there does appear to be a salary penalty, of approximately 5%-7%.

To consider this result further we look at the impact of not having an article in each of the top five journals separately. Conditional on overall research output, not having an Econometrica article in a portfolio imposes a penalty of 11% -13% across the nine salary models considered. (34) In contrast, none of the top three journals show any statistically significant salary penalty from being absent from a portfolio while the Review of Economic Studies shows a small penalty, which is statistically significant in just one of the nine salary models. We interpret this as a premium for demonstrating the technical competence and theoretical rigor that is typically associated with Econometrica articles. In contrast, AER, JPE, and QJE are closer substitutes for each other, and a dominant trend for all three of these journals over recent decades has been the rising share of empirical articles (Hamermesh 2013). Consequently, an article in any of the top three journals may not stand out in influencing the perception of a research record in the way that an Econometrica article stands out. Once this perception is gained, there is no further gain from additional Econometrica articles that is over and above the payoff to published research which is already captured by the number of quality-, age- and co-author-adjusted journal pages.

VI. CONCLUSIONS

In this article, we have compared nine different sets of economics journal assessment measures to find which is most consistent with labor market outcomes. These journal assessment measures differ according to the journals covered, the ranking, and the rate at which the weights decline for lower-ranked journals. The most aggressive schemes, in terms of either ignoring or down-weighting lower-ranked journals, exclude more than one-quarter of the lifetime output of our sample of University of California economists and imply a very substantial penalty for publishing in lower-ranked journals. For example, an article in a journal like Economic Inquiry that is just outside the top 50 journals is equivalent to less than 10% of a similarly sized article in the American Economic Review or the Quarterly Journal of Economics.

The clear picture that emerges from the empirical results is that the labor market does not reward publication in ways consistent with the weights implied by the most aggressive journal assessment measures. Instead, it is more inclusive schemes, where the weights for each journal do not decline sharply with the rank of the journal, such as from Mason, Steagall, and Fabritius (1997) and the CLm index of Combes and Linnemer (2010), that have the greatest congruence with academic salaries. This finding is robust to different assumptions about the age discounting and co-authorship discounting of articles and also holds if we restrict attention just to the best departments. Indeed, this last result, that a model using high convexity weights is rejected against a less convex alternative if tested on the top four University of California economics departments is contrary to the claim of Combes and Linnemer (2010) that a high convexity index is more suited for comparing the best departments.

We view this congruence with labor market information as an important criterion for the reasonableness of a journal ranking scheme. While journal ranking schemes that are very aggressive in down-weighting lower-ranked journals can be derived from citation data, such data may be manipulated by editors who coerce authors to add superfluous citations and by authors and editors who collude in cross-citation clubs to raise the citation counts for particular journals or particular individuals. In contrast, labor market decisions have costly consequences, so the evidence for more inclusive journal ranking schemes that is revealed by labor market outcomes is a pattern that comes from information that should be less prone to strategic manipulation.

ABBREVIATIONS
AER:   American Economic Review
AIC:   Akaike's Information Criteria
BIC:   Bayesian Information Criteria
CVs:   Curriculum Vitae
KLIC:  Kullback-Leibler Information Criterion
QJE:   Quarterly Journal of Economics
RePEc: Research Papers in Economics


doi: 10.1111/ecin.12107

Online Early publication May 29, 2014
APPENDIX

TABLE A1

Variable Definitions, Means (M), and Standard Deviations (SD)

Variable                 M       SD     Description

Salary                 156.66   57.07   Base salary in 2010 ($000)
Log (annual salary)    11.90    0.35    Logarithm of 2010 base salary
Experience (years)     18.00    12.43   Years since first appointment
                                          (or receipt of PhD if
                                          earlier)
Seniority (years)      12.15    9.97    Years of employment at
                                          current university
Male                    0.83    0.37    Person is male (=1) or
                                          female (=0)
Holder of a named       0.19    0.40    Person holds an endowed or
  chair                                   named position or a
                                          distinguished chair
Not standard pay        0.03    0.16    Person is not on a standard,
  scale                                    9-month, academic year
                                          pay scale
Nobel Prize winner      0.00    0.07    Winner of the Nobel Prize
PhD field not           0.08    0.27    Person holds a PhD granted
  economics                               from a department that is
                                          not economics
PhD rank (Amir and     31.61    34.55   Score for PhD-granting
  Knauff 2008)                            department (100 = best)
                                          using the placement-based
                                          ranking of Amir and Knauff
                                          (2008)
PhD rank                3.78    1.63    Score for PhD-granting
  (Goldberger.                            department (5 = best) in
  Maher, and                              the 1995 National Research
  Flattau 1995)                           Council rankings

Note: N = 223.


REFERENCES

Abelson, P. "The Ranking of Economics Journals by the Economic Society of Australia." Economic Papers, 28(2), 2009, 176-80.

Amir, R., and M. Knauff. "Ranking Economics Departments Worldwide on the Basis of PhD Placement." Review of Economics and Statistics, 90(1), 2008, 185-90.

Bratsberg, B., J. Ragan, and J. Warren. "Negative Returns to Seniority: New Evidence in Academic Markets." Industrial and Labor Relations Review, 56(2), 2003, 306-23.

--. "Does Raiding Explain the Negative Return to Seniority." Economic Inquiry, 48(3), 2010, 704-21.

Card, D., and S. DellaVigna. "Nine Facts about Top Journals in Economics." Journal of Economic Literature, 51(1), 2013, 144-61.

Chang, C.-L., M. McAleer, and L. Oxley. "What Makes a Great Journal Great in Economics? The Singer Not the Song." Journal of Economic Surveys, 25(2), 2011, 326-61.

Combes, P.-P., and L. Linnemer. "Where Are the Economists Who Publish? Publication Concentration and Rankings in Europe Based on Cumulative Publications." Journal of the European Economic Association, 1(6), 2003, 1250-308.

--. "Inferring Missing Citations: A Quantitative Multi Criteria Ranking of All Journals in Economics." Groupement de Recherche en Economic Quantitative d'Aix Marseille (GREQAM), Document de Travail, no 2010-28, 2010.

Coupe. T. "Revealed Performances: Worldwide Rankings of Economists and Economics Departments, 1990-2000." Journal of the European Economic Association, 1(6), 2003, 1309-45.

--. "What Do We Know about Ourselves? On the Economics of Economics." Kyklos, 57(2), 2004, 197-216.

Ellison, G. "How Does the Market Use Citation Data? The Hirsch Index in Economics." American Economic Journal: Applied Economics, 5(3), 2013, 63-90.

Engemann, K., and H. Wall. "A Journal Ranking for the Ambitious Economist." Federal Reserve Bank of St. Louis Review, 91(3), 2009, 127-39.

Gibson, J. "Research Productivity in New Zealand University Economics Departments: Comment and Update." New Zealand Economic Papers, 34(1), 2000, 73-88.

Goldberger, M. L., B. A. Maher, and P. E. Flattau, eds. Research-Doctorate Programs in the United States: Continuity and Change. National Research Council. Washington, DC: National Academy Press, 1995.

Greene, W. Econometric Analysis. 7th ed. New York: Prentice Hall, 2012.

Hamermesh, D. "Six Decades of Top Economics Publishing: Who and How?" Journal of Economic Literature, 51(1), 2013, 162-72.

Hamermesh, D., and G. Pfann. "Reputation and Earnings: The Roles of Quality and Quantity in Academe." Economic Inquiry, 50(1), 2012, 1-16.

Hamermesh, D., G. Johnson, and B. Weisbrod. "Scholarship, Citations and Salaries: Economic Rewards in Economics." Southern Economic Journal, 49(2), 1982, 472-81.

Hilmer, C., and M. Hilmer. "How Do Journal Quality, Co-authorship, and Author Order Affect Agricultural Economists' Salaries?" American Journal of Agricultural Economics, 87(2), 2005, 509-23.

Hilmer, C., M. Hilmer, and J. Lusk. "A Comparison of Salary Structures between Economics and Agricultural Economics Departments." Applied Economic Perspectives and Policy, 34(3), 2012a, 489-514.

Hilmer, C., M. Hilmer, and M. Ransom. "Fame and the Fortune of Academic Economists: How the Market Rewards Influential Research in Economics." Discussion Paper No. 6960, IZA, Bonn, 2012b.

Hirsch, J. "An Index to Quantify an Individual's Scientific Research Output." Proceedings of the National Academy of Sciences, 102(46), 2005, 16569-72.

Kalaitzidakis, P., T. Mamuneas, and T. Stengos. "Rankings of Academic Journals and Institutions in Economics." Journal of the European Economic Association, 1(6), 2003, 1346-66.

--. "An Updated Ranking of Academic Journals in Economics." Working Paper 9/2010, Economics Department, University of Guelph, Guelph, Canada, 2010.

--. "An Updated Ranking of Academic Journals in Economics." Canadian Journal of Economics, 44(4), 2011, 1525-38.

Kodrzycki, Y., and P. Yu. "New Approaches to Ranking Economics Journals." B.E. Journal of Economic Analysis and Policy: Contributions to Economic Analysis and Policy, 5(1), 2006, Article 24.

Laband, D., and M. Piette. "The Relative Impact of Economics Journals." Journal of Economic Literature, 32(2), 1994, 640-66.

Liebowitz, S. J. "Willful Blindness: The Inefficient Reward Structure in Academic Research." Economic Inquiry, 2013. DOI: 10.1111/ecin. 12039.

Liebowitz, S. J., and J. P. Palmer. "Assessing the Relative Impact of Economics Journals." Journal of Economic Literature, 22(1), 1984, 77-88.

Macri, J., and D. Sinha. "Rankings Methodology for International Comparisons of Institutions and Individuals: An Application to Economics in Australia and New Zealand." Journal of Economic Surveys, 20(1), 2006, 111-56.

Mason, P., J. Steagall, and M. Fabritius. "Economics Journal Rankings by Type of School: Perceptions versus Citations." Quarterly Journal of Business and Economics, 36(1), 1997, 69-79.

Moore. W., R. Newman, and G. Turnbull. "Do Faculty Salaries Decline with Seniority?" Journal of Labor Economics, 16(2), 1998, 352-66.

--. "Reputational Capital and Academic Pay." Economic Inquiry, 39(4), 2001, 663-71.

--. "Academic Pay in the United Kingdom and the United States: The Differential Returns to Productivity and the Lifetime Earnings Gap." Southern Economic Journal, 73(3), 2007, 717-32.

OECD. "Performance-based Funding for Public Research in Tertiary Education Institutions," Workshop Proceedings, OECD Publishing, 2010. Accessed July 6. 2012. http://dx.doi.org/10.1787/9789264094611-en

Oswald, A. "An Examination of the Reliability of Prestigious Scholarly Journals: Evidence and Implications for Decision-Makers." Economica, 74(296), 2007, 21-31.

Pesaran, M. "On the General Problem of Model Selection." Review of Economic Studies, 41(2), 1974, 153-71.

Ransom, M. "Seniority and Monopsony in the Academic Labor Market." American Economic Review, 83(1), 1993, 221-33.

Sauer. R. "Estimates of the Returns to Quality and Coauthorship in Economic Academia." Journal of Political Economy, 96(4), 1998, 856-66.

Tressler, J., and D. Anderson. "The Merits of Using Citations to Measure Research Output in Economics Departments: The New Zealand Case." Agenda, 19(1), 2012, 17-37.

University of California. "Facts About the University of California, Total Compensation for UC Employees," October 2009. Accessed June 6, 2013. http:// compensation.universityofcalifornia.edu_total_comp_ facts_nov2009.pdf.

Vuong, Q. "Likelihood Ratio Tests for Model Selection and Non-nested Hypotheses." Econometrica, 57(2), 1989, 307-33.

Wilhite, A. W., and E. A. Fong. "Coercive Citation in Academic Publishing." Science, 335, 2012, 542-43.

SUPPORTING INFORMATION

Additional Supporting Information may be found in the online version of this article:

Table S1. Salary Regressions for UC Economists with Lifetime Output of Journal Articles Not Age Discounted

Table S2. Salary Regressions for UC Economists with Lifetime Output of Journal Articles Square-Root-Age Discounted

Table S3. Salary Regressions for UC Economists at the Top Four Departments (Articles Square Root-Age Discounted)

(1.) For an international review of performance-based research funding in tertiary education see OECD (2010).

(2.) Government research assessment exercises tend to rely more on peer review assessments. Critiques of the use of journal rankings in evaluating the research they publish are Oswald (2007) and Chang, McAleer, and Oxley (2011).

(3.) Journal rankings based on expert opinion have also been important. Examples in economics include assessments provided by Mason, Steagall, and Fabritius (1997) and the Economic Society of Australia (Abelson 2009).

(4.) See also the supporting analysis available from www.sciencemag.org/cgi/content/full/335/6068/542/DCl.

(5.) The journal Technological and Economic Development of Economy rose to third in the economics category of the ISI Journal Citation Reports in 2010 by having 60% of citations coming from five journals from the same university and another 23% of citations from journals published by a neighboring university.

(6.) Coupe (2004) provides a somewhat dated review of research on the market for academic economists.

(7.) Academic earnings equations have also been used to compare returns to research productivity between countries (Moore, Newman, and Terrell 2007) and to compare rankings of departments (Gibson 2000).

(8.) Kodczychi and Yu also provide assessment measures based on citations from a set of journals they refer to as "policy" oriented, which we do not use in this article.

(9.) This is an update of the widely used Kalaitzidakis, Mamuneas, and Stengos (2003) measures. In our analysis the journal impact factors used are taken from their 2010 Working Paper.

(10.) The list we use was obtained from http//homepages. ulb.ac.be/~tCoupe/update/journals.html (on August 20, 2007).

(11.) Coupe (2003) also uses direct citation measures to provide rankings of departments and individual economists.

(12.) The list we use was obtained on May 6, 2012 and as of July 2012 RePEc had grown to cover 1,004 journals.

(13.) For example, the relative weight for the Southern Economic Journal compared with the AER is from 0.02 to 0.20, with an average of 0.12, for eight of the nine schemes that we use. But the MSF scheme gives it a weight of 0.73.

(14.) These are all academic articles, as captured in the union of the EconLit, RePEc, and Web of Science databases, so the under-coverage of articles is not because they are not part of the peer-reviewed scientific literature.

(15.) Combes and Linnemer (2010, 2) suggest that their high convexity index "is useful to compare the best departments", while their medium convexity index is better suited to study middle-ranked departments.

(16.) See for example, University of California (2009). http://compensation.universityofcalifornia.edu/total_comp_ facts_nov2009.pdf.

(17.) This calculation is limited to individuals with 3 years of data for each type of salary and with no decline in base salary over time (which may signal only a partial year's employment as could occur from someone moving to another position). For these individuals, the squared deviation of annual salary from the 3-year average has a mean (median) of $1,443 ($312) for total salary and $1,093 ($213) for base salary (all in millions).

(18.) At UC Berkeley the salaries for the nine steps of the Faculty-Ladder ranks for the Law School Professor Series are about $40,000 higher per academic year than the same nine steps on the series for Business/Economics/Engineering.

(19.) We did gather the number of authored books (but not edited volumes) from EconLit but this proved to have no explanatory power in the earnings equations.

(20.) In a few cases the 2010 base salary was lower than in 2009 or 2007, which may signal partial year employment as could occur from someone moving to another position, so for these individuals we used their maximum base salary from 2007 or 2009.

(21.) Of course some of these factors may also differ within departments. For example, faculty with good PhD students who publish in the best journals may have reduced salary relative to others who get paid more as compensating differential for working with lower quality students who publish in lower-ranked journals. The RePEc Genealogy database may eventually allow linking all of these academics with their students, as a topic for future research.

(22.) Page correction factors were supplied by Joseph Macri, based on his work with the late Dependra Sinha. A value of 0.72 was used for journals with no factor available. This is the average page size for the lowest-ranked journals in Gibson (2000), and these are typically the ones without their own size correction factors available.

(23.) This result is broadly consistent with other estimates in the literature of the value of research output in the academic labor market. See Coupe (2004,208) for a brief review of such results.

(24.) The UCLA fixed effects become larger (and more significant) positive values once output measures are included.

(25.) The model using Coupe weights is also weakly rejected against the model using MSF weights (at p = .099).

(26.) If the models with different values of 6 were instead estimated after using the CLh weights to summarize lifetime output, the log-likelihood would maximize at [delta] = 0.3 (although at just 0.35 points above the log-likelihood at [delta] = 0.4). So this evidence of weak age discounting is not specific to the journal weighting scheme used.

(27.) See Coupe (2004, 202) for an overview of the earlier literature on co-author discounting.

(28.) We also used CLh weights and included 8 = 0.3 (the age discount factor that maximized the likelihood for those weights) and find very similar patterns to what is shown in Figure 3.

(29.) The results of these sensitivity analyses are available from the authors.

(30.) The result closest to statistical significance comes from the salary regressions using the Kodrzycki and Yu (2006) scheme that is restricted to economics journals, where the interaction variables have f-statistics of 1.56 (age discounted) and 1.79 (no age discounting).

(31.) These five journals are the top journals whose trends have recently been described by Card and DellaVigna (2013).

(32.) Professors with no top five journal articles are predominantly at UC Riverside, but a total of eight UC departments had at least one of these professors without an article in a top five journal.

(33.) This noneffect is the same no matter which journal ranking scheme is used for calculating research output.

(34.) We also considered whether the salary data favor modelling a penalty for not having an article in Econometrica or a premium for each Econometrica article. For the best-fitting salary regression with CLm weights, Cox-Pesaran non-nested tests show a salary model using the number of Econometrica articles is rejected against a model using the dummy for no Econometrica articles (p = .00) while there is no reverse rejection (p = .30).

JOHN GIBSON, DAVID L. ANDERSON and JOHN TRESSLER, We are grateful to the editor, two anonymous referees, Peter Timmer, Dean Scrimgeour, Izi Sin, and audiences at the Melbourne Institute, Motu and the University of Canterbury for helpful comments.

Gibson: Professor, Department of Economics, Waikato Management School, University of Waikato, Hamilton 3240, New Zealand. Phone 64-7-838-4289, Fax 64-7-829-5931, E-mail jkgibson@waikato.ac.nz

Anderson: Emeritus Professor, School of Business, Queen's University, Kingston, Ontario K7L 3N6, Canada. Phone 1-790-470-0408, Fax 1-613-533-2301, E-mail dla@queensu.ca

Tressler: Associate Professor, Department of Economics, Waikato Management School, University of Waikato, Hamilton 3240, New Zealand. Phone 64-7-856-2889, Fax 64-7-838-4331, E-mail tressler@waikato.ac.nz
TABLE 1

Indicators of the Inclusiveness of Various Journal Ranking Schemes,
as Applied to Publications of UC Economists

                                        Percentage of
                                      Lifetime Articles
Source of Impact Factor               with Zero Weight

MSF: Mason, Steagall, and Fabritius         33.3
CLm: Coomes-Linnemer (medium)               10.9
CLh: Coomes-Linnemer (high)                 10.9
RePEc Simple Impact Factor                  12.6
Coupe (2003)                                22.0
K&Y_all: Kodrzycki and Yu (2006)            26.4
K&Y_econ: Kodrzycki and Yu (2006)           26.0
KMS: Kalaitzidakis. Mamuneas,               29.6
  and Stengos (2011)
LP94: Laband and Piette (1994)              36.4

                                             Rank-Elasticity
                                              Regression (a)

Source of Impact Factor               Elasticity    SE    [R.sup.2]

MSF: Mason, Steagall, and Fabritius     -0.22      0.01     0.74
CLm: Coomes-Linnemer (medium)           -0.67      0.01     0.97
CLh: Coomes-Linnemer (high)             -1.35      0.01     0.97
RePEc Simple Impact Factor              -1.17      0.04     0.76
Coupe (2003)                            -0.78      0.03     0.81
K&Y_all: Kodrzycki and Yu (2006)        -1.48      0.05     0.85
K&Y_econ: Kodrzycki and Yu (2006)       -1.53      0.05     0.84
KMS: Kalaitzidakis. Mamuneas,           -1.74      0.07     0.80
  and Stengos (2011)
LP94: Laband and Piette (1994)          -1.92      0.09     0.82

                                      Average Lifetime
                                       Output for UC
Source of Impact Factor                Economists (b)

MSF: Mason, Steagall, and Fabritius        143.6
CLm: Coomes-Linnemer (medium)              106.3
CLh: Coomes-Linnemer (high)                 69.7
RePEc Simple Impact Factor                  65.2
Coupe (2003)                                45.5
K&Y_all: Kodrzycki and Yu (2006)            40.0
K&Y_econ: Kodrzycki and Yu (2006)           37.0
KMS: Kalaitzidakis. Mamuneas,               38.2
  and Stengos (2011)
LP94: Laband and Piette (1994)              35.9

Note: Author's calculations based on 5,721 journal articles
produced by 213 University of California economists and journal
weights from the sources noted.

(a) Estimated over the journals with non-zero weights for each
scheme using log(relative weight) = [alpha] + [beta]log(rank).

(b) Total number of standardized pages (with co-authors given 1
In) published in career through 2010, where journals are weighted
such that the highest-ranked journal for each scheme has weight
1.0 and there is no age-discounting for older articles.

TABLE 2

Salary Regressions for UC Economists: Individual Characteristics,
Salary Attributes, and LocationFixed Effects

                                     (1)          (2)          (3)

Seniority (years)                   -0.020       -0.020       -0.020
                                  (3.05) **    (3.02) **    (3.04) **
Seniority squared (/100)            0.023        0.023        0.024
                                    (1.21)       (1.22)       (1.24)
Experience (years)                  0.052        0.052        0.052
                                  (10.94) **   (10.99) **   (10.97) **
Experience squared (/100)           -0.072       -0.072       -0.072
                                  (6.06) **    (6.14) **    (6.17) **
Male                                0.060        0.058        0.058
                                    (1.47)       (1.41)       (1.43)
PhD field not economics             -0.037
                                    (0.90)
PhD rank (Amir and Knauff 2008)                  0.000
                                                 (0.33)
PhD rank (Goldberger, Maher,                                  0.001
and Flattau 1995)
                                                              (0.08)
Nobel prize winner

Holder of a named chair

Not standard pay scale

Davis                               -0.449       -0.445       -0.448
                                  (8.71) **    (8.38) **    (8.65) **
Irvine                              -0.392       -0.386       -0.391
                                  (8.74) **    (8.27) **    (8.81) **
Merced                              -0.454       -0.458       -0.465
                                  (9.08) **    (8.10) **    (8.62) **
Riverside                           -0.472       -0.466       -0.472
                                  (8.83) **    (8.17) **    (8.63) **
San Diego                           -0.200       -0.199       -0.200
                                  (3.85) **    (3.78) **    (3.84) **
Santa Barbara                       -0.406       -0.400       -0.405
                                  (7.86) **    (7.50) **    (7.84) **
Santa Cruz                          -0.439       -0.435       -0.440
                                  (7.87) **    (7.69) **    (7.93) **
Los Angeles                         0.033        0.036        0.033
                                    (0.60)       (0.65)       (0.60)
Constant                            11.696       11.689       11.692
[R.sup.2]                         (210.07) **  (202.75) **  (189.47) **
                                     0.70         0.69         0.69

                                     (4)          (5)

Seniority (years)                   -0.020       -0.022
                                  (2.95) **    (3.38) **
Seniority squared (/100)            0.024        0.029
                                    (1.25)       (1.61)
Experience (years)                  0.052        0.051
                                  (10.72) **   (10.21) **
Experience squared (/100)           -0.073       -0.074
                                  (6.06) **    (6.16) **
Male                                0.058        0.066
                                    (1.44)      (1.65)+
PhD field not economics

PhD rank (Amir and Knauff 2008)

PhD rank (Goldberger, Maher,
and Flattau 1995)

Nobel prize winner                  0.431
                                  (7.37) **
Holder of a named chair                          0.146
                                               (3.84) **
Not standard pay scale

Davis                               -0.449       -0.410
                                  (8.67) **    (8.16) **
Irvine                              -0.389       -0.362
                                  (8.71) **    (7.89) **
Merced                              -0.462       -0.445
                                  (9.07) **    (6.61) **
Riverside                           -0.470       -0.412
                                  (8.61) **    (7.42) **
San Diego                           -0.200       -0.142
                                  (3.85) **    (2.72) **
Santa Barbara                       -0.422       -0.358
                                  (8.46) **    (7.07) **
Santa Cruz                          -0.440       -0.384
                                  (7.99) **    (7.30) **
Los Angeles                         0.034        0.069
                                    (0.62)       (1.27)
Constant                            11.695       11.662
[R.sup.2]                         (209.67) **  (207.02) **
                                     0.70         0.71

                                     (6)          (7)

Seniority (years)                   -0.020       -0.021
                                  (3.06) **    (3.28) **
Seniority squared (/100)            0.024        0.030
                                    (1.26)       (1.60)
Experience (years)                  0.052        0.050
                                  (10.79) **   (9.89) **
Experience squared (/100)           -0.072       -0.075
                                  (6.08) **    (6.02) **
Male                                0.065        0.071
                                    (1.61)      (I.81)+
PhD field not economics

PhD rank (Amir and Knauff 2008)

PhD rank (Goldberger, Maher,
and Flattau 1995)

Nobel prize winner                               0.346
                                               (6.36) **
Holder of a named chair                          0.137
                                               (3.70) **
Not standard pay scale              0.113        0.115
                                   (2.17) *     (2.07) *
Davis                               -0.432       -0.397
                                  (7.97) **    (7.62) **
Irvine                              -0.376       -0.347
                                  (7.99) **    (7.18) **
Merced                              -0.451       -0.428
                                  (8.35) **    (6.38) **
Riverside                           -0.463       -0.404
                                  (8.39) **    (7.19) **
San Diego                           -0.185       -0.131
                                  (3.43) **     (2.45) *
Santa Barbara                       -0.389       -0.359
                                  (7.18) **    (7.00) **
Santa Cruz                          -0.425       -0.373
                                  (7.44) **    (6.92) **
Los Angeles                         0.048        0.084
                                    (0.84)       (1.49)
Constant                            11.678       11.647
[R.sup.2]                         (202.31) **  (202.41) **
                                     0.70         0.72

Notes: Dependent variable is log of base salary for the 2010
academic year, as reported at: http://www.sacbee.com/statepay/
with economists at UC Berkeley as the excluded group for the
fixed effects. N = 223, robust t statistics in parentheses.

(+) Significant at 10%; *significant at 5%; **significant at 1%.

TABLE 3

Comparisons of Academic Earnings Equations Using Different
Journal Assessment Weights to Compute Lifetime Output

                                            Academic Earnings
                                            Equation Regression

                         Semi-Elasticity (a)   Robust SE    [R.sup.2]

MSF: Mason, Steagall            0.009            0.001        0.77
  and Fabritius
CLm:                            0.013            0.002        0.78
  Coomes-Linnemer
  (medium)
CLh:                            0.017            0.003        0.77
  Coomes-Linnemer
  (high)
RePEc Simple Impact             0.016            0.002        0.77
  Factor
Coupe(2003)                     0.023            0.004        0.76
K&Y all: Kodrzycki              0.023            0.004        0.76
  and Yu(2006)
K&Y_econ:                       0.025            0.004        0.76
  Kodrzycki and Yu
  (2006)
KMS: Kalaitzidakis,             0.024            0.005        0.76
  Mamuneas, and
  Stengos (2011)
LP: Laband and Piette           0.027            0.004        0.77
  (1994)

                                                          Bayesian
                           Maximized      Akaike's Info     Info
                         Log-Likelihood     Criteria      Criteria

MSF: Mason, Steagall         83.66           -133.33       -75.40
  and Fabritius
CLm:                         87.49           -140.99       -83.07
  Coomes-Linnemer
  (medium)
CLh:                         83.06           -132.12       -74.20
  Coomes-Linnemer
  (high)
RePEc Simple Impact          80.00           -126.00       -68.07
  Factor
Coupe(2003)                  79.28           -124.57       -66.65
K&Y all: Kodrzycki           79.07           -124.13       -66.21
  and Yu(2006)
K&Y_econ:                    78.92           -123.84       -65.92
  Kodrzycki and Yu
  (2006)
KMS: Kalaitzidakis,          75.63           -117.27       -59.34
  Mamuneas, and
  Stengos (2011)
LP: Laband and Piette        80.27           -126.53       -68.61
  (1994)

Notes: The results are from nine separate regressions, where each
includes all of the variables in column 7 of Table 2 plus the
total number of standardized pages (with co-authors given 1 In)
published in each economist's career through 2010, where journals
are weighted such that the highest-ranked journal for each scheme
has weight 1.0 and there is no age-discounting for older
articles. Full results of the regressions are reported in Table
SI. N = 223.

(a) The semi-elasticity shows the percentage increase in annual
(academic year) salary for a 10-page increase in total career
output of weighted journal articles, where the weight for each
journal in each ranking scheme is used to convert into pages of
QJE-equivalent quality.

TABLE 4

Vuong Test Results Comparing Academic
Earnings Functions with Different Journal
Assessment Weights Used to Calculate Lifetime
Output

MSF         (a)    (a)     (b)    (c)     (d)
CLm         (b)   -1.40
CLh         (c)    0.17    2.27#
RePEc       (d)    1.28    3.00#   1.19
Coupe       (e)    1.65@   2.41#   0.95    0.29
K&Y_all     (f)    1.31    2.80#   1.41    0.56
K&Y_econ    (g)    1.34    2.81#   1.46    0.68
KMS         (h)    1.94@   3.49#   3.11#   1.45
LP94        (i)    0.80    2.24#   1.11   -0.08

MSF          (e)     (f)      (g)      (h)
CLm
CLh
RePEc
Coupe
K&Y_all      0.08
K&Y_econ     0.13    0.49
KMS          0.89    1.13     1.11
LP94        -0.23   -0.37    -0.42    -1.37

Notes: Cell values are z-statistics, calculated from the models
reported in Table S1. Significant positive values favor the model
in the column against the model in the row and negative values
favor the row model over the column model. Test values in bold
are statistically significant at 5% level, those in italics are
significant at 10% level.

Note: Test values in bold are indicated with #.

Note: Those in italics are indicated with @.

TABLE 5

Vuong Test Results When Lifetime Output Is
Calculated with Journal Articles
Square-Root-Age Discounted

MSF         (a)    (a)     (b)    (c)     (d)
CLm         (b)   -0.18
CLh         (c)   1.45    3.29#
RePEc       (d)   1.20    1.76@   -0.70
Coupe       (e)   1.71@   1.56    -0.14   0.47
K&Y_all     (f)   2.22#   2.69#   1.01    2.33#
K&Y_econ    (g)   2.14#   2.60#   0.92    2.21#
KMS         (h)   2.80#   3.94#   3.26#   2.62#
LP94        (i)   1.47    1.93@   0.18    0.68

MSF          (e)     (f)     (g)     (h)
CLm
CLh
RePEc
Coupe
K&Y_all     1.42
K&Y_econ    1.30    -0.72
KMS         1.88@   1.36    1.44
LP94        0.26    -0.81   -0.72   -1.89@

Notes: Cell values are z-statistics, calculated from the models
reported in Table S2. Significant positive values favor the model in
the column against the model in the row and negative values favor the
row model over the column model. Test values in bold are statistically
significant at 5% level, those in italics are significant at 10% level.

Note: Test values in bold are indicated with #.

Note: those in italics are indicated with @.

TABLE 6

Vuong Test Results for the Subsample at the Top
Four University of California Departments
(Lifetime Output Calculated with Journal
Articles Square-Root-Age Discounted)

MSF        (a)    (a)     (b)     (c)     (d)
CLm        (b)   0.95
CLh        (c)   2.20#   3.00#
RePEc      (d)   1.67@   1.00    -1.21
Coupe      (e)   2.04#   0.98    -0.51    0.41
K&Y_all    (f)   2.64#   2.09#    0.71    2.20#
K&Y_econ   (g)   2.57#   2.02#    0.63    2.09#
KMS        (h)   3.03#   3.35#    2.86#   2.69#
LP94       (i)   1.35    0.85    -0.94    0.13

MSF         (e)     (f)     (g)     (h)
CLm
CLh
RePEc
Coupe
K&Y_all     1.75@
K&Y_econ    1.62    -0.63
KMS         2.13#    1.49    1.55
LP94       -0.15    -1.38   -1.30   -2.42#

Notes: Cell values are z-statistics, calculated from the models
reported in Table S3. Significant positive values favor the model
in the column against the model in the row and negative values
favor the row model over the column model. Test values in bold
are statistically significant at 5% level, those in italics are
significant at 10% level.

Note: Test values in bold are indicated with #.

Note: those in italics are indicated with @.

TABLE 7

Portfolio Effects from Absence of Top Journals in
Publication Records of UC Economists

                                                      Salary Penalty
                                 Percent without     from Not Having
                                   Articles in      an Article in Top
                                  Top Journals         Journals (a)

                               Full        Full          Using CLm
                              Sample    Professors        Weights

Top three journals             44.4        28.4       -0.002 (0.08)
Top five journals              32.7        17.3       -0.052(1.71) (+)
Quarterly Journal of           70.0        59.1        0.012 (0.42)
  Economics
American Economic Review       56.1        40.2       -0.012 (0.40)
Journal of Political           71.3        57.5       -0.022 (0.67)
  Economy
Econometrica                   72.2        62.2       -0.112(3.48) **
Review of Economic Studies     74.9        66.9       -0.030 (0.93)

                                        Salary Penalty
                                       from Not Having
                                       an Article in Top
                                         Journals (a)

                              Most Significant       Average of
                                    Model             9 Models

Top three journals              -0.025 (0.83)      -0.008 (0.26)
Top five journals               -0.068 (2.21) *    -0.061 (1.95) (+)
Quarterly Journal of            -0.022 (0.75)       0.005 (0.39)
  Economics
American Economic Review        -0.036(1.16)        -0.018(0.55)
Journal of Political            -0.034(1.05)        -0.024 (0.71)
  Economy
Econometrica                   -0.134 (4.15) **     -0.119(3.54) **
Review of Economic Studies     -0.062 (2.01) *      -0.043 (1.33)

Notes: Author's calculations based on 223 University of
California economists. Robust t-statistics in parentheses.

(a) Estimates of the salary penalty are based on the
specifications reported in Table S2, which have been augmented
with a dummy variable for whether the individual had not
published in any of the journals or groups of journals listed in
the first column. These regressions condition on overall research
output (weighted by journal quality, age of the article and
number of co- authors). The results summarize 63 regressions
(seven journals or groups of journals crossed with nine different
journal ranking schemes), with column 4 reporting the results for
the overall best-fitting model (using CLm weights), column 5
reporting the most significant result for that particular
journal(s), and column 6 reporting the average over all nine
regressions for each of the journal(s).

(+) Significant at 10%; * significant at 5%; ** significant at 1%.
联系我们|关于我们|网站声明
国家哲学社会科学文献中心版权所有