首页    期刊浏览 2024年11月15日 星期五
登录注册

文章基本信息

  • 标题:Auction markets for evaluations.
  • 作者:Wilson, Bart J.
  • 期刊名称:Southern Economic Journal
  • 印刷版ISSN:0038-4038
  • 出版年度:2005
  • 期号:July
  • 语种:English
  • 出版社:Southern Economic Association
  • 摘要:As the technology of electronic exchange advances, new opportunities emerge for developing markets for products and services whose innate properties hamper efforts to do so in traditional settings. Even though one such good, product evaluations, has been long used for durable and nondurable goods alike, the transaction costs associated with providing and disseminating product evaluations have limited the scope of their use. Historically, evaluations have been limited to "word of mouth" exchanges among acquaintances and the reports of paid critics. (1) The Internet not only has the potential to significantly reduce the transaction costs for evaluation sharing, but it can also significantly reduce the costs for forming a centralized market mechanism to allocate evaluations. Although numerous evaluations are currently freely available on the Internet for everything from CDs and books to articles and professors, these services are incomplete. (2) For example, Amazon.com provides free book reviews by other customers who voluntarily provided the review, but there are numerous books that have not been reviewed. Ratemyprofessors.com enables students to share evaluations of their college instructors, but not every teacher has been rated. A market for evaluations creates the incentives for individuals to provide evaluations that might otherwise not be provided.
  • 关键词:Auction firms;Financial markets

Auction markets for evaluations.


Wilson, Bart J.


1. Introduction

As the technology of electronic exchange advances, new opportunities emerge for developing markets for products and services whose innate properties hamper efforts to do so in traditional settings. Even though one such good, product evaluations, has been long used for durable and nondurable goods alike, the transaction costs associated with providing and disseminating product evaluations have limited the scope of their use. Historically, evaluations have been limited to "word of mouth" exchanges among acquaintances and the reports of paid critics. (1) The Internet not only has the potential to significantly reduce the transaction costs for evaluation sharing, but it can also significantly reduce the costs for forming a centralized market mechanism to allocate evaluations. Although numerous evaluations are currently freely available on the Internet for everything from CDs and books to articles and professors, these services are incomplete. (2) For example, Amazon.com provides free book reviews by other customers who voluntarily provided the review, but there are numerous books that have not been reviewed. Ratemyprofessors.com enables students to share evaluations of their college instructors, but not every teacher has been rated. A market for evaluations creates the incentives for individuals to provide evaluations that might otherwise not be provided.

Avery, Resnick, and Zeckhauser (1999) provide a first step in creating a pricing mechanism to induce the efficient provision of evaluations. They discuss how, without a market for evaluations, risk neutral agents will provide a suboptimal level of evaluations because consumption is nonrivalrous (3) and because one person's experience might not perfectly predict another's outcome. The nonrivalrous nature of evaluations could also result in an inefficient ordering of the evaluators themselves, if the potential evaluators have heterogeneous opportunity costs for producing an evaluation. The optimal quantity can also depend on what the early evaluations reveal about the value of the product.

Avery, Resnick, and Zeckhauser (1999) show that a market mechanism can be used to solve these various social problems. They consider both sequential and batch evaluation games and demonstrate that a centralized broker who knows the pool of values and opportunity costs for the players can set a price for providing and receiving information so as to eliminate the social inefficiency. This paper relaxes the assumption that a central broker holds all of the pertinent information, which, as a point of practicality, would not hold in an actual implementation of such a mechanism. Using the experimental method, we study how efficiently four different market mechanisms voluntarily elicit an evaluation in exchange for payment, without any centralized information about the pool of values. This subjects the market mechanisms to the challenging task of inducing the optimal person to voluntarily provide a nonrivalrous evaluation. As an initial foray into implementing a market for evaluations, we consider variations of and then compare the efficiency and prices for uniform price sealed bid, discriminatory price sealed bid, English clock auction, and Dutch clock auction.

We follow Avery, Resnick, and Zeckhauser (1999) in considering a framework in which people have already made the decision to enter a product market. As a simplifying assumption for this exploratory research, the individuals in the markets have identical tastes for whether a product is "good" or "bad" and the same ability to discern those tastes. Individuals, however, differ in their values for a "good" or "bad" product. Thus, unlike Avery, Resnick, and Zeckhauser (1999), a single evaluation perfectly reveals whether the product is "good" or "bad." This assumption eliminates the inefficient ordering and optimal quantity problems and makes the underprovision problem binary, thereby allowing us to focus on the crucial issue of how the mechanisms aggregate private information, determine prices, and affect efficiency.

In our controlled test, we find that the provision of evaluations is markedly inefficient without a market mechanism, but not nearly as inefficient as what is predicted. Additionally, we find that each of the four mechanisms succeeds at increasing market efficiency by encouraging the socially optimal agent to undertake the costly evaluation when no one else is willing to do so. Finally, we observe that the four mechanisms are behaviorally equivalent with respect to the prices received by the evaluator. The structure of the paper is as follows. Section 2 presents the experimental design that we consider, and section 3 outlines the treatments and procedures. Section 4 discusses our results, and section 5 briefly concludes.

2. Experimental Design

Suppose there are two identical risk-neutral individuals considering reading a book. A particular book could be "good," resulting in a payoff of g > 0, or "bad," resulting in a payoff of b < 0 with [absolute value of b] > [absolute value of g]. If each outcome is equally likely, then both individuals have a negative expected value for reading the book and hence should opt for their next best alternative, provided that alternative payoff c is greater than (g + b)/2. However, it is not necessarily socially optimal for neither person to read the book. Consider the case in which one individual decides to read the book and then provides an evaluation to the other person. In this case, the evaluator expects to receive (g + b)/2, but the person receiving the evaluation can make a more informed decision. This informed person will read the book if it is "good" and receive a payoff of g. However, if the evaluator reports that the book is "bad," the other person will not read it, choosing the opportunity cost c over the bad payoff b. Hence the person receiving the evaluation has an expected payoff of (g + c)/2, and the social payoff from one person evaluating the book is [(g + b)/2] + [(g + c)/2], which is greater than 2c so long as c < (2g + b)/3.

The solution to this social problem lies with the creation of a market that allows one potential consumer to compensate another for undertaking the costly evaluation. Avery, Resnick, and Zeckhauser (1999) present a formal treatment of this problem and demonstrate the existence of an equilibrium price that attains the socially efficient outcome. (4) This price is paid to the evaluator by the person waiting to make an informed decision. In the example presented above, the unique equilibrium price, P, is (c - b)/4. This price is found by equating the expected payoff of the evaluator to the expected payoff of the person who waits, or P insures that (g + b)/2 + P = (g + c)/2 - P, thereby making the two identical people indifferent between evaluating and waiting.

Institutions

Equipped with the theoretical foundation that a market can solve this problem, the next step is to identify what market institutions should be implemented in practice. The identification of the market price in the preceding paragraph requires collective knowledge of three pieces of information for each individual i, namely [g.sub.i], [b.sub.i], and [c.sub.i], plus the probabilities that the product or service is "'good" or "bad." In practice, these pieces of information are typically private and unobservable (or unverifiable). Thus, one role of a functioning market is to aggregate private information and coordinate behavior. Therefore, any market institution must determine (i) who will be the evaluator and (ii) the price paid or received by each agent. As discussed in Smith (1994) the institution can significantly influence behavior and therefore market outcomes. For example, the four most common types of private value auctions, uniform price sealed bid, discriminatory price sealed bid, English clock, and Dutch clock, are all theoretically revenue equivalent under certain assumptions; yet, there is widespread evidence from the laboratory that these distinct formats elicit different behaviors, which affect market performance. (5) This paper takes the next step toward constructing markets for evaluations by developing variants of these four well-known market institutions and comparing the performance of each. An important distinction between our environment and that of the standard private value auction is that the evaluation is nonrivalrous. (6) Within the controlled confines of the laboratory, these wind tunnel tests directly compare these institutions to each other, as well as to a baseline case in which no market exists.

No-market Baseline

In the baseline case, a fictitious product is available for evaluation for a limited time, T. If at any point during this time one of the n individuals consumes and evaluates the product, then the payoff state, good or bad, is revealed to everyone. As a simplification, if the product is good, then everyone who waits receives their own good payoff [g.sub.i], but if the product is bad, all of the people who wait receive their opportunity cost el. The evaluator also receives [g.sub.i] if the product is good, but when it is bad, the evaluator receives [b.sub.i]. In this situation, the dominant strategy is to wait and see whether the others evaluate, regardless of how much time is remaining, assuming that the opportunity cost is sufficiently high. Formally, let [[lambda].sub.i] denote a player i's belief about the probability someone else will evaluate during the remaining time. The expected payoff to waiting is [[lambda].sub.i]([g.sub.i] + [c.sub.i])/2 + (1 - [[lambda].sub.i])[c.sub.i], which is greater than ([g.sub.i] + [b.sub.i])/2, the expected value of evaluating, if [[lambda].sub.i] > 1 + ([b.sub.i] - [c.sub.i])/([g.sub.i] - [c.sub.i]). This condition is equivalent to [[lambda].sub.i] [greater than or equal to] 0 if [c.sub.i] > ([g.sub.i] + [b.sub.i])/2. We now describe four distinct institutions for providing an evaluation.

Uniform Price Sealed Bid Auction

For the uniform price sealed bid auction in an independent private value environment, each bidder privately submits a bid. The bids are arrayed from highest to lowest and the winners are the bidders who submitted the highest bid. Each winner pays the same price, which is determined by the lowest winning bid and/or the highest loosing bid. In a market for a nonrivalrous evaluation, each of the n agents submit a single bid [[beta].sub.i]. This bid represents the minimum amount the person is willing to accept for evaluating the fictitious product. Once the n bidders have submitted their bids, the bids are ranked in ascending order. Let [[beta].bar] and [??] denote, respectively, the lowest and second lowest submitted bids. The agent submitting bid [[beta].bar] is chosen as the evaluator, and the price he receives for evaluating is ([[beta].bar] + [[beta].bar])/2. The other n - 1 agents wait for the evaluation, and each pays an equal portion of the price. The price paid by each individual for the information is ([[beta].bar] + [[beta].bar])/2(n - 1). Hence when entering a bid, agent i knows that [[beta].sub.i]/(n - 1) is the maximum amount she will have to pay for the evaluation. In the event that multiple agents submit the lowest bid, the evaluator is chosen randomly from that subset of bidders.

Discriminatory Sealed Bid Auction

Like the uniform price sealed bid auction, in this auction, each bidder privately submits a bid, and the winners are the bidders who submit the highest bids. However, in the independent private values environment, each winner pays a price equal to his own bid. In an evaluations market, a bid [[beta].sub.i], in this institution represents a minimum price that agent i will be paid to evaluate, and [[beta].sub.i]/(n - 1) represents a maximum amount that agent i might have to pay for the information, as in the uniform price auction. Again, the bids are ranked from lowest to highest and the person submitting the lowest bid is chosen to evaluate the product. However, the price paid by each agent who waits depends on his bid and the evaluator's bid. Let j denote the person who submitted the lowest bid 13. Agent i [not equal to] j pays ([[beta].bar] + [[beta].sub.i])/2(n - 1) to the evaluator. The total amount paid to agent j, the evaluator, is ([[beta].bar]/2) + [[summation].sub.i [not equal to]j][[[beta].sub.i]/2 (n - 1)]. Again, ties are broken randomly.

Descending (English) Clock Auction

An English clock auction with independent private values begins with a low price that the bidders are willing to pay. As long as there are more bidders who are willing to pay the current price than there are units for sale, the price increases. Once there is no excess demand at the current price, the auction ends and all remaining bidders buy the final price. An English clock auction for evaluations is also operationalized by setting an initial price, such that multiple agents are willing to accept the proposed price, and then moving the price in the less favorable direction until all but one agent drops out of the auction. Because an evaluation market is attempting to procure an evaluation, the process works in the reverse of that in an independent private value environment. The initial price on the clock is set sufficiently high such that multiple agents are willing to evaluate the fictitious product and then the price falls until all but one person has indicated a preference to wait for the information (i.e., withdraw from the market and not evaluate) at the current price on the clock. Once a bidder signals to wait, a bidder cannot re-enter the market for that period. The number of active bidders is not publicly stated as the price decrements and bidders withdraw. The clock price, like the bid amount in the sealed bid institutions, refers to the amount that the evaluator receives. Those who wait pay 1/(n - 1) of the final clock price. Again, ties are settled randomly. Unlike the sealed bid institutions, this mechanism requires up-front parameterization in the form of a starting price, the amount by which the clock decrements, and the rate of time for the decrement. Additionally, a stopping rule is imposed in the event that the clock price reaches zero with at least two agents still in the market. Because a clock price of zero indicates that at least two agents are willing to undertake the evaluation for no payment, one of these agents is randomly selected to provide the information for free to the remainder of the group. It is worth noting that although truthful revelation is a dominant strategy in the independent private value environment, this is not the case in an evaluations market because the price at which the penultimate bidder exits the market affects the price that bidder will pay for the information.

Ascending (Dutch) Clock Auction

As with an English clock auction, a Dutch clock auction also involves systematically changing the price until a winner is declared. With this auction, the price is initially set such that nobody is willing to accept the transaction. The price is then gradually improved until some agent accepts the terms. With independent private values, the price begins high and then is lowered until a bidder accepts the current price. However, in the market for evaluations the price is set sufficiently low such that everyone wants to wait initially, then the price is increased until the first agent agrees to undertake the evaluation for the price shown on the clock. Again, the clock price refers to the price received by the evaluator. Each of the n - 1 agents who did not indicate a willingness to evaluate pays 1/(n - 1) of the amount received by the evaluator. This institution also requires additional parameters for the starting price, minimum amount of the price increment, and rate at which the price is incremented.

Parameters

We chose to compare the institutions with n = 4 participants in each market. The experimental literature is ripe with examples of markets in which four sellers or four bidders can be considerably competitive, depending on other details of the environment (see, e.g., Cox, Roberson, and Smith 1982; Isaac and Walker 1985; Isaac and Reynolds 2002: Thomas and Wilson 2002; and Deck and Wilson 2003). Table 1 reports the good values, bad values, and opportunity costs for each of four participants in a market. We continue to assume that the product is good or bad with a 50% probability.

Within a group of four agents is one type 1, one type 2, and two type 3 agents. (7) Each agent has an expected value of -10 for evaluating, which is less than the common opportunity cost of 24. However, agents differ in their expected values from another agent evaluating. Table 2 lists the expected social surplus depending on which agent, if any, evaluates.

The socially efficient payoff is for one of the two type 3 agents to undertake the evaluation, and because there are two type 3 agents, a unique price exists that supports this outcome, assuming all agents are risk neutral. (8) Because the type 3 agents are identical, the equilibrium price is such that the two type 3 agents are indifferent between evaluating and waiting. That is, the price structure satisfies ([g.sub.3] + [b.sub.3])/2 q-price received for evaluating = ([g.sub.3] + [c.sub.3])/2 -price paid for information. When the price paid for information is 1/(n - 1) times the price received by the evaluator, the equilibrium price that supports the efficient outcome is 54 for the parameters in Table 1. This price prediction is applicable to the uniform price sealed bid, descending English clock auction, and ascending Dutch clock auction because each agent that waits pays the same price. However, this price prediction does not hold for the discriminatory sealed bid auction because the choices of the type 1 and type 2 agents will affect the price the evaluator receives. Furthermore, an explicit price prediction would depend on the beliefs bidders have about the likely bids and the values of others. (9) Nevertheless, as an exploratory exercise we include it in our comparison of the other three institutions.

We chose the payoffs for the type 1 and 2 agents so that they have the same expected value for evaluating as the type 3 agents, but if a type 1 or 2 agent evaluates, the expected social loss is nontrivial and dependent on which of the two evaluates. The opportunity cost of 24 satisfies the desirable property that [b.sub.i] + [g.sub.i] < 24: for each agent type. It also creates an essential separation between the social payoff in the case in which no one evaluates and the efficient case, while maintaining the property that a person who receives one good payoff, one bad payoff, and one opportunity cost payoff will experience nonnegative earnings. (10) Also, an opportunity cost of 24 leads to an integer price prediction that is not a natural focal point.

As discussed above, some of the institutional treatments also require parameterization. In the no-market baseline, the time available for product evaluations is T = 30 seconds. Both clock institutions require an increment and an initial price. The clock increment/decrement is 1 and updated every second. The initial prices are set such that, a priori, the mechanisms would last as long as the baseline no-market treatment. In the no-market baseline, no one should evaluate and a period lasts 30 seconds. Therefore, the starting price in the English clock auction is 54 + 30 = 84, and the starting price in the Dutch auction is 54 - 30 = 24 (which is conveniently equal to the opportunity cost). For the two sealed bid institutions, subjects also have 30 seconds to enter their bids. If a subject does not enter a bid in the allotted time, then his previous bid serves as the default. (11) For the payoffs to be comparable across treatments, we desired that each laboratory session consist of the same number of periods and thus each should last approximately the same amount of time.

3. Experimental Procedures

A market consisted of four subjects who were constantly and anonymously matched throughout the 48 decision periods of the experiment. Subjects retained the same agent type each period and never knew the payoff parameters associated with the other subjects or even the distribution of those parameters. Before the experiment began, each subject was given a set of written instructions. After all subjects completed the instructions and had the opportunity to ask questions, the computerized experiment began.

For the first 24 periods, all subjects, regardless of institution treatment, participated in the no-market baseline. This insures that before introducing the market mechanism the subjects have substantial experience with the payoff implications for evaluating and not evaluating with good and bad values. When making decisions in the baseline environment, subjects knew only their own payoff parameters and the time remaining in the period.

After each period, subjects received feedback about their own payoff and whether the product was good or bad that period if and only if someone evaluated. Subjects were not told who evaluated or anyone else's payoff. At any point during the experiment, regardless of treatment, a subject could scroll through a table that displayed for all previous periods their own action, whether or not someone else had evaluated: the payoff state if revealed: and their own profit. After the first 24 periods were completed, subjects in the baseline treatment continued in this environment for an additional 24 periods. Subjects in the other four treatments were given additional written instructions about a single market mechanism that would be in place for the remaining 24 periods. After all subjects read the instructions and had the opportunity to ask questions, the computerized experiment resumed. In addition to the information revealed in the baseline case, subjects in the four market treatments were told the price they paid for waiting or the price received for evaluating. This design constitutes a nontrivial challenge to the auction mechanisms. First, every subject is identical in expected value terms (-10) for not evaluating, and second, subjects have no information at all on the values of the other participants in the market. Subjects were not told the number of periods in the experiment or a portion thereof, nor were they informed in advance that a mechanism would be imposed in the latter part of the session.

A total of 25 sessions were conducted, five for each of the five treatments. We held constant across all sessions a random sequence of 48 good and bad value states; that is, one sequence was randomly determined in advance and then employed in all sessions. This serves to reduce the variation across sessions.

The 100 participants in this study were undergraduates from the general student population at George Mason University, where the experiments were conducted. For participating in the one-hour experiment, each subject was paid $7 for showing up on time, plus his or her salient earnings. All payoffs, parameters, and prices were denoted in terms of experimental dollars (EXP). At the conclusion of the experiment, a subject's cumulative profit was converted into U.S. dollars at the rate of EXP 200 = US$1, which was stated to the subjects before the beginning of the experiment. (12) The average earnings in the experiment were approximately $15.25, excluding the $7 show-up fee.

4. Experimental Results

The data consist of observed behavior in 720 periods under the no-market baseline and 120 periods under each of the four market mechanisms. We found that each of the market mechanisms was successful at increasing efficiency relative to the baseline by increasing the frequency with which the optimal agent type evaluated. This result is presented as a series of findings, each with supporting analysis that treats each session as an independent observation. To control for learning, the analysis focuses exclusively on data from the latter half of the periods in a particular institution (periods 13-24 and 37-48).

The first 24 periods in each session consist solely of the baseline situation. If observed differences in the latter part of the experiment are attributable to the institution and not subject heterogeneity, behavior should be similar across the groups before the auction mechanisms are introduced. The first finding is largely a calibration result demonstrating that subject behavior is indeed similar across all treatments before the implementation of a market mechanism. To compare the choices of individuals, and hence performance of the five institutions, we use the metric of average ex ante efficiency from the appropriate periods within a session. This is a measure of the expected social welfare conditioned on the frequency with which, if any, agents of a particular type evaluated. Thus, two sessions in which the same numbers of each type undertook the evaluation would be considered as performing identically, even though realized surplus might vary across the sessions depending on who evaluated when the product was good or bad. Because each session is independent of the others, this metric allows for a comparison of independent observational units.

FINDING 1. Ex ante efficiency is statistically indistinguishable across all five treatments in the first half of the experiment before an auction mechanism is introduced.

SUPPORT. For the null hypothesis that ex ante efficiency in the initial no-market phase of each session did not differ by treatment, we employ a Kruskal-Wallis test on the 25 average ex ante efficiency observations (one for each session) for periods 13-24. The test statistic, corrected for ties in the ranking, is 0.467, which cannot be rejected in favor of the two-sided null at any standard level of significance.

The frequency with which subjects choose to evaluate is nontrivial in periods 13-24. Figure 1 illustrates when the subjects choose to evaluate during the 30-second period without a market mechanism in place. The hatched bar indicates how many times no one evaluated, and the solid bars indicate how many evaluations occurred for the six, five-second blocks of time. Notice that when a subject evaluates, it is most often within the first 5 or last 5 seconds of the period. This suggests that our choice of a 30-second period is not too short because subjects choose either to evaluate quickly or to wait until the end of the period. Having established that the sessions do not differ before the implementation of a market mechanism, any differences across treatments can be attributed directly to the institutions. Therefore, our focus now turns to the effect of the market mechanisms as observed over the last 12 periods of each session.

[FIGURE 1 OMITTED]

FINDING 2. The introduction of a market mechanism significantly increases efficiency.

SUPPORT. Figure 2 displays the average ex ante efficiency over the last 12 periods by session. We use the Wilcoxon rank sum test (W statistic) to determine whether each mechanism improves efficiency relative to the no market baseline. In each pairwise comparison, the W statistic was 40, the largest value possible. Thus, the null hypothesis of no change in efficiency can be rejected at the 99% confidence level in favor of the alternative that the mechanism improved efficiency for each of the four market mechanisms considered.

[FIGURE 2 OMITTED]

The above analysis clearly shows that implementing a market for evaluations increases the expected efficiency. The next finding explores the differences between the four market mechanisms in terms of efficiency. A separate finding then addresses the source of the institutions' success.

FINDING 3. The four market mechanisms are statistically indistinguishable with respect to efficiency.

SUPPORT. Given no a priori ordering of ex ante efficiency by institution treatment, we employ a Kruskal-Wallis test to test the null of no difference by mechanism against the alternative that efficiency differs for some institution. Adjusting for ties in the efficiency rankings of individual sessions, the test statistic is 0.352, which means that the null hypothesis cannot be rejected at standard levels of significance.

Market mechanisms necessarily assign one person to evaluate the fictitious product in periods 25-48. For example, in either sealed bid auction someone must submit the lowest bid. Given the payment schemes, participation is individually rational for each agent. This is a distinct advantage of a market because it avoids the worst-case scenario in terms of ex ante efficiency, namely no one evaluating, which can and does happen without a market. It is reasonable to ask whether the greater efficiency is due simply to this aspect of the design. To answer this question, we determine what efficiency would have been in the no-market baseline if one person had been randomly selected as the evaluator when no one volunteered. With this more conservative accounting for the no-evaluation outcomes, efficiency with the tour market mechanisms is still statistically greater than efficiency in the no-market baseline, although the increase is only 5.3 percentage points. Formally, let [M.sub.j] denote the frequency over periods 37-48 at which no one evaluated in session j in the no-market baseline groups, and let [m.sub.ij] denote the frequency at which an agent of type i was observed to evaluate voluntarily over the same period in session j. For this conservative test, we recalculate the ex ante efficiency for the no-market baseline treatment by allocating one agent, in proportion to the types listed in Table 1, to be the evaluator for each of the periods in which no one voluntarily evaluated. More specifically, the frequency of type i agents evaluating in a no-market baseline session is imputed to be [m.sub.ij] + [[theta].sub.i][M.sub.j], where [[theta].sub.i] = 0.5 if i = 3 and [[theta].sub.i] = 0.25 for i = 1 and 2. Let [E.sub.i] denote the ex ante efficiency when a type i agent evaluates (see Table 2). The recalculated ex ante efficiency for baseline session j is thus [[summation].sub.i] ([m.sub.ij] + [[theta].sub.i][M.sub.j])[E.sub.i].

We use a Wilcoxon rank sum test to compare the recalculated efficiency of the baseline to the observed aggregated efficiency of the market treatments. The null hypothesis of no effect from a market can be rejected at the 95% confidence level in favor of the alternative hypothesis that the market mechanisms increase ex ante efficiency (W = 38, p = 0.0331). This demonstrates that an auction market increases efficiency by more than would be expected from merely randomly and involuntarily assigning one person to evaluate. (13) It should be emphasized that in the market mechanism, the subjects are volunteering to evaluate because they each have the choice in their bid to indicate their willingness to evaluate or wait.

Although someone will be chosen to evaluate with the market institutions, these auctions do not always induce the optimal agent to evaluate in each period as show in Figure 2. Also, it is not the case that type 1 and 2 agents are never willing to undertake the evaluation for free. In fact, a suboptimal agent evaluated 30% of the time on average for the last 12 periods in the no-market baseline treatment. Thus it remains to be explained the extent to which the markets succeed. Our conjecture is that the performance increase is due to the mechanisms encouraging type 3 agents to evaluate when no one else will. Finding 4 discusses this explanation formally.

FINDING 4. The auction mechanism induces the optimal agent to undertake the evaluation when no one else is willing to do so, thereby increasing efficiency.

SUPPORT. Figure 3 illustrates the qualitative support for this finding. The hatched (solid) bars for each treatment indicate the frequency with which each agent type evaluated in periods 13-24 (37-48). The "None" category specifies how many times no agent voluntarily evaluated and contains the same data displayed in Figure 1 with the hatched bars. In the no-market baseline treatment, all types of agents reduce their evaluations, most noticeably the type 1's (the least efficient evaluators). In marked contrast, the number of evaluations by type 3 agents increases substantially for all of the market institutions and generally by the number of times no one is willing to evaluate in the no-market periods 13-24. This finding is supported quantitatively by a comparison of observed market behavior with behavior in the no-market baseline in which type 3 agents are assumed to have evaluated if no one else volunteered. For brevity, this requires an imputation similar to that discussed following Finding 3. Specifically, the frequency of type i evaluating in a no-market baseline session j was recalculated as [m.sub.ij] + [[phi].sub.i][M.sub.j], where [[phi].sub.i] = 1 if i = 3 and [[phi].sub.i] = 0 otherwise. Because the conjecture involves the likelihood that a type 3 agent evaluated, the employed metric is the frequency with which a type 3 agent evaluated over the last 12 periods and not ex- ante efficiency. We do not suppose that the four institutions are identical under this metric. Thus, a Kruskal-Wallis test is used to test the null hypothesis that the frequency of type 3 evaluations is the same across all five treatments against the alternative hypothesis that this frequency differs for some treatment. The tie-corrected rank test statistic is [[chi square].sub.3] = 0.352, which cannot be rejected (p = 0.9499). Hence, the mechanisms encourage the appropriate agents to undertake the evaluations. (14) *

[FIGURE 3 OMITTED]

With strictly private information on values and heterogeneous attitudes toward risk, the market mechanism induces the socially optimal person to evaluate the product. This is nontrivial in that no explicit cues from the environment are driving this behavior. Rather, it is the incentives created by the market mechanism that lead the socially optimal agent to provide an evaluation when without a market such an agent would not. Note that the market mechanism could also have induced other agents to evaluate: agents who have the same risk-neutral expected value for evaluating and not evaluating the product. But this is not we what find.

Many experimental studies have shown that efficiency is quite high, generally greater than 95%, for uniform and discriminatory sealed bid and Dutch and English clock auctions with independent private values. Prices, however, often differ substantially by auction format, even though all four institutions are theoretically revenue equivalent with risk-neutral bidders (see footnote 3 for references). Our next finding compares the prices received by the evaluator for each institution. We employ a linear mixed effects model for analyzing the data with repeated measures as the basis for quantitative support. (15) Sessions are indexed by j = 1, ..., 20 and periods by t = 37, ..., 48. This parametric estimation treats each session as a random effect [[epsilon].sub.j] and each institution as a fixed effect [[beta].sub.j]. The dependent variable [Price.sub.jkt] is the price received by evaluator k in period t of session j. Within the session random effect, we also include a random effect [e.sub.k] for subject k within the session that submits the winning bid to be the evaluator. Specifically, we estimate the model

[Price.sub.jkt] = [[beta].sub.0] + [[beta].sub.1] [Descending.sub.j] + [[beta].sub.2] [Uniform.sub.j] + [[beta].sub.3] [Discriminatory.sub.j] + [[epsilon].sub.j] + [e.sub.k] + [u.sub.jkt], where [[epsilon].sub.i] ~ N(0, [[sigma].sup.2.sub.[epsilon],j]), [e.sub.k] ~ N(0, [[sigma].sup.2.sub.e]), and [u.sub.jkt] ~ N(0, [[sigma].sup.2.sub.u]). (16)

FINDING 5. The null hypothesis of identical market prices across all four market mechanisms cannot be rejected. Additionally, the theoretical risk-neutral price prediction is included in the confidence interval for any standard level of significance.

SUPPORT. Table 3 reports the fixed effects parameter estimates with the ascending Dutch clock auction serving as the basis for comparison. The considerable lack of significance on [[beta].sub.1] [[beta].sub.2], and [[beta].sub.3] fixed effects indicates that, on average, the Descending English clock and Uniform price and Discriminatory price sealed bid auctions result in the same price, respectively, as the Ascending Dutch clock auction. The second part of the finding is supported by a t-test of the null hypothesis that [[beta].sub.0] = 54 against the two-sided alternative. The test statistic is 0.59 with 178 degrees of freedom, which cannot be rejected at standard levels of significance.

Although the average price is statistically indistinguishable under each treatment, it is important to notice that there is considerable variation in observed prices. Figure 4 depicts this price variability. Over the last 12 periods, one session in each treatment has a median price less than 35. Also, over the last 12 periods, one session in each treatment has a median price over 75. Overall, the lowest median price in a session is 3.25, and the highest is 109.17. Even after controlling for variation from the random effects of sessions and the evaluators within each session, considerable variation in the observed prices remains, as evidenced by the size of the standard error on [[beta].sub.0] reported in Table 3. This suggests that while the theoretical price may characterize the central tendency of behavior in these institutions, it is not the case that people are behaving in strict accordance with the prediction. This can also be gleamed from Figure 2. If each subject behaves as theoretically predicted then in sessions with an operating market institution the ex ante efficiency should be 100%, which it clearly is not.

[FIGURE 4 OMITTED]

The same heterogeneity among subjects in terms of willingness to evaluate that explains the relatively low efficiency could also contribute to the price variability. Of the 100 subjects in the experiment, 12 clearly enjoyed providing the evaluation, taking on the risk at least 50% of the time in periods 13-24, the last 12 periods before a market mechanism was introduced. (17) Table 4 compares the frequency with which these risk takers evaluated when a market was in operation with the frequency of evaluation by others of the same type in the same institution. (18) Except in two cases, subjects who frequently volunteered to provide an evaluation for free were also more likely to be paid for providing an evaluation than subjects in similar situations. But we also note that of the six type 1 and 2 (inefficient) agents that evaluate more than 50% of the time in the premarket periods, only one evaluates more than 50% of the time after a market mechanism is introduced.

5. Conclusion

As demonstrated by Avery, Resnick, and Zeckhauser (1999), a pricing mechanism for product evaluations can increase efficiency by voluntarily eliciting an evaluation that would otherwise not be provided. With a controlled laboratory experiment, we evaluate the performance of four market mechanisms for providing product evaluations: uniform price sealed bid, discriminatory price sealed bid, descending (English) clock auction, and ascending (Dutch) clock auction. Our results indicate for this nonrivalrous information good that (i) each of these institutions improves social welfare and (ii) the four mechanisms are behaviorally equivalent with respect to the prices received by the evaluator. Each of the four institutions improves efficiency by inducing the socially optimal agent to undertake the evaluation when no one else is willing to do so. In our design, this is not an easy feat. All of the subjects had strictly private information on their values and the same (negative) expected value for evaluating the good. Moreover, the subjects might have heterogeneous attitudes toward risk, and yet, a market mechanism leads the socially optimal agent to voluntarily provide an evaluation when without a market such an agent would not. However, it is not the case that the four market mechanisms achieve 100% efficiency. There is heterogeneity in our subjects, leading some socially suboptimal agents to evaluate. This is consistent with naturally occurring voluntary evaluations, such as book evaluations provided (and not provided) by Amazon.com.

These efficiency and price results contrast with the standard experimental results of these auction formats for independent values. In laboratory auctions with independent private values, prices clearly separate with a relatively small variance across sessions, and efficiency is consistently greater than 95%. In those experiments, buyers' values and surplus are induced with certainty each period. In the case of nonrivalrous good such as evaluation, there is risk in the product value, which might be a source of the price variability that we find here. The large heterogeneity in willingness to evaluate often led agents that were suboptimal from an expected payoff perspective to evaluate. Therefore, the possible efficiency improvement that the market mechanisms could generate was limited. This might account for the similar efficiency performance across the four institutions. Collective or individual heterogeneity of subjects, uncertainty about the outcome, and the nonrivalrous nature of evaluations could explain why the market outcomes here differ from standard auction results in the laboratory. Of course, only a systematic analysis of these differences can reveal the explanation, but that we leave to future research.

Appendix

Experiment Instructions

This is an experiment in the economics of decision making. Various research foundations have provided funds for this research. The instructions are simple, and if you understand them, you may earn a considerable amount of money that will be paid to you in CASH at the end of the experiment. Your earnings will be determined partly by your decisions and partly by the decisions of others. If you have questions at airy time while reading the instructions, please raise your hand and a lab monitor will assist you.

This is what your screen will look like in the experiment. In each period of the experiment you will be matched with three other people, your counterparts. All four of you have a decision to make: either evaluate a fictitious product or wait for a counterpart to do so. Your payoff will be determined in part by the decisions you and your counterparts make.

You and your counterparts for the period will have 30 seconds to decide if you want to evaluate. At any point during the 30 seconds, you or your counterparts can click on the Evaluate button. Only one of you can evaluate during a period. If the clock expires without any of you choosing to evaluate, then all of you have waited for that period.

How is your payoff determined? The fictitious product will either be Good or Bad. When you and your counterparts are making your decision to either evaluate or not, you will not know if the product is good or bad. You and your counterparts will know whether it is good or bad only after one of you has chosen to evaluate. There is a 50% chance that the fictitious product will be Good and a 50% chance that it will be Bad. Your payoff depends on whether the product is Good or Bad, if you or a counterpart has chosen to evaluate. If any of you evaluate the good, then your payoff is higher if the product is Good than if it is Bad.

Now we will go through an example of how to read the payoffs which are listed in the table. The payoffs in these instructions are for illustrative purposes only. The payoffs in the experiment will be different from those displayed here.

Suppose that you chose to evaluate. Then your payoff would be determined by the first row of the table. The payoffs depend on whether the product is Good or Bad. If the product is Good, your payoff would be 140. If the product is Bad, you payoff would be -180.

Suppose that while you are waiting, your counterpart chooses to evaluate. In this case the second row displays your payoff. If the product is Good, your payoff would be 140. If the product is Bad, your payoff would be 30.

Finally, is none of you decide to evaluate, then your payoff is shown on the bottom row. Notice that your payoff will be 30, regardless of the product being Good or Bad. None of you will know the state of the product that period because none of you chose to evaluate it.

At the end of the period you will have 5 seconds to review the results. At the end of that time, the clock will reappear and the next period will begin. Your counterparts' payoffs may or may not be the same as yours.

At the end of the experiment, your experimental dollars will be converted into cash at the rate of 200 experimental dollars for US$1.

Any questions? If not, please raise your hand to indicate that you have finished reading the instructions.

Uniform Price Sealed Bid Auction Instructions

For the next portion of the experiment, the way you and your counterparts' payoffs are determined based on the fictitious good will remain the same, but all of you will be bidding to pay to Wait and to get paid to Evaluate. In each period of the experiment you will continue to be matched with the same three counterparts.

This is what your screen will look like in the next portion of the experiment. Each period there will be three people who will wait and one person who will evaluate. Whether you wait or evaluate depends upon the bids that you and the other three people submit.

At the beginning of each period you will submit a bid for the "Price Received to Evaluate." This is the amount you are willing to be paid to evaluate. One third of this bid also serves as the price you are willing to pay to wait.

Once all of the bids have been submitted, the computer ranks the bids from lowest to highest. The person who is chosen to evaluate is the person who submits the lowest bid. (Any ties will be broken randomly.) This person will receive as payment the average of her bid and the second lowest bid. Hence, the person who evaluates will always be paid at least as much as the bid that she submitted. This amount will be recorded in the "Price" column and will be added to the "'My Payoff" column for that period.

The three people who submitted the three highest bids will each pay 1/3 of the amount received by the evaluator. This amount will be recorded in the "Price" column and will be subtracted from the "My Payoff" column for that period. Notice that these throe people will not pay more than the amount they each submitted as the "Price Paid to Wait."

Just as before, the evaluator will also receive a payoff depending upon whether the product is Good or Bad.

At the end of the period you will have 5 seconds to review the results. At the end of that time, the next period will begin. If you do not submit a new bid before the time on the clock expires, the computer will use last period's bid as the bid for the current period.

At the end of the experiment, your experimental dollars will also be converted into cash at the rate of 200 experimental dollars for US$1.

Any questions? If not, please raise your hand to indicate that you have finished reading the instructions.

Discriminatory Price Sealed Bid Auction Instructions

For the next portion of the experiment, the way you and your counterparts' payoffs are determined based on the fictitious good will remain the same, but all of you will be bidding to pay to Wait and to get paid to Evaluate. In each period of the experiment you will continue to be matched with the same three counterparts.

This is what your screen will look like in the next portion of the experiment. Each period there will be three people who will wait and one person who will evaluate. Whether you wait or evaluate depends upon the bids that you and the other three people submit.

At the beginning of each period you will submit a bid for the "Price Received to Evaluate." This is the amount you are willing to he paid to evaluate. One third of this bid also serves as the price you are willing to pay to wait.

Once all of the bids have been submitted, the computer ranks the bids from lowest to highest. The person who is chosen to evaluate is the person who submits the lowest bid. (Any ties will be broken randomly.)

The three people who submitted the three highest bids will each pay the evaluator. The amount that each waiter pays has two factors. The first factor is the average of their own bid and the lowest bid. This amount is then multiplied by 1/3 (or equivalently divided by 3) because each of the three waiters pays the evaluator. This amount will be recorded in the "Price" column and will be subtracted from the "My Payoff" column for that period. Notice that these three people will not pay more than the amount they each submitted as the "Price Paid to Wait."

The evaluator will receive as payment each of the amounts paid by the three waiters. Notice that the evaluator will always be paid at least as much as the bid that she submitted. This amount will be recorded in the "Price" column and will be added to the "My Payoff" column for that period.

Just as before, the evaluator will also receive a payoff depending upon whether the product is Good or Bad.

At the end of the period you will have 5 seconds to review the results. At the end of that time, the next period will begin. If you do not submit a new bid before the time on the clock expires, the computer will use last period's bid as the bid for the current period.

At the end of the experiment, your experimental dollars will also be converted into cash at the rate of 200 experimental dollars for US$1.

Any questions? If not, please raise your hand to indicate that you have finished reading the instructions.

Ascending Clock Auction Instructions

For the next portion of the experiment, the way you and your counterparts' payoffs are determined based on the fictitious good will remain the same, but all of you will be bidding to pay to Wait and to get paid to Evaluate. In each period of the experiment you will continue to be matched with the same three counterparts.

This is what your screen will look like in the next portion of the experiment. At the beginning of each period the "Price Received to Evaluate" starts at a price of 24 and then continues to increase by one experimental dollar each second. The "Price Received to Evaluate" will increase until the first person clicks on the Evaluate button. The first person who clicks on the Evaluate button will evaluate the product and receive as payment from the three counterparts the amount in the "Price Received to Evaluate" box. This amount will be recorded in the "Price" column and will be added to the "My Payoff" column for that period.

Just as before, the evaluator will also receive a payoff depending upon whether the product is Good or Bad.

The three other people who did not click on the Evaluate button will wait that particular period. The three waiters will each pay the amount next to the label "Price Paid to Wait." This amount will be subtracted from the "My Payoff" column. Notice that because there are three waiters, the amount paid by each waiter is 1/3 of the price received by the evaluator.

At the end of the period you will have 5 seconds to review the results. At the end of that time, the prices will again start at 24 and will increase until one person clicks on the Evaluate button.

At the end of the experiment, your experimental dollars will also be converted into cash at the rate of 200 experimental dollars for US$1.

Any questions? If not, please raise your hand to indicate that you have finished reading the instructions.

Descending Clock Auction Instructions

For the next portion of the experiment, the way you and your counterparts' payoffs are determined based on the fictitious good will remain the same, but all of you will be bidding to pay to Wait and to get paid to Evaluate. In each period of the experiment you will continue to be matched with the same three counterparts.

This is what your screen will look like in the next portion of the experiment. At the beginning of each period the "Price Received to Evaluate" starts at a price of 84 and then continues to decrease by one experimental dollar each second. The "Price Received to Evaluate" will decrease until the first three people click on the Wait button. The remaining person who has not clicked on the Wait button will evaluate the product and receive as payment from the three counterparts the amount in "Price Received to Evaluate" box. This amount will be recorded in the "Price" column and will be added to the "My Payoff" column for that period.

Just as before, the evaluator will also receive a payoff depending upon whether the product is Good or Bad.

The three other people who clicked on the Wait button will wait that particular period. The three waiters will each pay the amount next to the label "Price Paid to Wait." This amount will be subtracted from the "My Payoff" column. Notice that because there are three waiters, the amount paid by each waiter is 1/3 of the price received by the evaluator.

If the "Price Received to Evaluate" falls to zero and there are still at least two people who have not clicked on the Wait button, then one of the people who has not clicked the button will be randomly selected to be the evaluator and will receive zero (because by not clicking the Wait button the person has indicated that they are willing to receiving nothing for evaluating).

At the end of the period you will have 5 seconds to review the results. At the end of that time, the prices will again start at 84 and will decrease until the first three people click on the Wait button.

At the end of the experiment, your experimental dollars will also be converted into cash at the rate of 200 experimental dollars for US$1.

Any questions? If not, please raise your hand to indicate that you have finished reading the instructions.
Table 1. Market Values

 Good Bad
 Agent Value Value Opportunity
Agent i Type (g) (b) Cost (c)

 1 1 320 -340 24
 2 2 220 -240 24
 3, 4 3 100 -120 24

 Expected Value Expected Value if
 if Agent Another Agent
Agent i Evaluates Evaluates

 1 -10 172
 2 -10 122
 3, 4 -10 62

Table 2. Expected Social Surplus by Type of Evaluator

Type of Expected Efficiency
Agent Social (= Social Payoff - No Evaluation Payoff)/
Evaluating Payoff (Maximum Social Payoff - No Evaluation Payoff)

None 96 0%
Type 1 236 56%
Type 2 286 76%
Type 3 346 100%

Table 3. Estimation Results for Linear Mixed Effect Model of Price
([Price.sub.jkt] = [[beta].sub.0]+ [[beta].sub.1] [Descending.sub.j]
+ [[beta].sub.2] [Uniform.sub.j] + [[beta].sub.3]
[Discriminatory.sub.j] + [u.sub.j] + [e.sub.k] + [[epsilon].sub.jkt])

 Standard Degrees of
Parameter Estimate Error Freedom t p

[[beta].sub.0] 47.06 11.763 178 4.00 <0.0001
[[beta].sub.1] 11.84 16.647 16 0.711 0.4871
[[beta].sub.2] -5.44 16.615 16 -0.326 0.7475
[[beta].sub.3] 4.39 16.661 16 0.263 0.7957

Table 4. Market Behavior of Subjects Who Evaluated
[greater than or equal to] 50% of the Time in Periods 13-24

 % of Periods Evaluated with Auction
 (Periods 37-48)

 No.
Agent of Auction By These By All Others of
Type Individuals Treatment Agents the Same Type

 2 1 Uniform 58 22
 2 1 Uniform 8 22
 3 1 Uniform 33 30
 2 1 Discriminatory 25 3
 3 2 Discriminatory 13 35
 2 1 Ascending clock 33 23
 3 2 Ascending clock 88 18
 1 2 Descending clock 42 11


This paper has benefited from conversations with Kevin McCabe and David Meyer and from comments by the Editor, two anonymous referees, Glenn Harrison, and participants at the 2003 Summer Economic Science Association meetings. We thank Stephen Salena for research assistance in running these experiments. B.J.W. also thanks the Office of the Provost at George Mason University for summer research funding and for additional support from the International Foundation for Research in Experimental Economics. The data are available on request.

Received February 2004; accepted November 2004.

(1) Additionally, businesses have long attempted to use techniques such as brand image and advertising as an indication of quality. However, when new products or services are introduced within a brand line, buyers can still face uncertainty about their value for the product or service.

(2) Both amateur and professional reviews can suffer from a lack of technical knowledge on the part of the reviewer and can reduce social welfare if, given the heterogeneity of values, the actual reviewers are not the socially optimal ones.

(3) As Avery, Resnick, and Zeckhauser (1999) point out in their footnote 3. positive evaluations for some products (e.g., stocks or restaurants) can increase demand and hence could be rivalrous. In such situations, agents could have an incentive not to truthfully reveal.

(4) Avery, Resnick, and Zeckhauser (1999) show that markets can solve much more complex problems as well, such as the case in which someone else's positive experience only serves as a signal about the probability that one's own experience will be positive. However, as a first step in understanding behavior in markets for evaluations, this study focuses exclusively on the case in which everyone's opinion of the outcome, but not necessarily their payoff from it, is the same.

(5) For a more comprehensive discussion, the reader is directed to Kagel and Roth (1995). Two early studies test the strategic equivalence of first-price and Dutch auctions. Coppinger, Smith, and Titus (1980) and Cox, Roberson, and Smith (1982) both find that (i) prices are higher in first-price auctions than in Dutch auctions and (ii) bidding is consistent with risk-averse behavior. The predicted isomorphism between English and second-price auctions also fails to be observed. Coppinger, Smith, and Titus (1980) and Kagel, Harstad, and Levin (1987) find that bidding in the English outcry auction conforms to the theoretical predictions quite well, whereas in the one-shot second-price auction, bidders consistently bid higher than the dominant strategy prediction, even with experience in the auction mechanism (Kagel and Levin 1993).

(6) Because one bidder will necessarily be selected as the evaluator and receive payment from the other participants, these mechanisms can be classified as fair division games (see Guth and van Damme 1986: Guth el al. 2002).

(7) The type 3 agents are similar to those used by Avery and Zeckhauser (1997) in a discussion of a market for evaluations in a similar setting.

(8) Identical expected values for evaluating generate a nontrivial environment to test how well a market institution induces a socially optimal evaluation by a type 3 agent.

(9) As a simplifying assumption, theoretical and experimental work on private value auctions have assumed that values are distributed uniformly and that this is common knowledge among the market participants. In contrast, we chose a challenging environment that parallels the naturally occurring economy, in which participants only have private information on their own values, a la a typical double auction market experiment. It is well known that the double auction institution quickly achieves competitive outcomes (at or near 100% efficiency), all with strictly private information on all values and costs. In this paper, we also want to investigate a rather difficult test of a market mechanism when the participants only have strictly private information, as is reasonably the case in the diffuse and impersonal world oil the Internet. The reader is referred to Guth and van Damme (1986) and Guth et al. (2002) for the derivation of the optimal bid strategies in fair value games for agents when there is common knowledge on the distribution of values.

(10) This positive gain property helps prevent a loss of control over a subject's motivation because negative earnings cannot be enforced. This is particularly important in the early stages of the experiment when subjects are relatively inexperienced.

(11) Participants in an unpaid pilot experiment indicated that more than 30 seconds was too long and rarely was the 30-second time limit binding. Also, the starting price of 84 in the English clock auction should discourage type l and type 2 agents from evaluating because risk-neutral agents of both types would prefer to exit the auction immediately. The results of the experiments are interpreted accordingly.

(12) The use of an exchange rate allows the clock prices to adjust over finer increments. Previous research has shown that clock speed and increment can be significant factors that influence market prices in standard clock auctions, which is also why we calibrated the starting prices for the clock auctions at 54 [+ or -] 30 experimental dollars.

(13) An alternative metric for evaluating market performance is to select an evaluator for the no-evaluation periods in the same proportion as when there was a volunteer, or, [[theta].sub.i] = [m.sub.ij]. A Wilcoxon rank sum test with this metric leads to the same conclusion (W = 37, p = 0.0284).

(14) The approach of conducting ex ante efficiency analysis similar to that discussed following Finding 3 would assume that the efficient type is assigned to evaluate (i.e., [[theta].sub.i] = 1 if i = 3 and [[theta].sub.i] = 0 otherwise. In this case, the null hypothesis that the market mechanisms generate the same level of efficiency as the baseline cannot be rejected in favor of the two-sided alternative with the Wilcoxon rank sum test. This also suggests that the markets induce the optimal agent to evaluate when no one volunteers (W = 71, p = 0.6832).

(15) See Longford (1993) for a description of this technique that is commonly employed in experimental sciences. A linear mixed effects model is not appropriate for the efficiency analysis because ex ante efficiency is discrete, taking on only one of three values each period.

(16) The linear mixed effects model for repeated measures treats each session as I degree of freedom with respect to the treatments. Hence, with four parameters, the degrees of freedom for the estimate of the institution treatment fixed effects are 16 = 20 sessions--4 parameters. This estimation accommodates sessionwise heteroscedastic errors when estimating the model via maximum likelihood. Adjusting the model to include first-order autoregressive errors in [u.sub.jkt], does not significantly increase the efficiency of the estimates.

(17) No subject volunteered to evaluate more than 83% of the time during these periods, and in many periods, no evaluations were provided.

(18) Table 4 only includes 11 of the 12 subjects that were in the auction mechanism treatment. The 12th was in the baseline treatment.

References

Avery, Christopher, Paul Resnick, and Richard Zeckhauser. 1999. The Market for evaluations. American Economic Review 89:564-84.

Avery, Christopher, and Richard Zeckhauser. 1997. Recommender systems for evaluating computer messages. Communications of the ACM 40:88-9.

Coppinger, Vicki M., Vernon L. Smith, and Jon A. Titus. 1980. Incentives and behavior in English, Dutch and sealed-bid auctions. Economic Inquiry 18:1-22.

Cox, James C., Bruce Roberson, and Vernon L. Smith. 1982. Theory and behavior of single object auctions. In Research in experimental economics. Volume 2, edited by Vernon L. Smith. Greenwich, CT: JAI Press, pp. 1-43.

Deck, Cary, and Bart J. Wilson. 2003. Automating posted pricing markets in electronic posted offer markets. Economic Inquiry 41:208-23.

Guth, Werner, Radosveta Ivanova-Stenzel, Manfred Konigstein, and Martin Strobel. 2002. Bid functions in auctions and fair division games: Experimental evidence. German Economic Review 3:461-84.

Guth, Werner, and Eric van Damme. 1986. A comparison of pricing rules for auction and fair division games. Social Choice and Welfare 3:177-98.

Isaac, R. Mark, and Stanley S. Reynolds. 2002. Two or four firms: Does it matter? In Research in experimental economics, volume 9: Market power in the laboratory, edited by Charles Holt and R. Mark Isaac. Greenwich, CT: JAI Press, pp. 95-119.

Isaac, R. Mark, and James Walker. 1985. Information and conspiracy in sealed bid auctions. Journal of Economic Behavior and Organization 6:139-59.

Kagel, John, Ronald Harstad, and Dan Levin. 1987. Information impact and allocation rules in auctions with affiliated private values: A laboratory study. Econometrica 55:1275-1304.

Kagel, John, and Dan Levin. 1993. Independent private value auctions: Bidder behavior in first-, second-, and third-price auctions with varying numbers of bidders. Economic Journal 103:868-79.

Kagel, John, and Alvin Roth. 1995. The handbook of experimental economics. Princeton, NJ: Princeton University Press.

Longford, N. T. 1993. Random coefficient models. New York: Oxford University Press.

Smith, Vernon L. 1994. Economics in the laboratory. Journal of Economic Perspectives 8:113-31.

Thomas, Charles J., and Bait J. Wilson. 2002. A comparison of auctions and multilateral negotiations. RAND Journal of Economics 33:140-55.

Cary A. Deck * and Bart J. Wilson ([dagger])

* Department of Economics, Walton College of Business, The University of Arkansas, Fayetteville, AR 72701, USA.

([dagger]) Interdisciplinary Center for Economic Science, George Mason University, 4400 University Drive, MSN 1B2, Fairfax, VA 22030-4444, USA: E-mail bwilson3@gmu.edu: corresponding author.
联系我们|关于我们|网站声明
国家哲学社会科学文献中心版权所有