首页    期刊浏览 2024年11月25日 星期一
登录注册

文章基本信息

  • 标题:Evaluating the assessment outcomes in the principles of marketing course.
  • 作者:Gerlich, R. Nicholas ; Sollosy, Marc
  • 期刊名称:Academy of Educational Leadership Journal
  • 印刷版ISSN:1095-6328
  • 出版年度:2009
  • 期号:January
  • 语种:English
  • 出版社:The DreamCatchers Group, LLC
  • 关键词:Educational assessment;Educational evaluation;Educational programs;Marketing

Evaluating the assessment outcomes in the principles of marketing course.


Gerlich, R. Nicholas ; Sollosy, Marc


INTRODUCTION

Increasingly, academic programs look to assessment as an important vehicle in determining the overall veracity of a program. It is a means by which the deliverer (instructor) of an academic program can determine if the goals and objectives of the program have been achieved by the recipients (students). Institutional effectiveness is concerned with the extent to which intended outcomes are being achieved (Black & Duhon, 2003). There are two (2) fundamental schemas of assessment delivery. The first employs an instrument developed externally and is standardized across a multitude of dimensions, like the Educational Testing Service's (ETS) exam in business. The second approach employs an embedded, internally developed, instrument that explicitly measures the specific outcomes associated with a program or course.

The purpose of this paper is to examine the various factors or drivers that influence a student's performance on an embedded, course specific assessment (local). These factors or drivers are described along two (2) dimensions, environmental or external, and internal student drivers. The assessment was administered to multiple sections of students in the same foundation or principles level course. The delivery of the actual course material was done both in the traditional in class model, as well as the increasingly popular Online delivery method. The results of this study are derived in the College of Business at a public university located in the Southwestern part of the United States. The institution is mid-sized with a total enrollment of approximately 7,500 total students, 1,000 undergraduate business students, 350 graduate business students.

This manuscript is organized as follows: First, a literature review is provided. The second section of the manuscript describes the data and the model. The next section provides empirical results testing the performance of students on the assessment exam while controlling for format of delivery, professor, student grade point average (GPA), international vs. domestic status, standardized test scores (ACT), transfer students, and gender. The final section offers conclusions and implications.

LITERATURE REVIEW

Research abounds on the subject of program assessment. Assessment is a "systematic collection, review, and use of information about educational programs undertaken for the purpose of improving student learning and development" (Palomba & Banta, 1999). Collegiate business programs are increasingly tasked with the need for ongoing assessment of student performance in their programs (Adams, et al, 2000; Bagamery, et al, 2005; Martell & Calderon, 2005; Terry, et al, 2008; Trapnell, 2005). Increasingly, since the mid-1980s, there has been a shift towards student-centered and learning-oriented assessments and accreditation (Lubinescu, et al, 2001). In fact, the AACSB imposes standards for program learning goals upon collegiate business programs aspiring to attain or maintain AACSB accreditation. These programs utilize direct measures in order to demonstrate student achievement of the stipulated goals (Martell, 2007; Pringle & Michel, 2007).

As assessment increases to build momentum, it is important to identify the internal and external audiences who will utilize the results in shaping and refining the assessment process. A comprehensive assessment process provides an institution with information that can be both shared and utilized to satisfy the needs of internal and external constituents. The internal audience (faculty, students, assessment committees, administrators and alumni) benefits by helping to define successful ongoing programs, implementing similar programs, and for improving less successful programs. Externally, the assessment data are used to demonstrate to accreditation organizations, government officials, government boards and other constituents the institution's effectiveness and accountability (Aloi, et al, 2003).

There are two principle types of assessment tests, standardized and local. Of the two, local tests require more faculty effort and other resources for test development, scoring, reporting, and improving. However, the advantage to the local instrument is that it can be tailored to a specific course or program so that the actual scores more accurately reflect the extent to which specific learning objectives are being met (Black & Duhon, 2003), along with impact of local and specific influences or drivers.

An important aspect for any instrument involves validity. Validity exists when the scores on the instrument accurately reflect achievement along the various dimensions the institution is seeking to evaluate. Validity has several dimensions, including content and context. Content validity exists when the test or instrument covers and measures the specific program or course curriculum. Context validity examines the extent to which scores (outcomes) logically correlate with other, external, variables expected to be associated with student achievement. An additional facet to validity is criterion validity, the extent to which scores of the test correlate with other variables one would expect to be associated with test performance ( Black & Duhon, 2003).

The Educational Testing Service's (ETS) exam in business has become, to many, the de facto standard of standardized assessment instruments in collegiate business programs. The literature reveals an almost universal agreement as to the principle variables examined as predictors of student performance on the ETS exam. These variables include: grade point average (GPA), standardized test scores (ACT/SAT) and gender. In addition to these variables, Mirchandani, et al, (2001) include transfer GPA and student grades in quantitative courses.

It is possible to extrapolate and utilize the same variables when examining the results on a local instrument. Terry, et al, (2008) developed a model based upon a production view of student learning to examine the determinants of performance on the business major field achievement ETS exam. Their model controlled for grade point average (GPA), standardized test scores (SAT/ACT), junior college transfer students, gender. Their findings were consistent with much of the previous research in this area, that academic ability as measured by grade point average (GPA) and scores on standardized tests (SAT/ACT) are the primary determinants of student performance on the ETS exam.

Black and Duhon's (2003) study of ETS scores conducted during three (3) semesters in 1996-1997 included an examination of an incentive as a driver in student performance. In that study, students scoring at the national 50th percentile, or better, were given an extra-point bonus, which was used in the calculation of the student's final course grade. The exclusion of this, or some other, form of incentive is not used; some students may not take the test seriously and by extension, the results may be misleading (Allen & Bycio, 1997). Terry (2007) included the impact of course formats, traditional campus courses, online courses and the newer hybrid courses on ETS scores.

METHODOLOGY

A 20-item multiple-choice comprehensive examination was given to students in each of three sections of the Principles of Marketing course at a regional university in the Spring 2008 semester. This exam was one component of a two-part final exam administered at the end of the semester, and was incentivized to motivate student preparation. The results for the three sections appear in Table 1.

The instrument was a series of questions that cover the entirety of topics presented in a typical Principles of Marketing course. The entire Marketing faculty agreed upon the corpus of questions in Fall 2007. Each of the questions were mapped to specific learning objectives for the course.

Two professors taught this course during the Spring 2008 semester. Professor 1 is a tenure-track professor with a PhD, and taught sections 1 (on-campus) and 3 (online). Professor 2 is a PTI (part-time instructor) with an MBA and 18 graduate hours of study in Marketing, and taught section 2.

Numerous studies have sought to map the relationship between student input variables and outcome assessments. Generally speaking, these models delineate multiple factors (or internal student drivers) impacting assessment scores, including Native Ability (intelligence, often measured by ACT and/or SAT scores), Student Effort (often measured by the student's cumulative GPA), and Student Traits (a vector of categorical demographic information, including gender, transfer status, and nationality). A fourth set of factors include Environment Variables, which accounts for course format (online, campus, etc.) and the individual professor of record. Figure 1 illustrates the relationships of this model. Four models were calculated in this study based on these relationships, with the broadest model specified as follows. The inclusion of specific variables is based on the work cited above in predicting ETS exam scores, with the exception of the Professor variable. This variable was included in an effort to extend the research and to search for possible hidden influencers.

[FIGURE 1 OMITTED]

The dependent variable in this study was the student's score on the 20-item exam, with each correct response earning five points. Scores could thus range from 0 to 100. Independent variables were compiled from the university Registrar's database after the semester ended, and included a variety of internally-scaled and binary variables. Cumulative GPA is on a scale from 0 to 4.00, and represents only the student's GPA since the time of their matriculation at this university. The ACT score is the student's composite score on the national entrance exam, and ranges from 1 to 36. The remaining independent variables (student nationality, transfer status, gender, course format, and professor) were binary variables coded as either a 0 or 1.

Based on upon the prior research of Terry, et al (2008), Mirchandani, et al, (2001), and Black & Duhon (2003), we hypothesize the following relationships:

H1: ACT will have a significant positive influence on a student's outcome assessment score. Terry, et al, and others are in unison in finding that ACT (or SAT), as a measure of native ability, is a strong predictor of student assessment outcomes.

H2: GPA will have a significant positive influence on a student's outcome assessment score. Similarly, student effort, as measured by cumulative GPA, is a strong predictor. Terry, et al, and numerous others concur that this variable will yield a positive and significant influence on outcomes scores.

H3: Transfer status will not have a significant influence on a student's outcome assessment score. Transfer status has been included in prior research, but has not yet been found to be a significant predictor. We thus do not predict any significant influence in either direction.

H4: Nationality will not have a significant influence on a student's outcome assessment score. Terry, et al, also included student nationality as a dichotomous variable, and found no significant influence. We likewise have no reason to predict any significance from this variable.

H5: Class Format will not have a significant influence on a student's outcome assessment score. While much research has been done comparing online and--on-campus outcomes, no trend has emerged favoring one over the other. We thus propose no significant influence from this variable.

H6: Gender will have a significant influence on a student's outcome assessment score. Terry, et al, found that females scored significantly higher on their ETS exam, and we likewise propose a similar significant relationship.

H7: Professor will not have a significant influence on a student's outcome assessment score. This variable has not been included in the prior cited work, but given that many schools employ multiple professors to teach their foundational courses, it is possible that an "instructor effect" may be apparent. In the absence of prior research, we posit there will be no difference, assuming similar effectiveness across instructors.

RESULTS

The initial sample was comprised of 151 students who took the assessment exam; this group was labeled the "full" model. The "partial" model was comprised of 63 students for whom the university had an ACT score on record. Since the university does not require an ACT score for incoming transfer students, and because a large percentage of students are transfers, the "partial" model was based on a much smaller sample.

Tables 1 and 2 contain course section means for the full and partial (ACT) groups. The overall mean score for the full group was 72.58, while the overall mean among those with ACT scores was 72.85. Given that the majority of students reporting an ACT score were four-year students at the university (and the non-ACT students were transfers), it can be concluded there is no difference in mean score between these two student groups.

Four linear regression models were then calculated, as summarized in Table 3. A 2 X 2 matrix of the models includes both "full" and "partial" models, as well as those with and without the effect calculated for the professor.

The first model (Basic Model; top-left) is built on the full data set and contains the basic vector of variables (GPA, Transfer, Nationality, Gender, and Format). In this model both Format and GPA were significant predictors at the Prob.=0.05 level, suggesting that the online format and higher overall student GPAs were positively related to student scores on the assessment exam. Only 25% of the variance is accounted for by the model, though. The data for the three sections suggested, though, that the explanatory power of Format may not be sufficient to capture other nuances of the data. A categorical "dummy" variable was created to account for the professor of these three sections, the results indicating in the second model (Basic Model + Prof; lower-left) that GPA and Professor are now significant at the Prob.=0.05 level. This manipulation resulted in the model accounting for 39% of the variance, a substantial improvement over the first model.

Of the four models created, the Partial + Professor model resulted in the highest R2 (0.458). In this model only the Gender (H6) and Professor (H7) variables were significant at the Prob.=0.05 level. GPA (H2) was marginal at Prob.= 0.06, and thus conditionally retained. Surprisingly, ACT (H1) was not significantly related, in spite of the findings in prior research. In this instance, it is possible that the Professor effect was more profound and offset any native ability students may have possessed going into the course.

Of particular interest is the effect of Gender. While we affirm the findings of Terry, et al, Gender did not appear as a significant variable in the Full Model (column 1), but did so in the Partial Model (column 2). A Key Influencers analysis of the variables revealed that female students were tightly clustered in the 76-84 score range, and once the Professor variable was added to the equation, Gender emerged as a very strong predictor.

Further analyses were conducted by calculating t-tests in both the Full and Partial models using the dichotomous independent variables with the dependent variable (test score). These results are summarized in Table 4, indicating Format and Professor yielded significant differences in test scores. The models above, though, indicate somewhat different results when variables are not isolated.

CONCLUSIONS AND FUTURE RESEARCH

The statistical models presented above illustrate the differences in outcomes scores that can accrue in one topical area in multiple format/professor situations. In this particular case, the most significant predictor was the professor, not the format. The implication of these findings is that departments and colleges engaged in the assessment process need to be cognizant of influences from different sources that may not otherwise have been observed or isolated.

In this particular study, differences in outcomes assessment were not so much a factor of course delivery format as they were who was teaching the course. The difference in scores may be attributed to the differences between a full-time doctorally-qualified professor and a part-time adjunct with a master's degree. This is not to say that non-PhD instructors are less effective than those with a PhD; rather, it illustrates the need for departments and colleges of business to carefully dissect the data when analyzing assessment scores.

This research is intentionally narrow in scope, and is thus not necessarily transferable across disciplines or schools as it stands. With only a small matrix of course sections, formats, and professors included in the study, there is room for further exploration by applying this model across disciplines, time, and more faculty. Of greater potential importance is the possibility of tracking these students through the remainder of their academic program, including their outgoing assessment exam in the capstone Strategy course.

REFERENCES

Adams, C., Thomas, R & King, K. (2000). Business Student's Ranking of Reasons for Assessment: Gender Differences, Innovation in Education and Training International.

Allen, J. S., & Bycio, P. (1997). An evaluation of the Educational Testing Service Major Field Achievement Test in Business. Journal of Accounting Education, 15, 503-514.

Aloi, S. L., Gardner, W. S. & Lusher, A. L. (2003). A framework for assess general education outcomes within majors. JGE: The Journal of General Education, 52(4), 237-252

Angelo, T.A., & Cross, K. P. (1993) Classroom assessment techniques: A handbook for college teachers (2nd ed). San Francisco: Jossey-Bass

Bagamery, B., Lasik, J., & Nixon, D. (2005). Determinants of success on the ETS Business Major Field Exam for students in an undergraduate multisite regional university business program. Journal of Education for Business, 81, 55-63.

Black, H. & Duhon, D. (2003). Evaluating and Improving Student Achievement in Business Programs: The Effective use of Standardized Assessment Tests. Journal of Education for Business, 79, 157-162.

Ebel, R. I. & Frisbie, D. A. (1991) Essentials of educational measures (5th ed.), Englewood Cliffs, NJ: Prentice-Hall.

Lubinescu, E.S., Ratcliff, J.L., & Gaffney, M.A., (2001). Two continuums collide: Accreditation and Assessment, New Directions for Higher Education, 113, 5-21.

Martell, K. (2007). Assessing student learning: Are business schools making the grade? The Journal of Education for Business, 82(4), 189-195.

Mirchandani, D., Lynch, R, & Hamilton, D. (2001). Using the ETS Major Field Test in business: Implications for assessment. The Journal of Education for Business, 77, 51-56.

Muchlitsch, J. F., & Sidle, M. W. (2002) Assessing Student Learning Outcomes: A Comparative Study of Techniques Used in Business School Disciplines. Journal of Education for Business, January/ February 2002, 125--130

Palomba, C.A. & Banta, T.W. (1999). Assessment essentials: Planning, implementing, and improving assessment in higher education. San Francisco: Jossey-Bass.

Pringle, C. & Michel, M. (2007). Assessment practices in AACSB--accredited business schools. The Journal of Education for Business, 82(4), 202-211.

Terry, N. (2000). The effectiveness of virtual learning in economics. The Journal of Economics & Economic Education Research, 1(1), 92-98

Terry, N., Mills, L., & Sollosy (2008). Performance of domestic versus international students on the ETS exam, Southwest Review of International Business Research, 19(1), 191-195

Trapnell, J. (2005). Forward. In K. Martell & T. Calderon Eds. Assessment of student learning in business schools. Tallahassee, FL: Association for Institutional Research.

R. Nicholas Gerlich, West Texas A&M University

Marc Sollosy, West Texas A&M University
Table 1: Exam Scores by Section, Format,
and Professor: Full Sample (161)

Section Avg. Score Class Size Format Professor

1 79.06 53 Classroom 1
2 60.27 55 Classroom 2
3 81.32 53 Online

Table 2: Exam Scores by Section, Format,
and Professor: ACT-Only Sample (68)

Section Avg. Score Class Size Format Professor

1 80.96 31 Classroom 1
2 60.83 24 Classroom 2
3 80.38 13 Online

Table 3:: Model Comparison

Full Model (all cases) Partial Model (ACT score available)
Basic Model Basic + ACT
n = 161 n = 68
[R.sup.2] = .206 [R.sup.2] = .274
Adj [R.sup.2] = .181 Adj [R.sup.2] = .203

 IV t prob IV t prob

Format 4.724 0.000 Format 2.087 0.041
GPA 3.124 0.002 GPA 3.249 0.002
Nat 1.303 0.194 Nat 1.046 0.300
Transfer -0.869 0.386 ACT 0.953 0.344
Gender 0.895 0.372 Transfer -0.643 0.532
 Gender 2.334 0.023

Basic Model + Prof Effect Partial Model + ACT + Prof Effect
n = 161 n = 68
[R.sup.2] = .397 [R.sup.2] = .458
Adj [R.sup.2] = .374 Adj. [R.sup.2] = .395

 IV t prob IV t prob

Format 1.116 0.266 Format 0.617 0.540
GPA 2.402 0.017 GPA 1.910 0.061
Nat 0.715 0.476 Nat 0.826 0.412
Transfer -0.435 0.664 ACT 0.901 0.371
Gender 1.276 0.204 Transfer -0.279 0.781
Prof 6.988 0.000 Gender 2.624 0.011
 Prof 4.520 0.000

Table 4: T-Tests

Full Model Partial Model

 IV t prob IV t prob

Format -4.747 0.000 Format -1.674 0.099
Gender -0.555 0.580 Gender -0.587 0.559
Transfer 0.411 0.682 Transfer 0.500 0.619
Nat -1.624 0.106 Nat -1.243 0.218
Prof -9.423 0.000 Prof -6.044 0.000


联系我们|关于我们|网站声明
国家哲学社会科学文献中心版权所有