100% satisfaction guarantee Immediately available after payment Both online and in PDF No strings attached 4.6 TrustPilot
logo-home
Exam (elaborations)

VALIDITY, RELIABILITY AND DIFFICULTY INDICES FOR INSTRUCTOR-BUILT EXAM QUESTIONS

Rating
-
Sold
-
Pages
5
Grade
A+
Uploaded on
10-08-2024
Written in
2024/2025

To assess the reliability of the tests, we needed to use a number of experts to mark the exam papers in order that the marking does not affect the marker’s opinion( seif 2004). In this study, we asked two instructors to mark the exam papers separately and used Kendal agreement coefficient to check the agreement of the two markings. Difficulty Coefficient and Discriminant Coefficient Because all of chemistry exam questions were open questions, we used the following formula for calculating the difficulty coefficient(DifCo). () () ( ) * Si Wi question i B i M M DifCoef N m + = Where MS(i)= sum of marks for Strong group in question i MW(i)= sum of marks for Weak group in question i NB=number of students in both groups mi=total mark of question i And the Discriminant Coefficient(DisCo) was calculated based on the following formula(Kiamanesh 2002). () () ( ) * Si Wi question i g i M M DisCoef n m − = Where MS(i)= sum of marks for Strong group in question i MW(i)= sum of marks for Weak group in question i ng=number of students in one group mi=total mark of question i Results The percentages of papers were almost equal in terms of students’ sex(49% males and 51% females). The characteristics of the exam questions is summarized in Table 1. Table 1. Exam characteristics by book chapters Characteristic knowledge concept application chapter mark percent mark percent mark percent total 1 41.5 11.5 78 21.7 8.25 2.3 127.75 2 36 10 85.5 23.7 2 0.6 123.5 3 30.75 8.5 20.5 5.7 2.25 0.6 53.5 4 32.5 9.1 22.75 6.3 0 0 55.25 total 140.75 39.1 206.75 57.4 12.5 3.5 360 Evaluation of Academic Activities in Universities 154 Table1 shows that most chemistry questions were on concept(57.4%) and percentages on knowledge(39.1%) and small percentage on application(3.5%).There were no questions on analysis, combination and evaluation in the exams. As stated before, the agreement of instructors evaluations was calculated using Kendal’s agreement coefficient. The value of the coefficient was 0.49 which was significant at p-value of 0.05. The Kendal’s agreement coefficient for face validity of the questions based on the evaluation of expert instructors was 0.42 and significant at p-value<0.05). The reliability coefficient based on markers’ evaluations was ) 0.971 and significant(p<0.0001). The minimum and maximum difficulty coefficients estimated were DifCoef(min)=0.14 and DifCoef(max)=1 with standard error of 0.16 which indicates that the questions have moderate difficulty level. The minimum and maximum discriminant coefficients were DisCoef(min)=0.07 and DisCoef(max)=0.98 with standard error of 0.20 indicating that the questions have good discriminant coefficient. We also found no significant difference for content validity and reliability between female and male instructors. Then we compared the Difficulty coefficient and discriminate coefficient between two sexes of instructors. The test results are shown in Tables 2 and 3. Table 2. Chi- sqaure test for comparison of difficulty coefficients between female and male instructors Difficulty level # of questions from female instruct

Show more Read less
Institution
VALIDITY, RELIABILITY AND DIFFICULTY
Module
VALIDITY, RELIABILITY AND DIFFICULTY









Whoops! We can’t load your doc right now. Try again or contact support.

Written for

Institution
VALIDITY, RELIABILITY AND DIFFICULTY
Module
VALIDITY, RELIABILITY AND DIFFICULTY

Document information

Uploaded on
August 10, 2024
Number of pages
5
Written in
2024/2025
Type
Exam (elaborations)
Contains
Questions & answers

Subjects

Content preview

Evaluation of Academic Activities in Universities




VALIDITY, RELIABILITY AND DIFFICULTY INDICES FOR
INSTRUCTOR-BUILT EXAM QUESTIONS1




Gholamreza JANDAGHI2
PhD, Associate Professor Faculty of Management, University of Tehran, Qom Campus, Iran




E-mail:


Fatemeh SHATERIAN3
MSc, Academic member of Islamic Azad University, Saveh Branch, Iran




E-mail:




Abstract: The purpose of the research is to determine college Instructor’s skill rate in
designing exam questions in chemistry subject. The statistical population was all of chemistry
exam shits for two semesters in one academic year from which a sample of 364 exam shits
was drawn using multistage cluster sampling. Two experts assessed the shits and by using
appropriate indices and z-test and chi-squared test the analysis of the data was done. We
found that the designed exams have suitable coefficients of validity and reliability. The level of
difficulty of exams was high. No significant relationship was found between male and female
instructors in terms of the coefficient of validity and reliability but a significant difference
between the difficulty level in male and female instructors was found(P<.001). It means that
female instructors had designed more difficult questions. We did not find any significant
relationship between the instructors’ gender and the coefficient of discrimination of the exams.

Key words: instructor-built exam; content validity; face validity; reliability; coefficient of
discrimination; coefficient of difficulty



1. Introduction

Examination and testing is an important part of a teaching-learning process which
allows instructors to evaluate their students during and at the end of an educational course.
Many instructors dislike preparing and grading exams, and most students dread taking them.
Yet tests are powerful educational tools that serve at least four functions. First, tests help you




151

, Evaluation of Academic Activities in Universities



evaluate students and assess whether they are learning what you are expecting them to
learn. Second, well-designed tests serve to motivate and help students structure their
academic efforts. Crooks (1988), McKeachie (1986), and Wergin (1988) report that students
study in ways that reflect how they think they will be tested. In last 40 years the most exams
used to evaluate the students have been designed by instructors. Some may have used tests
which have been designed by outsider exam designers. These tests have not had enough
efficiency (Seif 2004). Regarding the importance of instructor-designed test in evaluation
process of the students, many researches have been done in this area (Lotfabadi 1997). In
theory, the best test for a subject is a test that includes all educational objectives of the
course. But if the test is too long, its preparation is impractical. Therefore, instead of
including all content and objectives, one may choose some questions which are
representative of the whole subject to achieve all objectives. Such a test is said to have
content validity (Seif 2004).
Content validity of a instructor-designed test can be assessed by a sample of the
test questions. When a test does not have content validity two possible outcomes may occur.
First, the students can not present the skills that are not included in the test when they need.
Second, instead some unrelated question may be included in the test that are answered
wrongly. The important point here is that we should not mistake the face validity with
content validity. Basically the face validity is a measure that determines whether a test is
measuring logically and whether students think the test questions are appropriate ( Lotfabadi
1997).
Based on what is said, an ideal test in addition to measuring what is supposed to
measure, must be consistently constant in different times. This characteristic is called
reliability. Other measures of an ideal test are difficulty level and discriminant index. The
total percent of the individuals who answer the question correctly is known as difficulty
coefficient denoted by P (Seif 2004). The discriminant index is a measure of discrimination
between strong and weak groups. In this study, we intend to evaluate the extent of ideal
quality measures (validity, reliability,…) in instructor-designed test for first year college.


Materials and methods

The statistical population in this study consisted of all chemistry exam papers for
final chemistry exams in first and second semester for first year of college in Qom province
of Iran of which a sample of 364 was taken. A twostage cluster sampling was used to draw
samples. In first stage three colleges was randomly selected. In second stage a number of
exam papers from each college was selected according to the number of students in each
college.
In this study the content validity of the exam questions was assessed in two ways. In
the first method we used a two dimensional table. One dimension was educational goals
and the second dimension was the content of the course materials(Seif 2004). The second
method applied for assessing content validity was a questionnaire with Likert scale in which
two chemistry education expert evaluated the extent of compatibility of exam questions with
course contents. For assessment of face validity of instructor-built exams we used a 12-item
questionnaire answered by two chemistry experts.




152

Get to know the seller

Seller avatar
Reputation scores are based on the amount of documents a seller has sold for a fee and the reviews they have received for those documents. There are three levels: Bronze, Silver and Gold. The better the reputation, the more your can rely on the quality of the sellers work.
TIFFACADEMICS Liberty University
Follow You need to be logged in order to follow users or courses
Sold
673
Member since
2 year
Number of followers
375
Documents
6218
Last sold
1 day ago
REING SUPREME SCHOLARLY // ENLIGHTENED

Here we offer revised study materials to elevate your educational outcomes. We have verified learning materials (Research,Assignments,notes etc...) for different courses guaranteed to boost your academic results. We are dedicated to offering you the best services and you are encouraged to inquire further assistance from our end if need be. Having a wide knowledge in Nursing,trust us to take care of your Academic materials and your remaing duty will just be to Excel. Remember to give us a review,it is key for us to understand our clients satisfaction. We highly appreciate refferals given to us. Also clients who always come back for more of the study content your offer are extremely valued. ALL THE BEST.

Read more Read less
3.7

132 reviews

5
59
4
13
3
33
2
11
1
16

Recently viewed by you

Why students choose Stuvia

Created by fellow students, verified by reviews

Quality you can trust: written by students who passed their exams and reviewed by others who've used these revision notes.

Didn't get what you expected? Choose another document

No problem! You can straightaway pick a different document that better suits what you're after.

Pay as you like, start learning straight away

No subscription, no commitments. Pay the way you're used to via credit card and download your PDF document instantly.

Student with book image

“Bought, downloaded, and smashed it. It really can be that simple.”

Alisha Student

Frequently asked questions