enow.com Web Search

Search results

  1. Results from the WOW.Com Content Network
  2. Item analysis - Wikipedia

    en.wikipedia.org/wiki/Item_analysis

    The process of item analysis varies depending on the psychometric model. For example, classical test theory or the Rasch model call for different procedures. In all cases, however, the purpose of item analysis is to produce a relatively short list of items (that is, questions to be included in an interview or questionnaire) that constitute a ...

  3. Item-total correlation - Wikipedia

    en.wikipedia.org/wiki/Item-total_correlation

    An important goal of item analysis is to identify and remove or revise items that are not good indicators of the underlying trait. [2] A small or negative item-correlation provides empirical evidence that the item is not measuring the same construct measured by the assessment. Exact values depend on the type of measure, but as a heuristic, a ...

  4. Classical test theory - Wikipedia

    en.wikipedia.org/wiki/Classical_test_theory

    Item analysis within the classical approach often relies on two statistics: the P-value (proportion) and the item-total correlation (point-biserial correlation coefficient). The P-value represents the proportion of examinees responding in the keyed direction, and is typically referred to as item difficulty .

  5. Kuder–Richardson formulas - Wikipedia

    en.wikipedia.org/wiki/Kuder–Richardson_formulas

    Often discussed in tandem with KR-20, is Kuder–Richardson Formula 21 (KR-21). [4] KR-21 is a simplified version of KR-20, which can be used when the difficulty of all items on the test are known to be equal. Like KR-20, KR-21 was first set forth as the twenty-first formula discussed in Kuder and Richardson's 1937 paper.

  6. National Achievement Test - Wikipedia

    en.wikipedia.org/wiki/National_Achievement_Test

    The test is a system-based assessment designed to gauge learning outcomes across target levels in identified periods of basic education. Empirical information on the achievement level of pupils/students serve as a guide for policy makers, administrators, curriculum planners, principles, and teachers, along with analysis on the performance of regions, divisions, schools, and other variables ...

  7. Differential item functioning - Wikipedia

    en.wikipedia.org/wiki/Differential_item_functioning

    Differential item functioning (DIF) is a statistical property of a test item that indicates how likely it is for individuals from distinct groups, possessing similar abilities, to respond differently to the item. It manifests when individuals from different groups, with comparable skill levels, do not have an equal likelihood of answering a ...

  8. Automatic item generation - Wikipedia

    en.wikipedia.org/wiki/Automatic_Item_Generation

    AIG reduces the cost of producing standardized tests, [10] as algorithms can generate many more items in a given amount of time than a human test specialist. It can quickly and easily create parallel test forms, which allow for different test takers to be exposed to different groups of test items with the same level of complexity or difficulty, thus enhancing test security. [3]

  9. Compositional data - Wikipedia

    en.wikipedia.org/wiki/Compositional_data

    In a survey, the proportions of people positively answering some different items can be expressed as percentages. As the total amount is identified as 100, the compositional vector of D components can be defined using only D − 1 components, assuming that the remaining component is the percentage needed for the whole vector to add to 100.