Navigation Menu

Click the "+" to see inside a chapter or use the search to the right.

.

Skip to end of metadata
Go to start of metadata

The Exam Form Statistics in the Observer Dashboard describes the various statistics associated with every exam form in the project you are working on. The user can select an exam form from the drop-down and view the report for the corresponding exam form below in the Exam Form Observer Dashboard.

On selecting an exam form from the drop-down, information about the exam form is displayed below like the page displayed below:

The following statistics are displayed:

Exam Form Statistics:

  1. Exam Form: Shows the name of the exam form selected. The exam form ID is mentioned in parentheses following a # sign.
  2. Average Difficulty (Angoff): The average Angoff rating of the items in the exam form.
  3. SEM: The Standard Error of Measurement of the exam form.

    Icon
    SEM is an estimate of error to use in interpreting a candidate’s test score.



  4. Average Validation Rating for Test: The average validation rating for the exam form.

  5. Questions selected:
    1. Scored: The number of items scored in the exam form.
    2. Experimental: The number of experimental items in the exam form.
  6. Percentage Shared Among Versions: The percentage of items shared among the versions of an exam form.
  7. Metadata (% of Test): The percentage of metadata belonging to different metadata contents in an exam form.

This page also shows the details of the ratings assigned to items in the project:

Column

Description

Username

The user working on the project.

# Question Rated

The number of items rated by the user.

Avg Angoff

The average Angoff ratings of the items assigned by the user.

# Rating Outliers

The number of items with outlier rating on the exam form.

Rating Difference

The difference between the average Angoff and rating outliers.

Tracking user progress in a Standard Setting session

Icon

This table is a great tool to see which users have finished work and which still have work to do using the # Questions Rated column in the table above. If you know how many items are on the exam form, it will be easy to see which users you should ping to complete their work.

 

Another section included is the Validation Data Report.

This report shows the validation performed on the items belonging to the selected exam form in drop-down with the following details:

Column

Description

Seq.#

Sequence on exam form.

Question #

Question ID (not DB ID).

Revision Code

Displays the Revision Code number of the item.

Blueprint ID

Displays Blueprint ID of the item.

Question Author

Displays the Author name, Date and Time for the item.

Users who Approved Question

Displays the users who approved that item.

Avg. Importance

Average Importance Rating (using 0 -4 rating across all raters of current version of the item).

Avg. Criticality

Average Criticality Rating (using 0 -4 rating across all raters of current version of the item).

Avg. Frequency

Average Frequency Rating (using 0 -4 rating across all raters of current version of the item).

Avg. Angoff

Average Angoff Rating (Average Angoff Rating across all raters who rated the current version of this item).

#Raters

The number of raters who were used in calculating the averages listed above.

Raters

All of the raters whose ratings were used in calculating the values above.

 

 

  • No labels