flowchart LR
O[Objective<br/>What to evaluate] --> L[Learning<br/>Experience]
L --> T[Tools /<br/>Techniques]
T --> C[Criterion<br/>Standard]
C --> J[Judgment]
classDef default fill:#003366,color:#ffffff,stroke:#ffcc00,stroke-width:3px,rx:10px,ry:10px;
6 Evaluation Systems in Higher Education
Evaluation in higher education has four examined sub-topics:
- Elements and types of evaluation.
- Choice Based Credit System (CBCS) evaluation.
- Computer Based Testing (CBT).
- Innovations in evaluation systems.
6.1 Measurement, Assessment, Evaluation — the Three Layers
These three terms are often confused. NTA stems regularly ask the candidate to distinguish them.
| Term | What it does | Output | Example |
|---|---|---|---|
| Measurement | Quantification — assigning a number | A score | Marks out of 100 in a test |
| Assessment | Collecting evidence about learning | Multiple data points (qualitative + quantitative) | Tests + projects + observation |
| Evaluation | Judging value against a criterion or objective | A judgment / decision | “Pass / fail” or grade |
Measurement is neutral; evaluation is judgmental. Assessment is the bridge — gathering the evidence on which a judgment can be made.
6.2 Elements of Evaluation
Every evaluation act has four elements.
| Element | What it specifies | Example |
|---|---|---|
| Objective | What is being evaluated | “Identify the main idea in a 200-word passage” |
| Learning experience | How the learner was prepared | Lecture + tutorial + practice exercises |
| Tools / Techniques | Instrument used to gather evidence | Test, quiz, rubric, observation schedule |
| Criterion | Standard against which evidence is judged | Score ≥ 60 for pass; rubric levels 1–4 |
6.3 Types of Evaluation
Evaluation is classified along three independent dimensions: purpose, reference, and timing.
6.3.1 By Purpose
| Type | Purpose | When | Example |
|---|---|---|---|
| Placement | Decide where the learner should start | Before instruction begins | Diagnostic test before joining a programme |
| Diagnostic | Identify learning difficulties or misconceptions | During instruction | Concept-inventory test mid-semester |
| Formative | Monitor progress; give feedback to improve | Throughout instruction | Weekly quiz, assignment, peer review |
| Summative | Certify achievement at the end | At the end of instruction | End-semester exam, final project |
6.3.2 By Reference Standard
| Type | Reference | What it answers | Example |
|---|---|---|---|
| Norm-referenced | Performance compared to others | “Where does this learner stand relative to peers?” | Percentile ranking; bell-curve grading |
| Criterion-referenced | Performance compared to a standard | “Has the learner met the criterion?” | Driving licence test; mastery learning |
6.3.3 By Timing — Continuous vs Terminal
| Type | Pattern | Strengths |
|---|---|---|
| Continuous Comprehensive Evaluation (CCE) | Frequent low-stakes assessments throughout the term | Reduces exam pressure; tracks growth; integrates affective and psychomotor domains |
| Terminal evaluation | One high-stakes assessment at the end | Efficient; standardised; comparable across institutions |
flowchart TB
E[Evaluation] --> P[By Purpose]
E --> R[By Reference]
E --> T[By Timing]
P --> P1[Placement]
P --> P2[Diagnostic]
P --> P3[Formative]
P --> P4[Summative]
R --> R1[Norm-referenced]
R --> R2[Criterion-referenced]
T --> T1[Continuous CCE]
T --> T2[Terminal]
classDef default fill:#003366,color:#ffffff,stroke:#ffcc00,stroke-width:3px,rx:10px,ry:10px;
6.4 Choice Based Credit System (CBCS)
The University Grants Commission introduced the Choice Based Credit System for Indian higher education. CBCS allows learners to choose courses from core, elective and skill-based options and uses a credit system that is portable across institutions.
| Feature | What it specifies |
|---|---|
| Course types | Core, Elective (Discipline-Specific / Generic), Ability Enhancement, Skill Enhancement |
| Credit definition | One credit = one hour lecture / two hours tutorial or lab per week, per semester |
| Grading | Letter grades on a 10-point scale, mapped to grade points |
| Assessment split | Internal (formative, ~ 25 %) + External (summative, ~ 75 %) |
| Mobility | Credits earned at one institution transfer to another following CBCS |
6.4.1 CBCS 10-Point Grading Scale
| Letter grade | Grade point | Description |
|---|---|---|
| O | 10 | Outstanding |
| A+ | 9 | Excellent |
| A | 8 | Very good |
| B+ | 7 | Good |
| B | 6 | Above average |
| C | 5 | Average |
| P | 4 | Pass |
| F | 0 | Fail |
| Ab | 0 | Absent |
The summary statistics from this scale are SGPA (Semester Grade Point Average) and CGPA (Cumulative Grade Point Average).
- SGPA = Σ (credits × grade points) / Σ credits, computed per semester.
- CGPA = Σ (credits × grade points across all semesters) / Σ credits, computed cumulatively.
6.5 Computer Based Testing (CBT)
Computer Based Testing is the model the National Testing Agency (NTA) uses for UGC-NET, JEE Main, NEET-PG and other large examinations.
| Feature | What it does | Implication |
|---|---|---|
| On-screen presentation | Items appear on a computer monitor | Random order possible; adaptive difficulty possible |
| Automated scoring | Software grades MCQs immediately | Faster results; less subjectivity in scoring |
| Item bank | Items drawn from a pre-validated pool | Multiple sittings can have different papers of equal difficulty |
| Centralised security | Encrypted delivery; biometric authentication | Reduces leakage and impersonation |
| Statistics on the fly | Time spent per question, attempt rate | Diagnostic feedback for both learner and item designer |
| Dimension | Computer Based Test | Paper-and-pencil |
|---|---|---|
| Speed of result | Hours to days | Weeks |
| Logistics | Need terminal, software, power | Need only printed paper |
| Adaptive testing | Possible | Not possible |
| Question types | MCQ, drag-drop, hot-spot, simulation | Mostly MCQ and written |
| Risk | Hardware failure, connectivity | Paper leakage, transport |
6.5.1 Computer Adaptive Testing (CAT)
A subtype of CBT in which the software changes the difficulty of the next item based on whether the learner answered the previous item correctly. CAT shortens test length while preserving measurement precision. The Graduate Record Examinations (GRE) is the best-known CAT example.
6.6 Innovations in Evaluation Systems
| Innovation | What it does | When to use |
|---|---|---|
| Open-book examination | Reference material allowed; tests application | Application and analysis; postgraduate work |
| Online proctored examination | Webcam-based remote invigilation | Distance learners; pandemic conditions |
| Portfolio assessment | Learner curates a body of work over time | Studio, design, language, teacher education |
| Peer assessment | Learners evaluate each other against a rubric | Building evaluative judgment in the learner |
| Self-assessment | Learner judges own work against a rubric | Reflective practice |
| Rubric-based assessment | Pre-stated criteria with level descriptors | Open-ended tasks (essays, projects) |
| Competency-based assessment | Tests mastery of named competencies | Professional courses; outcome-based education |
| AI-powered plagiarism detection | Software compares against reference corpora | Theses, term papers, journal submissions |
flowchart TB
E[Evaluation<br/>Innovations] --> O[Open-book exam]
E --> P[Online proctoring]
E --> PA[Portfolio]
E --> PE[Peer assessment]
E --> S[Self-assessment]
E --> R[Rubrics]
E --> CB[Competency-based]
E --> AI[AI plagiarism check]
classDef default fill:#003366,color:#ffffff,stroke:#ffcc00,stroke-width:3px,rx:10px,ry:10px;
6.7 Outcome-Based Education and NEP-2020
NEP-2020 endorses outcome-based education (OBE) — defining the learning outcomes of every programme and aligning curriculum, teaching and assessment to those outcomes (Ministry of Education, Government of India, 2020). The candidate should recognise the OBE alignment triangle.
flowchart TB
L[Programme<br/>Learning Outcomes] --> C[Curriculum]
L --> T[Teaching<br/>Methods]
L --> A[Assessment]
C -.-> A
T -.-> A
classDef default fill:#003366,color:#ffffff,stroke:#ffcc00,stroke-width:3px,rx:10px,ry:10px;
The triangle is sometimes called the constructive alignment model (Biggs). The senior teacher writes the outcomes first, then designs the assessment, and only then designs the teaching — assessment-led design.
6.8 Validity, Reliability, and Practicability
A good evaluation tool has three essential properties.
| Property | What it asks | Threat if absent |
|---|---|---|
| Validity | Does the tool measure what it claims to? | Wrong inference about the learner |
| Reliability | Does the tool give consistent results across administrations and graders? | Score depends on which day or which grader |
| Practicability | Can the tool be administered with available time, cost and expertise? | Tool exists on paper but not in practice |
A reliable test that measures the wrong thing is useless precision. A valid test that gives different results each time is useful in principle but unworkable. A perfect test that needs three hours of one-to-one administration with each learner is valid and reliable but impracticable.
6.9 Practice Questions
Which of the following is not a type of evaluation by purpose?
View solution
A teacher gives a weekly quiz so that students get feedback during the term. This is best classified as:
View solution
In CBCS, one credit usually corresponds to:
View solution
Match the term with its function:
| (i) | Measurement | (a) | Judgment of value against a criterion |
| (ii) | Assessment | (b) | Quantification — assigning a number |
| (iii) | Evaluation | (c) | Collecting evidence about learning |
View solution
A test that gives different scores when administered by two different graders is low in:
View solution
In Computer Adaptive Testing (CAT):
View solution
Which of the following is a feature of an open-book examination?
View solution
In the UGC's 10-point grading scale, which letter grade carries the highest grade point?
View solution
- Measurement → number; Assessment → evidence; Evaluation → judgment.
- Types by purpose: Placement, Diagnostic, Formative, Summative (“PDFS”).
- By reference: Norm-referenced (vs peers), Criterion-referenced (vs standard).
- CBCS: 10-point grading (O = 10, A+ = 9, … F = 0); SGPA per semester, CGPA cumulative.
- One credit = 1 hour lecture/week/semester (or 2 hours lab/tutorial).
- CBT features: on-screen, automated scoring, item bank, encrypted delivery, statistics.
- CAT (Computer Adaptive Testing) — difficulty adapts in real time.
- Three properties of a good tool: Validity, Reliability, Practicability.
- NEP-2020 → outcome-based education + constructive alignment.