Abstract
This study focuses on students' performance in a distance-learning computer-based C course offered by EPGY (Education Program for Gifted Youth) at Stanford University. Its main goal is testing the feasibility of automated evaluation. Three automated methods consisting of a criterion set and weighted scoring methods were developed empirically, then applied to four hundred programs of ten types of problems and thoroughly analyzed. For most of the assignments, the more significant criteria for evaluation were related to execution time features, (i.e. number of operations and calls to functions executed by the program), memory storage required to store the variables of the program, and code's length. The results showed that the criterion measures often differed by several orders of magnitude between student's programs and were dependent on the specific problem domain. A low correlation was found between the evaluation criteria indicating that the criteria are nearly independent. Consequently, in qualitative tests for achievement and ability students' performances exhibited high dimensionality.
Original language | English |
---|---|
Pages (from-to) | T1B/7-T1B/12 |
Journal | Proceedings - Frontiers in Education Conference, FIE |
Volume | 1 |
State | Published - 2002 |
Event | 32nd Annual Frontiers in Education; Leading a Revolution in Engineering and Computer Science Education - Boston, MA, United States Duration: 6 Nov 2002 → 9 Nov 2002 |
Keywords
- Automated criteria
- Program evaluation
- Students' individual differences
- Students' performances