Assessment of Architecture Design using Rasch Measurement Model
Transcript of Assessment of Architecture Design using Rasch Measurement Model
PeKA’12
Mengilham Harapan, Mencipta Masa Depan
Prosiding
Seminar Pendidikan Kejuruteraan dan Alam Bina 2012
Dianjurkan Oleh
Pusat Penyelidikan Pendidikan Kejuruteraan,
Fakulti Kejuruteraan dan Alam Bina
Universiti Kebangsaan Malaysia
Disunting Oleh
Siti Rozaimah Sheikh Abdullah
Adi Irfan Che Ani
Hafizah Husain
Siti Aminah Osman
Norngainy Mohd. Tawil
Nor Kamaliana Khamis
Zulkifli Mohd Nopiah
Afida Ayob
Nizaroyani Saibani
Roszilah Hamid
Mohd Shahbudin Mastar@Masdar
Noorhisham Tan Kofli
Nur Arzilah Ismail
FAKULTI KEJURUTERAAN DAN ALAM BINA
UNIVERSITI KEBANGSAAN MALAYSIA
BANGI 2012
Cetakan Pertama 2012
Hak cipta Fakulti Kejuruteraan dan Alam Bina
Universiti Kebangsaan Malaysia, 2012
Hak cipta terpelihara. Tiada bahagian daripada terbitan ini boleh diterbitkan semula,
disimpan untuk pengeluaran atau ditukarkan ke dalam sebarang bentuk atau dengan
sebarang alat juga pun, sama dengan cara elektronik, gambar serta rakaman dan
sebagainya tanpa kebenaran bertulis dari Fakulti Kejuruteraan dan Alam Bina,
Universiti Kebangsaan Malaysia
Diterbitkan di Malaysia oleh
FAKULTI KEJURUTERAAN DAN ALAM BINA
UNIVERSITI KEBANGSAAN MALAYSIA
43600 UKM Bangi, Selangor Darul Ehsan
Malaysia
Dicetak di Malaysia oleh
Penerbit UKM
Universiti Kebangsaan Malaysia
x
25. Pengukuran Hasil Pembelajaran Program Untuk Projek Bersepadu Melalui
Soal Selidik Pelajar Jabatan Kejuruteraan Kimia Dan Proses Tahun III
Sesi 2010/2011
Siti Kartom Kamarudin, Siti Rozaimah Sheikh Abdullah, Norhisham Tan
Kofli, Norliza Abdul Rahman, Siti Masrinda Tasirin, Jamaliah Jahim &
Rakmi Abdul Rahman
192
26. Assessment of Total Learning Experience in UKM on Engineering
Student
Azrilah Abdul Aziz, Nuraini Khatimin, Khairul Anwar Mastor & Azami
Zaharim
204
27. Penilaian Terhadap Kebolehpercayaan Instrumen Pengukuran Bagi Kursus
Kalkulus Vektor Untuk Pelajar Kejuruteraan Menggunakan Model
Pengukuran Rasch
Haliza Othman, Izamarlina Asshaari, Hafizah Bahaludin, Norgainy Mohd Tawil & Nur Azilah Ismail
212
28. Using Rasch Measurement Model for Standard Setting on Static and
Dynamic Final Examination in UKM
Nuraini Khatimin, Azrilah Abdul Aziz, Siti Aminah Osman & Azami Zaharim
219
29. Pengukuran dan Penilaian Hasil Pembelajaran Kursus Projek Penyelidikan
Tahun Akhir di Jabatan Kejuruteraan Awam & Struktur, UKM
Azrul A. Mutalib, Shahrizan Baharom, Siti Fatin Mohd. Razali & Wan
Hanna Melini Wan Mokhtar
226
30. Pembentangan Hasil Projek Latihan Ilmiah Melalui Seminar Selari
Nasharuddin Zainal, Muhammad Faiz Bukhori, Rosdiadee Nordin &
Mohd. Hafiz Baharuddin
233
31. Computer Based Analysis in the Design Process for the Second Year
Design Studio Universiti Kebangsaan Malaysia (UKM)
Nangkula Utaberta, Badiossadat Hassanpour, Aisyah Nur Handryant, Adi
Irfan Che Ani & Norngainy Mohd Tawil
238
32. Evaluating Students’ Performance In Complex Analysis Course Through
The Pre-Test Approach
Zulkifli Mohd Nopiah, Norain Farhana Ahmad Fuaad, Nur Suaidah Rosli,Nur Arzilah Ismail & Mohd Noor Baharin
245
33. Assessement of Architecture Design Studios using Rasch Measurement
Model
Nangkula Utaberta, Badiossadat Hassanpour, Adi Irfan Che Ani & Norngainy Mohd Tawil
252
Seminar Pendidikan Kejuruteraan dan Alam Bina 2012
Kongres Pengajaran & Pembelajaran
UKM2012 15 – 18 Disember 2012
http://www.ukm.my/kongres2012
252
Assessement of Architecture Design Studios using
Rasch Measurement Model
Nangkula Utabertaa, Badiossadat Hassanpour
b, Adi Irfan Che Ani
b
& Norngainy Mohd Tawilb
aArchitecture Department
Universiti Putra Malaysia
43400 UPM Serdang, Selangor, Malaysia bArchitecture Department
Faculty of Engineering & Built Environment
Universiti Kebangsaan Malaysia
43600, Bangi, Selangor, Malaysia
Abstract
Architecture as a multidisciplinary filed has a close relation with other fields like art and because of this nature, its education system is completely different with scientific majors. Students expose themselves to critique and evaluation in each session and in this way by gaining experience they try to upgrade their problem solving skills. Definitely such an education system needs different
evaluation and assessment model. Essentially different models and methods are currently implementing in universities and architecture departments all around the world with minimum variations. Unfortunately, despite of its considerable, minimum studies have been conducted and reported in literature. So in order to promote an open discussion and improve the assessment in architecture education, rasch measurement model and mini facet software have been used for the first time to assess the evaluation system at the current architecture studio. This research has been choosing Malaysia (as a model of developing multicultural country) for the study and select Universiti Kebangsaan Malaysia especially architecture department as case study. Keywords: architecture assessment; critique sessions; education; Rasch measurement model.
1. Introduction
There are different assessment methods and grading models that are designed and
implemented by pioneer universities in architecture education and followed by other
universities around the world. By gradually improving the definition of education and
assessment, these models have been upgraded and evaluated. One of the systems is
Nangkula Utaberta et al. / Prosiding - Seminar Pendidikan Kejuruteraan dan Alam Bina 2012
253
comparative method. In this appraisal model the student’s projects will be compared with
each other.
In fact jurors or the related tutors that are going to give marks in submission day,
judge the quality of projects holistically then they rank the projects. Grades follow in
descending form best project to worth one. Comparing students with each other is unfair,
because students are from different backgrounds and talents (Salama 2007). Students
deserve to be graded on the basis of the quality of their work alone, uncontaminated by
reference to how other students in the studio perform on the same or equivalent tasks, and
without regard to each student’s previous level of performance (Sadler 2005). In
comparative system, the holistically attitude to the projects judgment leads to neglect
student’s creativity and abilities in some contexts.
Students cannot be aware of their weak and strong points and by this way they cannot
do any effort to increase their marks and just lucky students who are skillful in graphic
design are able to impact jurors for better grades. On the other hand making pair-wise
comparisons just among small set of students submissions is possible but it will be very
difficult in large amount of projects and students.
In recent years, universities have made explicit overtures towards criteria-based
grading and reporting. Under these models, grades are required to evaluate student’s
achievement in fulfilling juror’s expectations. These expectations can be explain in
different form. We name these expectations as course objectives. The objectives are
assumed to provide the basis for the criteria, but exactly what the criteria are is in essence
left undefined (Sadler 2005).
These objectives should be known by instructors, students and especially external
jurors. Because invited jurors have their certain tendency and assumed objectives that
would be the base of their grading. This incoherency may lead to variant in given marks
by different instructors and students dissatisfaction. One of the implemented methods
under this way is grading system base on marking forms.
These grading criteria sheets (Montgomery 2002) typically do not map in any simple
way into course objectives. They are scoring rubrics which shows some tasks and their
marks portion. These tasks outline some of the knowledge and skills students ideally
should be able to exhibit by the end of the course. For instance, 3D model and executive
details, boards, oral presentation as tasks and 5 mark for each of them.
The given mark is based on the quality of presented documentation. This holistic
method cannot explain the expected details in each task and will leave the doors open to
enter the personal opinions and subjective decisions in evaluation. An underlying
difficulty is that the quality of performance in a course, judged holistically on the basis of
the quality of work submitted, may not be determinable well with the attainment of
course objectives.
Since the purpose of this study is evaluation of current appraisal model and
improvement of current framework to achieve better critique sessions and assessment
model, second year design studio students selected as critical stage in architecture
education for this study.
Therefore to demonstrate the proposed framework, we used Rasch measurement
model and Mini Facet Software to use the software’s prediction and estimation ability
besides evaluating the efficiency and quality of proposed and implemented model. The
main target of this research is to evaluate the implemented assessment and grading model
Nangkula Utaberta et al. / Prosiding - Seminar Pendidikan Kejuruteraan dan Alam Bina 2012
254
in Universiti Kebangsaan Malaysia and evaluate the efficiency and reliability of used
grading model.
2. Findings and Discussions
This research used assessment and grading sheet in second year design studio based on
the main framework and specification of criteria based assessment (Hassanpour, 2012). It
was designed based on the expected objectives of the course and each objective was
clearly defined and expanded by number of tasks. The rubrics of this evaluation sheet are
named as follows:
1. Critical Explanation
2. Logical Development
3. Proposal and recommendation
4. Oral and Graphic Presentation
On the other hand each of these tasks had different percentage mark based on the
importance of each and their priority to the objectives. Also the given marks were defined
into five different levels from fail, poor, average, good and excellent and their definition
to make the ranking as easy and fast as possible for the jurors.
To test this evaluation sheet, Rasch measurement model and mini facet software has
used.
Figure 1 shows Criteria measurement report. Item reliability is >0.7 given as 0.96 to
indicate that the items have a good spread of item difficulties to measure the design
progress of students’ project.
The difficulty level ranges from the maximum logit at 2.43 to the minimum logit (least
difficult) at -1.18logit with a standard deviation of (S.D.) 1.22.
Figure 1. Criteria Measurement Report
Nangkula Utaberta et al. / Prosiding - Seminar Pendidikan Kejuruteraan dan Alam Bina 2012
255
The item separation of 4.87, which can be round of to 5 is different levels of item
difficulty. The most difficult criteria is, M=Model, all the M items are located way above
the rest of the items, Highest at 2.43 logit and lowest at 2.16. Mean for M=7.02/3=2.34.
Where else, the oral and graphic presentation is mediocre, Mean for O= -1.37/3= -0.46
Design development is the easiest with the lowest mean value, Mean for D=-5.64/8=-0.71. Raw variance explained shown in figure 2, by measures is at 65.26%
higher than 40% gives an indication that the instrument is measuring one dimension (uni-
dimensionality) or in other words, measuring what is supposed to measure.
Figure 2. Measurable Data Summary.
To monitor tutors’ performance, statistical analysis has been done and the results are
revealed in figures 3 and 4. Figure 4 shows that teacher3 is the strictest assessor among
the other teachers since Teacher3 logit measure is at 3.17 logit.
And the next logit measure is only at 1.39 logit, a difference of 1.78 logit. Teacher4 is
the most lenient among them all with logit at 0.96 measures.
However further scrutiny reveals that Teacher4 has Z-Std or high negative residual.
Teacher4 under-rate easy criteria, and over-rate on difficult criteria.
Figure 3. Tutors’ Measurement Report
Nangkula Utaberta et al. / Prosiding - Seminar Pendidikan Kejuruteraan dan Alam Bina 2012
256
Teacher one rated student twelve, on item D1 with a ‘2’ but in actual teacher one is
expected to give student twelve a ‘3.7’ or ‘4’. Similarly, teacher one rated student18 on
item D6 a ‘4’ however teacher1 is expected to rate student18 a ‘1.9’ or ‘2’.
Teacher2 rated student1 on item M1 a ‘1’, but teacher two is expected (calculated by
Rasch) to rate student1 a ‘2.7’ or ‘3’.
Figure 4. Unexpected Responses
Figure 5. All Facet vertical Rulers
Nangkula Utaberta et al. / Prosiding - Seminar Pendidikan Kejuruteraan dan Alam Bina 2012
257
Figure 6. Probabilities of Grades
Figure 5 shows the results of all facet vertical rulers. The first column includes
measurement values which have calculated by the software. The next columns include
teachers, students, criteria and scales, respectively. Distributions of students in getting
marks in different criteria by different teachers are shown by the star icons.
It is observable that the measurement values of the teacher 1, teacher 2 and teacher 4
are close to each other while the measurement value of the teacher 3 is more than other
teachers.
Giving criteria in brief would have an important influence on the form of questioning,
but in tutors’ and jurors’ responses their first consideration was the strength and weakness
of the students’ design. The tutors and members of jury have their own preference and
priorities, and would tend to bias their questioning to accommodate these personal
interests.
Obtained data and analysis have shown that defined criteria for that specific sample
does fit properly and as Figure 6 reveals in any probabilty and any students’ measurement
all of marks could be in the range.
Based on this curve, we can expect that for even large sample of students and teachers
the criteria would work and easily we can predict or expect marks. In the tutor performance Rasch also confirms the concerns. As Figure 7 shows teacher
three with minimum experience and degree among other tutors has more unexpected
marks and teacher number two which was the studio master with degree and experience
has no unexpected marks and judgment.
Figure 7. Most Unexpected Responses
Nangkula Utaberta et al. / Prosiding - Seminar Pendidikan Kejuruteraan dan Alam Bina 2012
258
All these make the idea of tutor training, having more discussion session among
invited jurors before holding submission session and explain the jurors about the
objectives and expectations. Teacher4 under-rate easy criteria, and over-rate on difficult
criteria. Maybe the distance of tutors from practice world with ageing profile of untrained
instructors may cause the problem in evaluation and assessment process.
In proposed assessment model in this research the existence of objectives and different
portion of importance for each and informing this to students helped students to have
their best probability of mark for each task. Rasch also proved that each student obtained
his maximum rank of mark due to his effort and ability.
To make this understandable the average of three submission sessions which were held
on before the specific session which measured by Rasch were compared to the obtained
grade on the chosen session.
It is notable that in all three sessions the evaluation sheet were used but students were
not aware of details and also jurors used the evaluation sheet but the form of marking
were holistic, they just given the percentage for each of criteria-Based on obtained result
by Rasch which discussed in chapter four, student number 16 is the most performed
student and the least performed student is student 20.
The investigation as figure 8 reveals shows that student number four and number
twenty has maximum progress and student number 7 and 18 have regressed. But the
reason their regress seems to be lack of time for preparation in submission day. In their
evaluation sheet for instance jurors have mentioned that the unfinished 3D model or
absence of one or two elevations.
In most performed students group also we have some small progress, especially for
student number 12 and 16. In total, summations of progresses were more than regresses.
The summations of marks are the evidence of this claim. Summations of marks from
1204 unit go to 1261 unit. This means 4.73 percent progress. We chose 6 students. Six
because of the separation number given by Rasch and then would do the investigation on
the process of progress and the given marks in interim submissions.
Figure 8. Improvement of marks after using proposed model
Nangkula Utaberta et al. / Prosiding - Seminar Pendidikan Kejuruteraan dan Alam Bina 2012
259
3. Recommendations and Conclusion
The results from Rasch shown that criteria based assessment model and defined
criteria for course and assignment objectives worked well and was designed properly.
About tutor performance, Rasch shown that experience has an important role that can
affect the judgment. In summary, Rasch also confirmed the necessity of implementing
some recommendation like holding workshops for tutors, clarification in objectives and
criteria for teachers and students.
In fact, inviting part time teachers with practice base can help to have more effective
evaluation and assessment process but in the case they do not see their teaching as an
escape from the constraint of office life that d like to follow the strategic approach to
embedding in professional skills.
All these make the idea of tutor training, having more discussion session among
invited jurors before holding submission session and explain the jurors about the
objectives and expectations. In case of time, the most spent time goes to students’
presentation and giving comments which would be same in all architecture assessment
models. For grading time why there is no need for tutors to calculate the total mark or
transfer the chosen level of satisfaction between fail to excellent, this will decrease the
consuming time and leads to dedicate maximum time to measure the satisfaction level of
each task and criteria.
Rasch also can help to transfer qualitative evaluation like our case to accurate
interpretation of numbers. The results of this research imply that students are learning
from the tool of criticism in the design studio. The expectations are merely components
of a design studio, which allow for crit sessions to be an effective didactic situation.
References
Hassanpour, B., Utaberta, N. & Zaharim, A. (2012). Critical Analysis of Criteria-Based Assessment and
Grading in Architecture Education (Universiti Kebangsaan Malaysia as Case Study). Research Journal of
Applied Sciences, Engineering and Technology 4(14), 2046-2054.
Salama, A.M., & Wilkinson, N. (2007). Design Studio Pedagogy: Horizons for the Future. In A. M. Salama and
N. Wilkinson (Eds.). Gateshead, United Kingdom.: The Urban International Press.
Sadler, D.R , (2005). Interpretations of criteria-based assessment and grading in higher education. Assessment
and education in higher education 30(2), 175-193.
Masodi, S. & Gulman. (2008). Application of Rasch-based ESPEGS model in measuring generic skills of
engineering students. WSEAS TRANSACTIONS on Advances in Engineering Education 8(5), 591-602.
Sadler, D.R , 2005. Interpretations of criteria-based assessment and grading in higher education. Assessment and
education in higher education 30(2), 175-193.
Sudin, W.P. (2003). Architecture education in Malaysia: an historical overview. Architecture forum. Kuala
Lumpur. Malaysia.
.