Implicating the integrated format on reading test assessment: An evaluation of relevant factors

The present study in general evaluates relevant factors involved in assessing reading

performance in the context of English classroom teaching. Unlike the traditional way of placing the

reading text and test items separately in the split format, the test items were placed in accordance with

relevant parts of the reading text in the integrated format and this study compares reading performance

on the ground of the following variables: test formats, study subject, learning task, and pre-proficiency

level. Findings from the study firstly indicated the influence of test formats on reading performance

from the evidence that participants in the integrated format performed better than those in the split

format. Second, findings also showed significant effect of the interaction between test formats and

learning task as well as the interaction between pre-proficiency level and study subject in reading test

performance. Third, in combination with test formats, task design also had an influence on reading test

performance.

Implicating the integrated format on reading test assessment: An evaluation of relevant factors trang 1

Trang 1

Implicating the integrated format on reading test assessment: An evaluation of relevant factors trang 2

Trang 2

Implicating the integrated format on reading test assessment: An evaluation of relevant factors trang 3

Trang 3

Implicating the integrated format on reading test assessment: An evaluation of relevant factors trang 4

Trang 4

Implicating the integrated format on reading test assessment: An evaluation of relevant factors trang 5

Trang 5

Implicating the integrated format on reading test assessment: An evaluation of relevant factors trang 6

Trang 6

Implicating the integrated format on reading test assessment: An evaluation of relevant factors trang 7

Trang 7

Implicating the integrated format on reading test assessment: An evaluation of relevant factors trang 8

Trang 8

Implicating the integrated format on reading test assessment: An evaluation of relevant factors trang 9

Trang 9

pdf 9 trang xuanhieu 1620
Bạn đang xem tài liệu "Implicating the integrated format on reading test assessment: An evaluation of relevant factors", để tải tài liệu gốc về máy hãy click vào nút Download ở trên

Tóm tắt nội dung tài liệu: Implicating the integrated format on reading test assessment: An evaluation of relevant factors

Implicating the integrated format on reading test assessment: An evaluation of relevant factors
s of test performance under the division of factors with 
significant effects from ANOVA analyses. 
4. Results 
4.1. Descriptive statistics of test score performance 
 The following descriptive statistics firstly summarizes the number of participants allocated into test 
formats and study subjects, to be followed by the mean scores and its corresponding SD. The mean scores 
between two test formats overall indicate that participants in the integrated format scored higher than those 
in the split format: 
Table 3. Descriptive statistics of test score performance 
Test format x Study subjects 
civil read1 read2 read4 
0(split) 13 9 13 10 
1(integrated) 11 8 9 11 
Mean and SD 
 Test format Subject Mean SD 
1 0 Civil 5.53 1.19 
2 1 Civil 7.63 1.28 
3 0 read1 5.88 1.36 
4 1 read1 7.12 1.35 
5 0 read2 6.15 1.57 
6 1 read2 6.66 1.11 
7 0 read4 5.90 1.79 
8 1 read4 6.09 2.02 
4.2. Analyzing the combined effects test formats, study subject, and learning task on test score 
performance 
A three-way between subjects 2x4x2 ANOVA analysis was conducted to measure the effects of test 
formats (integrated and split formats; var. name: ver), study subjects (civil, read 1, read 2, read 4; var. name: 
class_numeric), and learning task (with and without recording exercise; var. name: condition) on test score 
performance. There was a significant effect of test formats on test score performance at the 0.01 level (F 
(1, 76)=9. 37, p=0.003). There was also an interaction between test formats and learning task on test score 
performance and this interaction is statistically significant at the 0.05 level (F (1, 76)=4.29, p=0.04). 
Table 4. Analysis of Variance for test formats, study subject, learning task, and test score performance 
Response: mydata$test_score 
 Df Sum Sq Mean Sq F value Pr(>F) 
mydata$ver 1 21.108 21.1077 9.37890.0030 ** 
mydata$class_numeric 1 0.047 0.0473 0.0210 0.885074 
mydata$condition 1 2.037 2.0373 0.9052 0.344399 
mydata$ver:mydata$class_numeric 1 2.203 2.2034 0.9790 0.325574 
mydata$ver:mydata$condition 1 9.671 9.6708 4.29710.041568* 
mydata$class_numeric:mydata$condition 1 1.683 1.6831 0.7479 0.389875 
mydata$ver:mydata$class_numeric:mydata$condition 1 0.876 0.8756 0.3891 0.534665 
Residuals 76 171.042 2.2505 
Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1 
4.3. Analyzing the combined effects of pre-proficiency level, test formats, study subjects, and learning 
task on test score performance 
The follow-up analysis on the combined effects of four factors on test score performance (var. name: 
test_score) of three groups of participants from Reading 2 (read 2), Reading 4 (read4), and British-American 
Civilization (civil) and was divided into two models. The first model involves the combined effects of pre-
proficiency level (in the form of pre-test scores; var. name: pre_score), learning task (with and without 
recording exercise; var. name: condition), and test formats (integrated and split formats; var. name: ver). 
A three-way ANOVA analysis was conducted to measure the effects of three factors on test score 
performance in the first model. Results of the test indicated no significant effects of pre-proficiency level 
(F(1,59)=0.73, p>0.05) and no combined effects in the interaction of pre-proficiency level with learning 
task (F(1,59)=0.34, p>0.05) as well as with test formats (F(1, 59)=0.02, p>0.05) at the 0.05 level. Apart 
from the significant effect of test formats on test score performance, the interaction between learning task 
and test formats was also significant at the 0.05 level (F(1,59)=4.94, p=0.02). 
Table 5. Analysis of Variance Table of Model 1 
Response: mydata$test_score 
 Df Sum Sq Mean Sq F value Pr(>F) 
mydata$pre_score 1 1.806 1.8060 0.7397 0.39322 
mydata$condition 1 1.003 1.0033 0.4110 0.52396 
mydata$ver 1 13.935 13.9348 5.7078 0.02011 * 
mydata$pre_score:mydata$condition 1 0.854 0.8538 0.3497 0.55654 
mydata$pre_score:mydata$ver 1 0.050 0.0504 0.0206 0.88627 
mydata$condition:mydata$ver 1 12.082 12.0820 4.9488 0.02995 * 
mydata$pre_score:mydata$condition:mydata$ver 1 0.258 0.2582 0.1058 0.74617 
Residuals 59 144.041 2.4414 
Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1 
The second model involves the combined effects of pre-proficiency level (in the form of pre-test 
scores; var. name: pre_score), learning task (with and without recording exercise; var_name: condition), 
and study subjects (civil, read 2, read 4; var. name: class_numeric). The second three-way ANOVA 
indicated that in the second model the effect of pre-proficiency level on test score performances was not 
significant (F(1,61)=0.69, p>0.05) at the 0.05 level. Meanwhile, the interaction of pre-proficiency level 
with study subjects was significant at 0.05 level (F(1,61)=4.29, p=0.04). 
Table 6. Analysis of Variance Table of Model 2 
Response: mydata$test_score 
Df Sum Sq Mean Sq F value Pr(>F) 
mydata$pre_score 1 1.806 1.8060 0.6984 0.40658 
mydata$condition 1 1.003 1.0033 0.3880 0.53568 
mydata$class_numeric 1 1.172 1.1717 0.4531 0.50341 
mydata$pre_score:mydata$condition 1 1.204 1.2041 0.4656 0.49759 
mydata$pre_score:mydata$class_numeric 1 11.105 11.1047 4.2943 0.04247 * 
Residuals 61 157.740 2.5859 
Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1 
4.4. Analyzing the single effect of learning task and of test formats on task performance 
 Previous findings showed significant effect in the interaction between pre-condition and test formats 
on test score performances. First, the analysis of task performance based on the single effect of learning 
task was conducted by means of comparing the differences in score performance of task between two 
conditions: with recording exercise (experimental; n=41; class_numeric: read 1 and civil) and without 
recording (control; n=43; class_numeric: read 2 and read 4). The task performance was analyzed using 
Welch two-sample t-test in two modes: single and dual task performance. Results of Welch two-sample t-
test in the first analysis indicated no significant effect of learning task on task performance at the 0.05 level 
even though participants in the experimental group scored higher than those in the control group (except 
for the performance of task 1 and in the pair of task 1 3). 
Table 7. The single effect of learning task on task performance 
 Mcontrol M experimental t df p-value 
Task 1 2.13 2.09 0.206 79.113 0.83 
Task 2 1.697 1.878 -0.7675 81. 871 0.445 
Task 3 2.348 2.512 -0.85 81.04 0.397 
Task 1 2 3.837 3.975 -0.436 81. 49 0.663 
Task 1 3 4.48 4.39 0.38 81.7 0.70 
Task 2 3 4.04 4.60 -1.91 81.802 0.059 
Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1 
 Second, the analysis of task performance based on the single effect of test formats was conducted by 
means of comparing the differences in score performance of task between two formats: split (n=45) and 
integrated (n=39). On similar analysis using Welch two-sample t-test, participants in the integrated format 
scored higher than those in the split format. However, significant effect of test formats on task performance 
was found in the performance of task 1 (t=-2.51, df=77.9, p=0.01), the performance of dual task 1 2 (t=-
2.94, df=76.99, p=0,004), and the performance of dual task 1 3 (t=-2.63, df=75.89, p=0.01) at the 0.05 level. 
Table 8. The single effect of test formats on task performance 
 M 
Split 
M 
integrated 
t df p-value 
Task 1 1.88 2.38 -2.51 77.9 0.01* 
Task 2 1.6 2.0 -1.72 80.77 0.08 
Task 3 2.37 2.48 -0.559 76.597 0. 577 
Task 1 2 3.48 4.38 -2.94 76.99 0.004* 
Task 1 3 4.13 4.79 -2.638 75.89 0.01* 
Task 2 3 4.11 4.56 -1.52 81.072 0. 13 
Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1 
5. Discussion and implications 
The purpose of the present study is to investigate the effects of test formats, study subjects, learning 
task, and pre-proficiency level on test score performance in the evaluation of combined effects and single 
effect. Findings of the study firstly revealed that test formats have a significant effect on test score 
performances, considering that participants in the integrated format performed better than those in the split 
format. Despite being different in research scope, while investigating the influence of material design on 
geometry test performance, Tindall-Ford et al. (2015) similarly found that participants in the integrated 
format earned a higher mean score than those in the split format. In a similar concern with their study, it 
could be the employment of self-management strategy by learners as participants while making their efforts 
in reducing the split-attention effect; in addition, participants in the present study may also understand how 
to respond the test items by integrating the questions with the relevant text. 
Furthermore, the present study also showed that there was a significant effect in the interaction 
between test formats and learning task, and this finding can be linked with the role of worked examples in 
lowering cognitive load for the purpose of relating isolated information from the reading material. In 
another study on the scope of material designing for an introductory accounting course, Sithole (2017) also 
indicated participants showed a better performance in the integrated format for the recall and transfer tests 
and suggested the usage of integrated textbook as an effective worked example for learning. 
Regarding the usage of recording task prior to the participation of the reading performance test for 
the experimental group, the present study refers to Pouw et al.’s (2019) study on the cognitive basis for 
split-attention effects to suggest the implementation of recording exercise as the means to lower the 
extraneous cognitive load for the purpose of maintaining and searching information in working memory. 
In other words, the recording exercise as the learning task in this study may reduce the amount of time for 
memory traces while learners perform the reading task in the assessment phase in classroom. 
In the present study, the evaluation of single effect for learning task in the treatment of recording 
exercise presented that there was no significant difference in the mean score between split and integrated 
formats. In correspondence with task requirements from the design of MCQs in split and integrated formats, 
this finding further raises practical concerns about how individual differences differentiate their reading 
performance in MCQs, considering the impact of MCQs task in magnifying reading self-efficacy (Solheim, 
2011). 
6. Conclusion 
 The following study is conducted to analyze relevant factors which are likely to affect reading test 
performance in the context of language classroom. Under the main involvement of the split and integrated 
test formats, findings from the study generally support the reduction of split-attention effect from the 
implication of integrating relevant texts with their relevant questions in the reading task. The study also 
sheds lights on contributors to reading test performance, considering that these contributors are likely to 
promote and demote the process of reading comprehension. 
 There are some suggestions for future research investigating similar scope. In manipulating the 
condition which can reduce the split-attention effects in reading assessment in the context of language 
classroom, future studies can refer to the integration of hypertext glosses as the technological aids of 
explaining word meanings in reading assessment (Chen, 2014). The assessment of reading comprehension 
in different formats can also be adapted to longitudinal design of extensive reading in the evaluation of its 
effectiveness in the development of reading proficiency and related variables (Jeon & Day, 2016) 
References 
Chandler, P., & Sweller, J. (1991). Cognitive load theory and the format of instruction. Cognition and Instruction, 
8(4), 293-332. 
Chen, I.J. (2014). Hypertext glosses for foreign language reading comprehension and vocabulary acquisition: 
Effects of assessment methods. Computer Assisted Language Learning, 29(2), 413-426. 
Huynh, C.M.H. (2015). Split-attention in reading comprehension: A case of English as a foreign/second language. 
6th International Conference on TESOL (pp. 1-12). SEAMEO. Ho Chi Minh City, Vietnam. 
Jeon, E.Y., & Day, R.R. (2016). The effectiveness of ER on reading proficiency: A meta-analysis. Reading in a 
Foreign Language, 28(2), 246-265. 
Paas, F., Renkl, A., & Sweller, J. (2004). Cognitive load theory: Instructional implications of the interaction 
between information structures and cognitive architecture. Instructional Science, 32(1), 1-8. 
Pouw, W., Rop, G., De Koning, B., & Paas, F. (2019). The cognitive basis for the split-attention effect. Journal of 
Experimental Psychology, 148(11), 2058-2075. 
Schnotz, W., & Kürschner, C. (2007). A reconsideration of cognitive load theory. Educational Psychology 
Review, 19(4), 469-508. 
Sithole, S.T. (2017). Enhancing students understanding of introductory accounting by integrating split-attention 
instructional material. Accounting Research Journal, 30(3), 283-300. 
Solheim, O.J. (2011). The impact of reading self-efficacy and task value on reading comprehension scores in 
different item formats. Reading Psychology, 32(1), 1-27. 
Sweller, J. (1994). Cognitive load theory, learning difficulty, and instructional design. Learning and Instruction, 
4(4), 295-312. 
Sweller, J., Ayres, P., & Kalyuga, S. (2011). Cognitive load theory. New York, NY: Springer. 
Tindall-Ford, S., Agostinho, S., Bokosmaty, S., Paas, F., & Chandler, P. (2015). Computer-based learning of 
geometry from integrated and split-attention worked examples: The power of self-management. Journal of 
Educational Technology & Society, 18(4), 89-99. 
NGHIÊN CỨU CÁC YẾU TỐ TRONG THỰC NGHIỆM 
DẠNG ĐỀ TÍCH HỢP VÀO ĐÁNH GIÁ ĐỌC HIỂU TIẾNG ANH 
Tóm tắt: Nghiên cứu thực nghiệm này được thực hiện nhằm định trị các nhân tố trong việc đánh giá kỹ 
năng đọc hiểu tiếng Anh ở môi trường lớp học. Dựa trên sự so sánh kết quả đánh giá đọc hiểu giữa hai 
dạng đề: dạng đề không phân vùng (split format) và dạng đề tích hợp (integrated format), nghiên cứu 
phân tích các nhân tố bao gồm dạng đề, môn học liên quan, bài tập tìm hiểu,và trình độ đọc hiểu. Kết 
quả thực nghiệm cho thấy người tham gia làm dạng đề tích hợp có kết quả cao hơn dạng đề không phân 
vùng. Bên cạnh đó, có mối liên hệ trong sự tương tác giữa dạng đề và bài tập tìm hiểu cũng như là giữa 
trình độ đọc hiểu và môn học liên quan. Nghiên cứu cũng chỉ ra cùng với dạng đề, việc thiết kế dạng 
bài đánh giá cũng có ảnh hưởng tới kết quả đánh giá. 
Từ khóa: Split-attention, integrated format, reading assessment, reading performance 

File đính kèm:

  • pdfimplicating_the_integrated_format_on_reading_test_assessment.pdf