Tool Usage and Efficiency in an Online Test
Gyanesh Jain
Playpower Labs, India
gyanesh.jain@playpowerlabs.com
Aditya Sharma
Playpower Labs, India
aditya@playpowerlabs.com
Nirmal Patel
Playpower Labs, India
nirmal@playpowerlabs.com
Amit.A.Nanavati
Ahmedabad University, India
amit.nanavati@ahduni.edu.in

ABSTRACT

In this study, we analyze data from the National Assessment of Education Progress (NAEP) digital test to understand how digital tool usage relates to the efficiency of answering questions. Digital testing software provides students with on-screen tools such as calculators and scratchpads. We found that students who used digital tools in NAEP were slower in solving the problems but more accurate when the question demanded using the tool. We also found that when students used the tool when it was not needed, they were more likely to be incorrect. Overall, our findings suggest that students need to be trained on how to use the tools and when to use them to make the most use of their testing time.

Keywords

Digital Assessments, Process Data, Student Behavior

1. INTRODUCTION AND PRIOR WORK

The study focuses on the relationship between on-screen tool usage and the efficiency of students. The software used for online testing typically provides tools such as scratchpads and calculators to the students to help them think, and solve problems. Students’ usage patterns of these tools can help us model their behavior and help instructors to help students identify tool usage. Data from the NAEP (National Assessment of Educational Progress) used by the US Govt, was used. Computer-based tests are now standard in large-scale assessments.

To ensure the digital competency of the test taker [14] students are subjected to lab-based studies where they are recorded while interacting with the assessment interface [2]. NAEP test provides some tools for the students to aid in problem-solving. Students are provided with quick training before the test starts. Two key tools in the NAEP test are Calculator and Scratchpad. All students in the NAEP test have access to a physical calculator (they are given one if they don’t bring their own), and students can also request pencils and scratch paper if needed. The NAEP UI (shown in Figure 1) also has a digital Calculator and a digital Scratchpad that substitute for their physical versions.

User interface of the NAEP digital test.
Figure 1: User interface of the NAEP digital test.

Efficient test-takers show distinct digital patterns in the log data. Sahin used Latent Profile Analysis to look at how students allocated times to different test items in the NAEP [5]. They discovered four distinct groups in their sample that they described as 1) little time on the first (problem) visit, 2) balanced time (across problem visits and revisits), 3) little revisit with more time in the end, and 4) little revisit with less time in the end. The researchers found that these four groups differed in the average outcomes, with group 4 scoring the lowest (pg. 21, ibid). Another recent study [3] showed that in the NAEP test, students were more likely to use the digital calculator on calculation-heavy items. When it was used in an ideal way, students were also likely to score higher.

We analyzed the process data from the 2018 NAEP test and compared the students who did not use digital on-screen tools with the ones who did. All students in the NAEP test had access to physical tools, so we wanted to understand student preferences in using digital tools. In our analysis, we compared the time taken by students who used digital tools with those who did not.

2. DATA

We used a random sample from the 2018 NAEP Mathematics test for Grade 8. Our sample had data N = 1642 students. The digital Math assessment consists of two 30-minute blocks, and our sample had data from the first block. There were a total of 20 questions in the first block. The entire test-taking process of the students was captured by collecting data points for each interaction event. Each student interaction in the digital assessment system resulted in one observation in the dataset. Each observation had seven different variables. They are listed in Table 1 below.

Table 1: Columns of the NAEP Process Data.
Variable Description
STUDENTID Unique identifier of the student
Block Block of the NAEP test, A or B
AccessionNumber Unique identifier of the question that the student is attempting
ItemType Type of the question e.g. MCQ, Fill in the Blank, etc.
Observable The name of the action that student took e.g. clicking, dragging, scrolling, typing, opening a calculator
ExtendedInfo Metadata of the student action
EventTime Time when the student interaction occurred

There were forty-two unique types of actions (Observable column in the data). These actions were further coded by us into six different categories: Answer (responding to the item), Navigation (switching between items), Timer (looking at the remaining time), Calculator (using the digital calculator), Scratchpad (using the digital scratchpad), Equation Editor (using the equation editor), and Readability (adjusting the readability of the on-screen text). We considered Calculator, Scratchpad, and Equation Editor as Digital Tools. We calculated the time students spent doing each type of action, and the combined time students spent using Digital Tools on a question, used to categorize them into tool users and non-tool users. The non-tool users had no tool used for a given question, while the tool users had one or more events related to the digital tools. Once the students were categorized for their tool usage for each question, we compared the two groups for each question, in how quickly they responded to the question items.

We had accuracy data available for eight multiple-choice questions, calculated by looking at the latest option clicked by the students and comparing it with the answer key. Further, the accuracy data was used to compare students who did and did not use the digital tools.

3. RESEARCH QUESTIONS AND METHOD

Our objective was to understand how students who used the on-screen tools differed from those who did not. We wanted to know whether students who used the digital tools were faster at responding to the test items. We also wanted to understand whether the tool usage behavior was different across items and if there were any differences in the proficiency of the students using the tools versus not using the tools.

For RQ1, we used the t-Test to compare the item response times of the groups of students who used the digital tools and those who did not. For RQ2, for each question, we compared the number of students who used digital tools for answering it with those who did not use the tools. Given that we had a random sample from the test, we expected that the student preference seen in our data would generalize to the target population. To answer RQ3, we used correctness data from the multiple choice questions of the test and calculated what proportion of the students using the tools were getting the items correct.

Comparison of item response times for students who used the digital tools and who did not use the digital tools
Figure 2: Comparison of item response times for students who used the digital tools and who did not use the digital tools. The answer key was only available for multiple-choice questions.

4. RESULTS

RQ1: We found that, on average, students who used the digital tools took more time to respond to the items than those who did not use the tools. Figure 2 summarizes our findings for each question. We can see that for some questions, the difference in the response time between non-tool and tool users is as much as double. We performed mean comparisons for items where we had correctness data for the correct, incorrect, and not attempted results. Appendix A contains the results of the t-Tests, where we see that most of the differences were statistically significant. This answers RQ1 and tells us that students who used the on-screen tools were slower than the ones who did not use the on-screen tools. We do not have precise data about what the non-tool users used. Maybe they did some guesswork (given the multiple-choice questions), or they used physical tools. Based on the content of the questions (provided in Appendix C), we can say that mental math may not be sufficient to solve most of the questions.

Table 2: Number of correct and incorrect students by their digital tool usage. We can see that for some questions, the total number of students who used the tools is higher than the ones who did not use the digital tools. We can also see that tool users scored more on average for some questions. In questions where tools were not required, the tool users scored less.
Question Tool Usage Correct Incorrect Prop.
(N) (N) Correct
VH098519 ToolsNotUsed 246 462 65.25
VH098519 ToolsUsed 251 665 72.60
VH098753 ToolsNotUsed 221 29 11.60
VH098753 ToolsUsed 1000 328 24.70
VH098759 ToolsNotUsed 261 18 6.45
VH098759 ToolsUsed 741 604 44.91
VH098783 ToolsNotUsed 130 161 55.33
VH098783 ToolsUsed 259 1049 80.20
VH098808 ToolsNotUsed 146 173 54.23
VH098808 ToolsUsed 653 657 50.15
VH098810 ToolsNotUsed 697 876 55.69
VH098810 ToolsUsed 39 30 43.48
VH098812 ToolsNotUsed 483 318 39.70
VH098812 ToolsUsed 461 324 41.27
VH098839 ToolsNotUsed 322 79 19.70
VH098839 ToolsUsed 637 438 40.74

RQ2: We found that students preferred to use tools in some questions and not use them in others. This is consistent with findings from [3], where they observed that calculator use was more in calculation-heavy items. Question 810, a recall question (shown in Appendix C), did not need a calculator at all, and we can see that out of the 1642 students, 1573 (95.8%) did not use any tool while answering this question. For Questions 753, 759, 783, and 808, more than 80% of the students preferred to use the tools (whether correct or incorrect). For Questions 812 and 519, the tool users were 49.5% and 56.4%, telling us that students did not clearly prefer to use tools for these questions. Looking at Appendix C, we can see that Question 812 may not require tool usage. In summary, Table 2 answers RQ2 and shows us that in some questions, tools were preferred, in some, they were not, and in others, there was mixed behavior.

RQ3: Overall, we found that the tool-using group of students scored more on average than the non-tool-using group. We can see in Table 2, students typically scored more in questions where tools were used. Appendix B shows where the students scored significantly more when using digital tools. The biggest difference was seen in Question 759, a question on calculating averages. Here, 44.9% of the tool users answered correctly, whereas only 6.5% of the non-tool users answered correctly. We do not know why so many students who did not use the tools got the question incorrect because, as per NAEP policy, they all had access to physical calculators. It may be possible that students’ digital tool use is an indicator of their other abilities. We know that students who answered without using the digital tools answered the question faster, though some of those responses could be guesses.

5. DISCUSSION

Based on the results, we can see that the students who did not use the tools provided in the NAEP digital interface were faster in answering the question - whether correct or incorrect. Since we did not have data on students’ outside activities, we cannot say anything about why the students not using the digital tools were faster. We also tried to sequence modelling on students, but the patterns were too complex for the scope of this poster (Appendix C).

For certain Questions tool usage was not required (based on the content of the question). These questions could be solved without tools, and it was seen that the proportion of correct and incorrect responses for non-tool users is statistically insignificant (Appendix B). For certain Questions which required calculation, we still found that the difference between correct and incorrect non-tool users was insignificant, which is interesting and should be investigated further. In all questions, students have to decide whether to use the tool, implying that student training needs to be conducted on not just how to use the tool but also when to use the tool. As the difference between incorrect and correct proportions when not using a tool is positive, it implies that many students were either hesitant to use the tools, inept at using them, or found the tools lacking.

Our analysis did not utilize the fine-grained process data provided by the NAEP system. The step-by-step data of student actions can show how they utilized the tools. To help students be more efficient while taking the test, we can provide personalized feedback based on their usage patterns. For the students who answered correctly, we can find tool usage patterns that were more efficient than theirs (if available) and provide the closest and fastest patterns as suggestions. This can guide the students in avoiding unnecessary steps while solving the problem and making the most of their time. It may be worthwhile to consider having practice tests where the digital test-taking interface disables access to tools for the questions when they are unnecessary. Appropriate tool usage can help students save time and have fewer digital distractions during the test.

6. CONCLUSION AND FUTURE WORK

Our study found that students taking the NAEP test differed in their on-screen tool usage behavior. Students who used the digital tools were typically slower in responding to the items than those who did not. If the question item demanded tool use, then students preferred to use the tools and also scored higher when they used the tools. Our findings show that when taking digital tests, students are better off if they know when to use the tools and when not to use them. A future study can analyze the nuanced processes of tool usage and compare efficient and inefficient digital tool usage. The process data can provide students with personalized recommendations on how to use the tools more efficiently and save time while taking the test. We could also look at the sequence of tool usage and non tool usage amongst students, attempting NAEP.

7. REFERENCES

  1. Antonio Calvani, Antonio Fini, Maria Ranieri, et al. Digital competence in k-12: theoretical models, assessment tools and empirical research. Anàlisi: quaderns de comunicació i cultura, pages 157–171, 2010.
  2. Richard P Duran, Ting Zhang, David Sanosa, and Fran Stancavage. Effects of visual representations and associated interactive features on student performance on national assessment of educational progress (naep) pilot science scenario-based tasks. American Institutes for Research, 2020.
  3. Yang Jiang, Gabrielle A Cayton-Hodges, Leslie Nabors Oláh, and Ilona Minchuk. Using sequence mining to study students’ calculator use, problem solving, and mathematics achievement in the national assessment of educational progress (naep). Computers & Education, 193:104680, 2023.
  4. Fanny Pettersson. On the issues of digital competence in educational contexts–a review of literature. Education and information technologies, 23(3):1005–1021, 2018.
  5. Füsun Şahin. Exploring the relations between students’ time management strategies and test performance. Paper presented at the Annual meeting of the National Council for Measurement in Education, 2019.

APPENDIX

A. RESPONSE TIME COMPARISON

A.1 For Correct Answers (T-value calculated by Time consumed by tool not used- tool not used)

Table 3: Response Time Comparison for Correct Answers
AccessionNumber VH098519 VH098753 VH098759 VH098783 VH098808 VH098810 VH098812 VH098839
T-Value -11.759 -6.733 -4.811 -9.962 -12.677 -4.275 -4.114 -6.251
P-value <2.2e-16 <2.2e-16 <2.2e-16 <2.2e-16 <2.2e-16 <2.2e-16 <2.2e-16 <2.2e-16

A.2 For Incorrect Answers (T-value calculated by Time consumed by tool not used- tool not used)

Table 4: Response Time Comparison for Incorrect Answers
AccessionNumber VH0985\19 VH098753 VH098759 VH098783 VH098808 VH098810 VH098812 VH098839
T-Value -10.689 -14.246 -15.657 -6.507 -12.565 -3.711 -6.49 -9.497
P-value <2.2e-16 <2.2e-16 <2.2e-16 <2.2e-16 <2.2e-16 <2.2e-16 <2.2e-16 <2.2e-16

A.3 For Unattempted Answers (T-value calculated by Time consumed by tool not used- tool not used)

Table 5: Time Spent on Unattemped Answers
AccessionNumber VH098519 VH098753 VH098759 VH098783 VH098808 VH098812 VH098839
T-Value -3.295 -3.423 -3.057 -3.802 -2.49 -2.965 -1.514
P-value 0.0046 0.0012 0.0121 0.0012 0.0472 0.0069 0.1355

B. CORRECTNESS PROPORTION FOR NON-TOOL USERS

Table 6: Proportion of Correct answers for non-tool users
Question No. Correct When Tool Not Used Incorrect When Tool Not Used Difference Count Correct Count Incorrect Proportion z-test p-value Sig
VH098519 0.410 0.495 0.085 1127 497 0.436 3.185 0.0025 Yes
VH098753 0.081 0.181 0.100 357 1221 0.158 4.541 0.0000 Yes
VH098759 0.029 0.260 0.232 622 1002 0.172 12.025 0.0000 Yes
VH098783 0.133 0.334 0.201 1210 389 0.182 8.944 0.0000 Yes
VH098808 0.208 0.183 -0.026 830 799 0.196 -1.307 0.1698 No
VH098810 0.967 0.947 -0.020 906 736 0.958 -1.996 0.0544 No
VH098812 0.495 0.512 0.016 642 944 0.505 0.638 0.3254 No
VH098839 0.153 0.336 0.183 517 959 0.272 7.538 <2.2e-16 Yes

C. SEQUENCE MODELLING

C.1 Where Calculator was used

Process Flow where Calculator was used
Figure 3: Sequence where calculator was used

C.2 Where no tool was used

Process Flow where no tool was used
Figure 4: Where no tool was used

D. LINK FOR THE NAEP TEST

QR Code for the NAEP Test
Figure 5: Scan to go to NAEP Test

© 2023 Copyright is held by the author(s). This work is distributed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0) license.