Professional Documents
Culture Documents
507
Authorized licensed use limited to: UNIVERSITY OF SHARJAH. Downloaded on September 22,2023 at 17:32:45 UTC from IEEE Xplore. Restrictions apply.
2023 8th International Conference on Business and Industrial Research 18 - 19 May 2023
(ICBIR2023) Bangkok, Thailand
Here VR medical education content with a gaze should focus on when viewing the medical education
visualization interface was developed using eye- content based on the reviewer feedback, and are listed
tracking technology to prevent the cognitive overload in Table 1.
issues that may occur in existing A research
hypothesis was derived that there would be TABLE I. AOI CLASSIFICATION
differences in learning performance depending on the AOI Number AOI Name
type of gaze visualization interface. Two gaze
visualization interface types, Circle Marker and AOI 1 Patient Band
Highlight, were applied to the developed VR medical AOI 2 Confirmation Form of Operation
education content. To prove the research hypothesis,
an experiment was conducted to analyze the AOI 3 Surgical Site
difference in learning performance between the two AOI 4 Anesthesia Machine
gaze visualization interface types. A mixture of
AOI 5 Anesthesia
quantitative and qualitative data was used for
effective evaluation and analysis of learning AOI 6 Stethoscope
performance. Therefore, based on previous studies, AOI 7 Defibrillator
learning performance indicators were derived for
learning concentration evaluated with eye tracking AOI 8 Monitor
data and learning flow evaluated with surveys. AOI 9 Surgical Count
Learning concentration was evaluated as AOI 10 Specimen
quantitative data using dwell time among eye tracking
metrics [25]. In this study, it was evaluated that the
higher the dwell time, the better the concentration of An instructor’s gaze visualization interface was
learning. Learning flow was evaluated with created to assist the learners in learning the medical
qualitative data using the FSS questionnaire. The FSS education content more effectively, with Circle
questionnaire proposed by Jackson and Marsh was Marker and Highlight gaze visualization interfaces
adapted for the study and evaluated on a 5-point produced and tested for this experiment. Tobii XR
Likert scale [21]. It was evaluated that the higher the SDK was used to visualize the instructor’s gaze. The
FSS score, the better the learning flow. Circle Marker gaze visualization interface is shown in
Fig. 1, and is executed when presenting the VR
Correlation analysis was conducted to analyze the medical education content as follows. While the
correlation between quantitative data (concentration instructor wears a head-mounted display (HMD) with
by eye-tracking data) and qualitative data (flow by an eye-tracking function and views the VR education
questionnaire data). As a result of the correlation content, a ray is created in the direction of the
analysis, it was found that there was a significant instructor’s gaze, with the visualized material
correlation between learning concentration and rendered in front of the object to create a gaze
learning flow. An independent samples t-test was visualization object. This gaze visualization object
conducted to determine if there was a significant appears as an object with a sprite texture, with the
difference in learning performance based on the gaze sprite shader render queue appearing first in the Unity
visualization interface type. The t-test results scene. This causes the gaze visualization object to
confirmed that there was a statistically significant appear first, even if there are other objects behind it.
difference between the degrees of concentration and
learning flow based on the gaze visualization
interface type, with the Circle Marker type having a
more positive effect on both learning concentration
and flow.
II. DEVELOPMENT OF VIRTUAL REALITY
MEDICAL EDUCATION CONTENT WITH A GAZE
VISUALIZATION INTERFACE
VR medical education content with a gaze
visualization interface was developed to prevent the
cognitive load that may occur in existing VR Fig. 1. Circle Marker Interface.
education and improve concentration. The VR
medical education content was developed using the The Highlight gaze visualization interface is
Unity game engine, and consisted of educational shown in Fig. 2, and is executed when presenting the
scenarios to prevent patient safety accidents that may VR medical education content as follows. When the
occur during surgery. The educational environment instructor’s gaze falls upon an object that has a
and content were reviewed by current medical collider applied to it, that object is highlighted in a
professionals to increase the objectivity of the color that is set by the instructor. The object then
medical education content development. Ten Areas of returns to its original color after the instructor’s gaze
Interest (AOIs) were selected as objects that learners shifts away from that object.
508
Authorized licensed use limited to: UNIVERSITY OF SHARJAH. Downloaded on September 22,2023 at 17:32:45 UTC from IEEE Xplore. Restrictions apply.
2023 8th International Conference on Business and Industrial Research 18 - 19 May 2023
(ICBIR2023) Bangkok, Thailand
A function to measure the learner’s concentration The process of the experiment is shown in Fig. 5.
was developed to analyze how well the learner In the experiment, the Vive Pro Eye HMD was used
followed the educational content and focused on the to track the subject’s gaze tracking data. Vive Pro Eye
learning objects based on the two gaze visualization has Tobii's eye-tracking sensor built-in, and the eye-
interface types. The G2OM algorithm in Tobii XR tracking specifications are as follows. Gaze data
SDK was used to develop a function to measure the output frequency = 120Hz, Accuracy = 0.5°–1.1°
dwell time from the eye-tracking data. When the (within FOV 20°), Trackable field of view = 110°.
learner views the VR educational content, the names Each subject was first seated and given a HMD, and
of the objects (AOIs) that are gazed by the learner and an eye calibration was performed. The experimenter
their associated dwell time is measured in real-time then briefly explained the guidelines and precautions
from the eye-tracking data, as shown in Fig. 3-(a). of the experiment. The subject then viewed the VR
The dwell time is measured as the time from when the medical education content, with the subject’s dwell
learner’s gaze focuses on an AOI that has a collider time data being measured in real-time; these data
applied to it to the time the learner’s gaze moves from were saved as a CSV file after the content was viewed.
that object. The dwell time for each AOI is saved as a The subject then completed a FSS questionnaire after
comma-separated values (CSV) file after the learner viewing the VR medical educational content.
views all of the VR medical education content, as
shown in Fig. 3-(b). These data are then analyzed to
determine the degree of concentration.
509
Authorized licensed use limited to: UNIVERSITY OF SHARJAH. Downloaded on September 22,2023 at 17:32:45 UTC from IEEE Xplore. Restrictions apply.
2023 8th International Conference on Business and Industrial Research 18 - 19 May 2023
(ICBIR2023) Bangkok, Thailand
flow was found to be r = 0.303, which indicates a TABLE IV. INDEPENDENT SAMPLES T-TEST RESULTS
(LEARNING FLOW)
positive correlation. A positive correlation between
learning concentration and flow therefore existed Group N Mean SD t p
based on the correlation coefficient standard, with this
A 28 4.23 0.91
correlation highlighting that learning concentration
2.589 0.012*
would improve (degrade) as the learning flow B 28 3.64 0.78
improved (degraded).
*p < 0.05, **p < 0.01, ***p < 0.001.
510
Authorized licensed use limited to: UNIVERSITY OF SHARJAH. Downloaded on September 22,2023 at 17:32:45 UTC from IEEE Xplore. Restrictions apply.
2023 8th International Conference on Business and Industrial Research 18 - 19 May 2023
(ICBIR2023) Bangkok, Thailand
Realization of Virtual Reality Contents using [16] Y. Rahman, S. M. Asish, N. P. Fisher, E. C. Bruce, A. K.
Multimodal Sensory Interface). Kulshreshth and C. W. Borst, “Exploring eye gaze
visualization techniques for identifying distracted students in
educational VR,”
REFERENCES IEEE Conf. Virtual. Real. 3D. User. Interfaces. Online, pp.
[1] J. Bailenson, Experience on demand: What virtual reality is, 868–877, March 2020.
how it works, and what it can do. NY, USA: WW Norton & [17] C. W. Borst, N. G. Lipari and J. W. Woodworth, “Teacher-
Company, 2018. guided educational vr: Assessment of live and prerecorded
[2] Z. Feng, V. A. González, R. Amor, R. Lovreglio, and G. teachers guiding virtual field trips,”
Cabrera-Guerrero, “Immersive virtual reality serious games IEEE Conf. Virtual. Real. 3D. User. Interfaces. Germany, pp.
for evacuation training and research: A systematic literature 467–474, March 2018.
review,” Comput. Educ. New York, vol. 127, pp. 252–266, [18] G. Makransky, T. S. Terkildsen and R. E. Mayer, “Adding
December 2018. immersive virtual reality to a science lab simulation causes
[3] Z. Merchant, E. T. Goetz, L. Cifuentes, W. Keeney- more presence but less learning,” Learn. Instr. London, vol.
Kennicutt, and T. J. Davis, “Effectiveness of virtual reality- 60, pp. 225–236, April 2019.
based instruction on students’ learning outcomes in K-12 and [19] P. Albus, A. Vogt and T. Seufert, “Signaling in virtual reality
higher education: A meta-analysis,” Comput. Educ. New influences learning outcome and cognitive load,” Comput.
York, vol. 70, pp. 29–40, January 2014. Educ. New York, vol. 166, pp. 1–16, June 2021.
[4] Y. L. Chen, “The effects of virtual reality learning [20] W. O. dos Santos, I. I. Bittencourt, S. Isotani, D. Dermeval,
environment on student cognitive and linguistic L. B. Marques and I. F. Silveira, “Flow theory to promote
development,” Asia-Pac. Educ. Res. Singapore, vol. 25, pp. learning in educational systems: Is it really relevant?,” Rev.
637–646, April 2016. Bras. Inform. Educ. Brazil, vol. 26, pp. 29-59, May 2018.
[5] C. Christou, Virtual reality in education,” in Affective, [21] S. A. Jackson and H. W. Marsh, “Development and
interactive and cognitive methods for e-learning design: validation of a scale to measure optimal experience: The
creating an optimal education experience. Hershey, PA, USA: Flow State Scale,” J. Sport. Exerc. Psychol. United States,
IGI Global, 2010, pp. 228–243. vol. 18, pp. 17–35, March 1996.
[6] M. B. W. Sinambela, Y. Soepriyanto, and E. P. Adi, “Taman [22] K. Holmqvist, M. Nyström, R. Andersson, R. Dewhurst, H.
Peninggalan Sejarah Berbasis Virtual Reality,” JKTP. Jarodzka, and J. Van de Weijer, Eye tracking: A
Malang, vol. 1, pp. 7–12, 2018. comprehensive guide to methods and measures. Oxford, UK:
[7] A. K. Tarmizi, H. Hasbiyati, and M. Hakim, “Pengembangan OUP, 2011.
media pembelajaran berbasis virtual reality pada mata kuliah [23] J. L. Rosch and J. J. Vogel-Walcutt, “A review of eye-
anatomi dan fisiologi manusia pada mahasiswa semester vi tracking applications as tools for training,” Cogn.
pendidikan biologi,” Bioshell. Jember, vol. 9, pp. 37–40, Technol. Work. London, vol. 15, pp. 313–327, June 2012.
October 2020.
[24] B. Albert and T. Tullis, Measuring the user experience:
[8] M. Supriadi, and L. V. Hignasari, “Pengembangan media collecting, analyzing, and presenting usability metrics.
pembelajaran berbasis virtual reality untuk meningkatkan Burlington, MA, USA: MK, 2013.
hasil belajar peserta didik Sekolah Dasar” KOMIK. Medan,
[25] S. Comu, G. Kazar and Z. Marwa, “Evaluating the attitudes
vol. 3, pp. 578–581, October 2019.
of different trainee groups towards eye tracking enhanced
[9] B. Boyles, Virtual reality and augmented reality in education. safety training methods,” Adv. Eng. Inform. United
Center For Teaching Excellence, United States Military Kingdom, vol. 49, pp. 1-10, August 2021.
Academy, West Point. New York, pp. 1–12, 2017.
[10] Y. Lin, X. Wang, F. Wu, X. Chen, C. Wang, and G. Shen,
“Development and validation of a surgical training simulator
with haptic feedback for learning bone-sawing skill,” J.
Biomed. Inform. New York, vol. 48, pp. 122–129, April
2014.
[11] M. Gopal, A.A. Skobodzinski, H.M. Sterbling, S.R. Rao, C.
LaChapelle, K. Suzuki, and V.R. Litle, “Bronchoscopy
simulation training as a tool in medical school education,”
Ann. Thorac. Surg. Amsterdam, vol. 106, pp. 280–286, July
2018.
[12] T. Hardcastle and A. Wood, “The utility of virtual reality
surgical simulation in the undergraduate otorhinolaryngology
curriculum,” J. Laryngol. Otol. London, vol. 132, pp. 1072–
1076, November 2018.
[13] K. Hagelsteen, A. Langegård, A. Lantz, M. Ekelund, M.
Anderberg and A. Bergenfelz, “Faster acquisition of
laparoscopic skills in virtual reality with haptic feedback and
3D vision,” Minim. Invasive. Ther. Allied. Technol. London,
vol. 26, pp. 269–277, April 2017.
[14] R. N. Uppot, B. Laguna, C. J. McCarthy, G. De Novi, A.
Phelps, E. Siegel and J. Courtier, “Implementing virtual and
augmented reality tools for radiology education and training,
communication, and clinical care,” Radiol. Oak Brook, vol.
291, pp. 570–580, April 2019.
[15] O. A. Meyer, M. K. Omdahl and G. Makransky,
“Investigating the effect of pre-training when learning
through immersive virtual reality and video: A media and
methods experiment,” Comput. Educ. New York, vol. 140,
p.103603, June 2019.
511
Authorized licensed use limited to: UNIVERSITY OF SHARJAH. Downloaded on September 22,2023 at 17:32:45 UTC from IEEE Xplore. Restrictions apply.