You are on page 1of 4

Sensing the Schoolyard: Using Senses and Sensors to

Assess Georeferenced Environmental Dimensions


Maria João Silva João Correia Lopes
Escola Superior de Educação INESC Porto & Faculdade de Engenharia
Instituto Politécnico do Porto Universidade do Porto
Rua Dr. Roberto Frias, 4200 Porto, Portugal Rua Dr. Roberto Frias, s/n, 4200-465, Porto, Portugal
+351225073480 jlopes@fe.up.pt
mjosilva@ese.ipp.pt
Maria José Marcelino
Pedro Moreira da Silva Centro de Informática e Sistemas da
INESC Porto & Faculdade de Engenharia Universidade de Coimbra, Pólo II,
Universidade do Porto Pinhal de Marrocos, 3030-290 Coimbra
Rua Dr. Roberto Frias, s/n, 4200-465, Porto, Portugal Portugal
pedro.da.silva@fe.up.pt +351239790067
zemar@dei.uc.pt

ABSTRACT Keywords
This paper presents the Usense2learn platform, a platform
Multisensory georeferenced information; mobile learning;
designed and implemented to enable children and teachers to use
wireless sensor networks; children.
georeferenced multisensory information together with information
acquired by sensors. Usense2learn places the creation of content
in the hands of children. While using Usense2learn, mobile 1. INTRODUCTION
sensors can be held across the exploration area and provide Human senses are our prime interface with the environment and
georeferenced environmental information such as air temperature sensory organs can be considered biological sensors [1].
and humidity. Using multimedia (video, image, sound and text) Furthermore, sensors are typically used as an extension of human
teachers and children can bring the outside world into their senses and nowadays electronic sensors are becoming everyday
classroom and share it with other classrooms across the globe. devices in diverse activities and situations [1]. However, sensing
Having the limited schools’ budgets in mind, content creation can the multiple dimensions of the environment to monitor its quality
be made anywhere without communication costs. Internet is a complex task for children and even for adults [2]. It requires
connection is only required, together with Google Earth, for the situated interpretation of the qualitative and/or quantitative
visualization. The Usense2learn platform was successfully used in data collected including its translation to environmental quality
a curricular context, engaging children and teachers in meaningful indicators [2].
environmental education activities. In these circumstances, the research presented in this paper uses
Information and Communication Technologies (ICT), such as
Categories and Subject Descriptors multimedia tools and virtual globes, to scaffold children in
H.5.2: [Information Systems]: Information Interfaces and recording, representing and interpreting multisensory
Presentation – User Interfaces environmental data [3]. This paper presents tools and activities
K.3.1: Computers and Education: Computer Uses in Education. that give multisensory information a context, defining where,
when and why multisensory approaches are developed.
General Terms In the following section some of the more meaningful projects in
Design, Experimentation, Human Factors. the area are overviewed. Afterwards the USense2Learn platform
is presented. Subsequently, the curricular activities developed to
assess the affordances of the developed platform are described.
Finally, the conclusions are presented and future work is outlined.
Permission to make digital or hard copies of all or part of this work for
personal or classroom use is granted without fee provided that copies are 2. RELATED WORK
not made or distributed for profit or commercial advantage and that In the past decade, mobile devices have been used in diverse
copies bear this notice and the full citation on the first page. To copy educational projects to explore spatial and sensory information in
otherwise, to republish, to post on servers or to redistribute to lists, the real world. In the Ambient Wood project, children used
requires prior specific permission and/or a fee. pervasive devices and multimodal displays — including probes,
COM.Geo 2010, June 21-23, 2010 Washington, DC, USA sensors and GPS —to find things out, to discover habitats, and
Copyright 2010 ACM 978-1-4503-0031-5...$10.00. biological processes in a woodland [4]. In the Savannah project,
GPS equipped PDAs allowed children to play the role of lions. Activities can take place in the absence of Internet connectivity.
Children could “see, hear, and smell” a virtual savannah to deeper Moreover, the platform can be used on a field trip since every
explore lion behavior [5]. component must be mobile and portable. Once there is no place to
“Sound Scavenging” is a project focused on soundscapes creation. plug in each component in the field, the platform must work for at
In a series of workshops children explored “how they listen, (…) least for 120 minutes – accounting for setup time, class duration
and how to map and link the sounds geographically” [6]. After the and a considerable safety margin.
workshops the students’ sounds and pictures were mapped using Visualization of created information via Web Browser (BW) or
the free Google Earth (GE) application [6]. Google Earth (GE) in both modes (ad-hoc or infrastructured)
In the SchoolSenses@Internet project elementary school children requires an Internet connection to download satellite images (the
used mobile phones to create georeferenced multisensory local experience's content can be provided by the local network)
messages [7] [8]. Those messages were produced as MMS [9]. If a school network is used, it is very likely that it will also
messages that were automatically displayed in GE which was provide Internet access. If an ad-hoc network is used, however,
embedded in the project collaborative Website [7] [8]. GE’s the other nodes will only have Internet connection if an Internet
interface was a fundamental engaging factor in these experiences gateway is configured.
[3]. Bearing the diversity of nodes in mind, each functionality (or set
The analyses of how the GPS-equipped mobile phones and GE of functionalities) is provided as a service – implemented as
were used to explore, create, and share multisensory geographic RESTfull Web Services. Compared with solutions based on
information showed that children: linked cognitive, emotional, binary protocols the creation of Web Services clearly
and physical experiences; explored and analyzed environmental distinguishes between interface and implementation. Furthermore,
elements; represented new dimensions of the environment and the use of Web Services provides programming language,
showed improved awareness of everyday things [7]. platform and operating system independence [9]. Since
communication at a lower level is made using sockets, Web
Services can be deployed on separate machines – ideal for
3. THE USENSE2LEARN PLATFORM increasing processing power (at the cost of adding more devices).
The USense2Learn platform was developed to allow students to
create in situ multisensory georeferenced content (image, video, The server node is where all information converges and collects
sound and text). The platform (see Figure 1) requires at least one all the georeferenced multisensory information. It stores it in a
sensor (Temperature and Humidity, in the present case – Temp, database and makes it available to other nodes. The database –
Hum), one netbook (or a mobile device with similar PostGIS – provides geographic functions crucial for data querying
characteristics) and one mobile phone for content creation [9]. All based on node location [9]. Without this functionality it would not
devices must have a GPS receiver for data georeferencing and a be feasible, as collected data increases, to query for the nearest
WLAN module for Wi-fi communication. Several netbooks can sensor information (time and spatial distance).
be used to increase battery life. Each netbook runs one or more The values provided by the sensors are automatically made
services. available to the students. Sensors may change their positions,
because the data sent to the Web Service is georeferenced. Since
sensors from different manufacturers can be used, all data is sent
in the format specified by the Web Service.

3.1 Content Production


A Web-based application was created with the goal of having a
uniform and easily upgradeable interface across the diverse
devices. Mobile devices that support GPS and WLAN will
probably support browser installation or incorporate a good
browser. Web page access restrictions mean that taking
photographs, recording sound and video, are tasks that are not
integrated into the Web page. Nevertheless, the solution was
designed with the future in mind, since major players are starting
to support Web access to such recording devices [9]: Nokia Web
Figure 1: Usense2learn Platform. (BW – Web Browser; GE –
Runtime (WRT), Palm WebOS, MotorolaWebUI, W3C
Google Earth; DHCP - Dynamic Host Configuration Protocol
MobileWeb Initiative (MWI) and, at an embryonic stage, OMTP
Server; Temp – Temperature Sensor; Hum – Humidity
BONDI.
Sensor; DB – Data Base)
This way, changing the interface is as simple as changing the
The network (Net) has two operating modes – ad-hoc and Web page. Using AJAX technology, information about the
infrastructured. A custom made DHCP (Dynamic Host environment is shown to the user to provide additional
Configuration Protocol) server automatically configures IP dynamically updated data that s/he can use for content creation.
addresses in ad-hoc mode and optionally in infrastructured mode. Also, because all devices are using the same page, no conflicts
An ad-hoc network is created only when no infrastructured will arise from using different versions. Moreover, only one
network is available, because of its poorer performance and Website needs to be created, although it is true that the style must
higher power consumption. adapt to screen size (which can vary) [9].
3.2 Content Visualization interface (see figure 2), children needed to rename and select the
Once all the information is collected it must be provided in such a created files [9]. Some children showed that they were not
way that it can be easily observed by children. The multisensory comfortable with this task, especially when there were a lot of
information created is exported as a package for each developed files in the directory. But they determinedly overcame this
activity for offline exploration. A KML (Keyhole Markup obstacle as well as other hardware constraints, such as the time
Language) file is generated and the content created by the needed to hold down the button for taking photos or recording
children (image, video, sound and text) is included in the package. video.
Sight and hearing are easily reproduced using current technology,
but the other senses are not. Text is provided to allow the
introduction of additional multisensory information. All media
files (except images) are converted to mp3 (audio) and Flash
video [9]. Text is placed inside the KML file but the other media
are made available as media files. The audio and video players are
also included into the package.
In order to avoid GE limitations and Flash Security issues, a
simple custom-made HTTP server is provided in the package – if
HTTP requests are used none of the issues caused by local files
access will arise. The HTTP server was built using Ruby to
provide a platform-independent solution.
Figure 2: Mobile Web horizontal interface.
4. SENSING THE SCHOOLYARD WITH
CHILDREN The produced messages showed “play niches” of the schoolyard –
To evaluate the Usense2learn platform experiences were designed such as places for running, for playing basketball, “human chess”
to have children producing content in outdoor activities and later – as well as facilities and problems related to solid waste.
visualizing it in the classroom. The developed experiences
The activities took place on two summer mornings and the
involved 25 children (13 boys and 12 girls), aged 10 to 11, from a
temperature rose as time passed. Both temperature and humidity
classroom of an elementary school of the city of Porto, Portugal. 3
showed very different values in places such as near the trees or
curricular activities were implemented in 3 classes of 90 minutes,
the building, on the basketball court or on the running track. The
held on 3 days, to assess the affordances of the developed
teachers supported children in verbally expressing what they were
platform. During the outdoor activities children worked in 8
sensing both with their senses and with the sensors in order to
groups, 7 groups of 3 and 1 group of 4 children.
enable them to relate abstract ideas to concrete data.
The Senses@School game was the strategy adopted to structure
the children’s activities [7]. The challenge presented to the
children was an open-ended task: to share multisensory
information about their schoolyard with other schools.
In the activities, an HTC TyTN II mobile phone (which has an
integrated keyboard), two laptops for children, based on the Intel
Classmate PC, and one sensor for air temperature and humidity
measurement (PASCO PASport PS-2124A) were used. To
overcome the lack of GPS, the sensor geoposition was acquired
through synchronization with the mobile phone.
Before going out to the schoolyard, children observed, using GE,
and commented on the georeferenced multisensory information
created by children in two other schoolyards [7]. In this
experience children could not only produce georeferenced
multisensory information, but also enrich such information using
sensors and producing videos while exploring their schoolyard.
The 4 groups of children that worked on the first day observed
and commented on the temperature and humidity data acquired at
several points in their schoolyard by the portable sensors and
produced georeferenced multisensory messages with a video, a Figure 3: Content exhibition in Google Earth (text message:
photo, a sound and a small text [9]. On the second day the 25 Running is healthy when it is not hot".)
children observed the first day messages. Afterwards, the other 4
groups of children produced their messages [9]. On the second day, since the temperature and humidity data were
Children used the mobile phone and the sensors in an engaged being updated on the mobile screen during the activity, children
way to explore the schoolyard. Since multimedia recording were explicitly asked to relate their messages to these data. Using
(image, sound and video) was not integrated into the Web this information, one of the groups wrote the following text:
“When we are under the shade of the trees, it gets cooler" [9].
Another group wrote that “running is healthy, when it is not hot” Future work will allow the exploration and social edition (e.g., "I
(see figure 3). like this", "I don't like this" and small text annotations) of the
A third session was implemented three weeks later. In this quantitative and qualitative georeferenced information. In order to
session, children used GE to observe, analyse and discuss all the make this possible, the USense2Learn platform will include: an
produced messages (see figure 4). One member of the research ontology that will support content exploration based on
team and the class teacher discussed with children the temperature geospatial, temporal and conceptual criteria; and a database for
and humidity data recorded in the georeferenced multisensory storing the sensor data, the multisensory content and the
messages, relating the data to when and where the messages were annotations produced by the children. This enhanced
produced. The children revisited the sensations experienced USense2Learn platform will be used in schools of diverse regions
during the development of each message and related them to the of Portugal to assess the environmental quality of each schoolyard
sensors’ values. and to support the creation and use of context aware interactive
objects. These objects will be created to improve the schools’
sustainability, whilst also contributing to scaffold the learning of
environmental complexity using a tangible experimentation
approach.

6. REFERENCES
[1] Gouveia , C. 2008 Collaborative environmental monitoring:
linking the senses. Doctoral Thesis. Universidade Nova de
Lisboa.
[2] Silva, M. J., Hipolito, J., and Gouveia, C. 2003. Messages
for environmental collaborative monitoring: The
development of a multi-sensory Clipart. In Human-Computer
Interaction INTERACT’03, M. Rauterberg, M. Menozzi, J.
Wesson, Eds. IOS Press, 896-899.
[3] Silva, M. J., Gomes, M., Brigas, C., Marcelino, M., and
Gouveia, C. 2008 Children as authors of georeferenced
multisensory information: towards the design of simulation
Figure 4: Google Earth – Content navigation. and editing tools for an educational context. In Proc. of the
ED-MEDIA 2008 (Vienna, June 2008), 1578-1586.
[4] Rogers, Y., Price, S., Fitzpatrick, G., Fleck, R., Harris, E.,
Afterwards, during this session, children were asked to give some Smith, H., Randell, C., Muller, H., O'Malley, C., Stanton, D.,
reasons that could influence other schools to develop similar Thompson, M. and Weal, M. 2004 Ambient Wood:
activities. 18 children (6 boys and 12 girls) mentioned the sensing Designing new forms of digital augmentation for learning
of temperature and humidity, 12 children (7 boys and 5 girls) outdoors. In Proc. of IDC 2004 (USA, Maryland, June
mentioned that the activities were fun, 11 children (5 boys and 6 2004), ACM Press, 3-10.
girls) pointed out that technology and media are reasons to value [5] Facer, K., Joiner, R. Stanton, D. Reidz, J. Hullz, R., and
the activities and 4 children (3 boys and 1 girl) said that the Kirk, D. 2004. Savannah: mobile gaming and learning?
activities were important to improve learning. Journal of Computer Assisted Learning, 20, 6 (December,
2004). 399–409.
5. CONCLUSION [6] Proboscis, Loren Chasse and the Jenny Hammond Primary
The Usense2learn platform was designed, developed and used in School 2006 Sound Scavenging Report.
order to make it possible for children and teachers, using location http://proboscis.org.uk/topographies/SoundScavenging_Repo
aware mobile devices, to create and visualize georeferenced rt.pdf
information acquired by sensors and human senses. The [7] Silva, M. J., Gomes, C. A., Lopes, J. C., Marcelino, M. J.,
Usense2learn platform was successfully used in a curricular Gouveia, C., Fonseca, Alexandra, and Pestana, B. 2009.
context, engaging children and teachers in meaningful Adding Space and Senses to Mobile World Exploration. In
environmental education activities. Mobile Technology for Children: Designing for Interaction
and Learning, A. Druin, Ed., Morgan Kaufmann, Boston,
The Usense2learn platform showed potential as a scaffolding tool
147-169
for the analysis of environmental information, since it allowed
[8] Silva, M. J., Pestana, B., and Lopes, J. C. 2008. Using a
children together with their teachers to explore, create,
mobile phone and a geobrowser to create multisensory
communicate and reflect on qualitative and quantitative, objective
geographic information, in Proc. of IDC 2008 (Chicago,
and subjective environmental information in their everyday
June 2008), ACM Press, 153-156.
context. The ICT tools used were found to be engaging for
[9] Silva, P. M., Lopes, J. C., and Silva, M. J. 2009. Ubiquitous
children and to allow (re)visiting sensory experiences in more
computing in children education: Framework for
formal and abstract contexts.
georeferenced multisensory information. In Proceedings of
XI International Symposium on Computers in Education,
(Coimbra, Portugal, November 2009). SIIE'09. Universidade
de Coimbra.

You might also like