Professional Documents
Culture Documents
Co
Seminar
Data Science
Jochen Kempfle, Jonas Pöhler, Lukas Wegmeth
Alexander Hölzemann & Marius Bock
• Contact
Time: Mo. 2pm – 4pm
Office: H-A 8106 | H-A 8104/05 | H-C 8318 |H-A 7117/2 | online
Mail: jochen.kempfle@uni-siegen.de
jonas.poehler@uni-siegen.de
lukas.wegmeth@uni-siegen.de
alexander.hoelzemann@uni-siegen.de
marius.bock@uni-siegen.de
https://moodle.uni-siegen.de/course/view.php?id=22277
Enrolment key:
UbiDS@22!
• Project description
– What is the goal of your project?
– What data/ resources do you have available?
• Approach (sketch)
– Milestones
– Mock-ups (if applicable)
– (Potential) problems & challenges
• Responsibilities
– Who will do what?
• Related work
– Summarize field(s) of research your project relates with
– What are popular publications?
– How does your project relate to them?
– Which ones influence your project & in what way?
• Implementation
– Summarize your implementation so far
– Which methods/ measures/ metrics are you using?
• Work documentation
– Who did what when (short tabular overview)
• Functional software (code + resources)
– Classes and functions are documented, i.e. there is a short description
for input, output, and purpose as comment
– Error-free compiling code
– Provided e.g. via GitHub (www.github.com)
• Scientific report
– Complete the report by systematically evaluating & summarising
findings of your project
– Revise the complete document according to the final feedback
– Chapters: Evaluation & Conclusion
• Mid-term Presentation
– Which milestones are completed? Which ones still have to be done?
– Methodology (data, algorithms, preliminary results)
– Short demo of the software (if applicable)
– Problems and challenges
• Final Presentation
– Summary of your project
– Demo of the software / presentation of final results
– Outlook & future research
• Project 4: Activity Recognition On Sensor Based Data Activity Data using the
TNDA-HAR Dataset
– Supervisor: Alexander Hölzemann (alexander.hoelzemann@uni-siegen.de)
• Task:
– Create a small AI-controlled Jump‘n‘Run game using Unity
• Create all necessary assets and the game logic
un
• Train the AI using the Unity machine learning toolkit (Unity agents)
– Evaluate your model and provide meaningful measures
‘ R
‘n
• Show and document the progress of your AI during learning
in k
T h
Outline:
– Provide challenges to the agent
• Obstacles like pits and walls
• Platforms (and interactable environment e.g. trampolines)
• Enemies (and collectibles)
– Start with easy tasks and increase difficulty towards the end of the level
• Gives the agent a chance to learn step by step
• Increases learning speed
– Come up with a meaningful reward function
• Distance to the goal
• Time
• Collected stuff
• Etc.
– Use a tile map as input to the neural network and actions as output
– Bonus: provide interesting challenges to the AI
Dataset
8 Channels Raw sensor data (EEG, EOG, GSR, Skin Temperature, IR Response,… )
3 Labels (Arousal, Dominance, Valence)
60 Subjects
Tasks
Understand and visualize the dataset
Preprocess raw signal data
If necessary compute adequate features
Implement a Multitask Model or an Model Ensemble to predict all three labels
Evaluate the model
Literature
https://ieeexplore.ieee.org/abstract/document/6127029
https://ieeexplore.ieee.org/abstract/document/8567470
1
8
Tasks:
1. Download and Read-In the TNDA-HAR Dataset:
https://ieee-dataport.org/open-access/tnda-har-0#files
2. Preprocess the data according the DL-HAR-Pipeline: https://arxiv.org/pdf/2110.06663.pdf
3. Implement and train a neural network which is capable to classify the sensor data
Difficulty: Intermediate
Note:
1. Developing an understanding of the dataset as well preprocessing the data correctly according the
pipeline that is defined in the paper seems to be easy but can be challenging.
2. Network definitions can be taken from the from other authors but need to be further developed and
the original source must be cited.
1
9
Advisor: Alexander Hölzemann
alexander.hoelzemann@uni-siegen.de
Seminar Data Science – Ubiquitous Computing – University of Siegen – 06.04.22 19
Project 5: Walk detection –
Do we walk differently when we are monitored?
Description:
7 subjects were monitored with a smartwatch weared at their wrist for one day. The first few minutes of every
subjects data, the subjects were asked to walk while monitored and were recorded via video. The rest of the
day, participants were not monitored any longer. Are we able to detect differences in their walking behavior
between monitored / unmonitored?
Tasks:
1. Inspect a dataset of 7 given participants with data from daily living and analyse if the subject walks
different in a monitored and non-monitored environment.
2. Preprocess the data according the DL-HAR-Pipeline: https://arxiv.org/pdf/2110.06663.pdf
3. Evaluate with deep learning if this is the case.
Note:
The project can be considered to be very challenging, therefore we expect participants to have significant prior
knowledge in deep learning, signal processing and programming in python as well as a high self-motivation.
However, results are supposed to be sent to either to IWOAR 2022 (https://iwoar.org/2022/: Deadline May
2
31th) or ISWC 2022 (https://iswc.net/iswc21: Deadline tba, but normally late June). 0
• Dataset:
- Data will be collected and processed by you!
- Climbing flights of stairs in e.g. Hölderlin Building
- Trying out different types of labels
Source: Andrej Karpathy
• Difficulty: Easy – intermediate
- Prior knowledge in Deep Learning is beneficial, but not
mandatory
Advisor: Marius Bock Investigated Architecture: DeepConvLSTM
marius.bock@uni-siegen.de
https://simpleassign.com/poll/-MzycQPcH7TokpLvLzHd
(link can be also found on Moodle)