Professional Documents
Culture Documents
Group Name:
Student 1 Name/ID: Student 2 Name/ID:
Student 3 Name ID: Student 4 Name ID:
Student 5 Name ID:
Introduction
•Artificial Intelligence, Machine Learning and Natural Language Processing are
all technologies aimed at achieving humanism in chatbots.
•Artificial intelligence (AI) is now widely used in a variety of industries,
including:
1. Fraud detection 4. Natural Language Processing
2. Visual synthesis 5. Natural Language Understanding.
3. Voice recognition
•There are over 40,000 machine intelligence conversation agents distributed across many platforms.
•Chat-bots provide a conversational experience by simulating real-life discussions with artificial intelligence and
natural language processing.
•Advancement in a software development improves replies over time by learning from previous interactions.
Introduction (cont’d)
•Humanistic chat-bots allow for human-like interaction to be supplied
through a flexible channel.
•Some of the best examples of current chat-bots in the market include:
1. Apple’s Siri
2. Microsoft’s Cortana
3. Google’s Google assistant and
4. Facebook’s Messenger chat-bot
5. ALICE
6. MYCin
Problem statement
•Experts in the field of artificial intelligence, machine learning, and software development are
constantly looking for a way to make chat-bots as humanistic as possible.
•Modern corporations are continuously attempting to expand their technological capabilities,
as seen by the structured literature study.
•Through technological improvements, the big business hopes to improve customer service
and expand service delivery.
•This is to obtain a competitive advantage over other capable corporations in terms of
profitable benefits and client base expansion.
•To assist consumers with their inquiries, a domain-specific chatbot will be built.
•In order to address the concerns of customer satisfaction caused by long lines and physical
encounters.
Problem statement (Cont’d)
•In order to address the concerns of customer satisfaction caused by long lines and physical
encounters.
•The chatbot will enable personalized and speedy reliable communication and their chatbots
to help them manage their accounts, orders, or tickets.
•For example, chatbots can help customers with bank statements when they need them,
answer questions, and schedule appointments with organizational managers.
•Using natural language used in messaging, the chatbot will make customers feel secure and
at ease when using the service, irrespective of their computer knowledge.
•It also offers very convenient and quality services because all transactions take place within a
single chat discussion.
•The use of chatbots helps in eliminating the need for the user to browse via a website.
Relevant literature
•The project adopted a cross-sectional literature review of over 500 peer-reviewed publications.
•The following four terms were used to conduct the literature review:
1. Context-awareness
2. Understanding of stated and implied expressions
3. Ability to manage cognitive load (ambiguity, complexity, nesting, and vagueness)
4. Ability to independent interpretation (lexical and semantic understanding)
•The terms are addressed, as well as the sub-terms included during the project's annotated
bibliography and NVIVO analysis stages.
•Additional futuristic questions have arisen as a result of the literature review conducted on a large number of
publications and studies published by the authors.
•A thorough study of the literature used to create an annotated bibliography was carried out.
2nd Presenter
Description of the methodology
•The aim of this particular study was to deliver an SLR report as the product
•The approach employed was aimed at having a detailed review of available
publications that focus on chatbots obtained from Gitlab
Project tasks
•The project involved the following tasks:
1. Querying the Gitlab repository for relevant articles related to conversation agents
2. Downloading the material into Endnote
3. Converting the endnote files into pdf for easier reading
Project tasks (cont’d)
4. Analyzing the articles to ascertain their appropriateness to the topic under study
Selecting the most ideal materials and discarding the rest
5. Preparing an annotated bibliography on the articles using the provided terms
6. Downloading Nvivo and registering to gain access
7. Uploading the files into Nvivo analysis software
8. Querying the files using the terms provided including new sub-terms
9. Going through the selected paragraphs and highlight then coding the paragraphs
10. Visualizing the data and including it in the final report
11. Write up the final 10000 words report.
How the product was developed
•We chose the research questions for this study because we wanted to provide
extensive explanations of both the benefits and shortcomings of the chatbots that are
already in use.
•The main goal was to see how similar some of the most recently produced chatbots
are to being human in terms of dialogue.
•In order to achieve this task, we divided the task into the following stages;
1. Article identification and extraction.
2. Information analysis and coding using NVIVO
3. Creating annotated bibliographies and
4. Report writing.
The Process adopted
Step 1:
•The Federation University of Australia repository was queried to identify and extract relevant
publications.
•Over 700 relevant publications were found in the university database, which we utilized during the
research.
•We used the terms "context awareness," "cognitive load," "complexity," "vague," "nesting," and
"ambiguity" to help us find the materials we were looking for. "lexical understanding" and "semantic
understanding" "Emotion," "expression," "interpretation," "comprehension," "humanness," and
"humanistic" are all terms used to describe how people feel.
•Whenever we clicked the query button, the terms we used produced over 100 relevant articles.
•We were able to gather approximately 700 articles in XML format, which we then imported into the
EndNote software.
The Process adopted (Cont’d)
•The articles were converted into PDF files that could be opened using Adobe Acrobat Professional.
•The papers would then be imported into the Nvivo analysis program.
Step 2:
•Relevance, credibility, reliability, and transferability were to be evaluated further on the collected papers
collected.
•We used the Nvivo software to categorize the papers and prepare the repository for coding.
•We checked the articles' legitimacy by running them via scholarly databases such as Google Scholar.
•We decided to use a few of the articles we found on Google Scholar because they were credible.
•Out of the 2,000 items we collected from the Federation University of Australia database, we only deleted
100.
•The findings aided us in gaining full knowledge of our study's themes, narratives, interpretations, and
contexts.
3rd Presenter
The Process adopted (Cont’d)
•The findings aided us in gaining full knowledge of our study's themes, narratives, interpretations,
and contexts.
•To acquire relevant information about our research topic, we employed the NVIVO qualitative
analysis program.
•The software is primarily used for qualitative analysis of a variety of content-rich sources,
including surveys, peer-reviewed publications, fragmented texts, and social media, among others.
•Nvivo is a software program developed by QSR International that can code vast amounts of data
into specific categories.
•In this study, for example, we utilized the software to analyze over 500 articles to find only those
that were relevant to our course.
•We were able to augment the relevant articles and develop an annotated bibliography afterward.
The Process adopted (Cont’d)
•The procedure for using Nvivo involved:
1. Registering on the software using our emails and entering the product key
2. Creating a new project with a unique project title or name
3. Importing data into the software; importing the articles
4. Creating a new folder and dumping the articles/files into it
5. Developing cases
6. Creating codes
7. Creating themes to be used
8. Creating search/text codes
9. Querying the files
10. Finally coding the relevant information to align code.
11. Saving the project into a new folder outside NVIVO.
12. Closing Nvivo.
The Process adopted (Cont’d)
•The software's highlighted information was then coded according to how significant it was to the
research topic.
•The coding or allocating a paragraph to an appropriate theme or code took about three days on
average.
Step 3
•We created an annotated bibliography based on a combination of two terms we allocated ourselves
during the project's initial stages. A pair's key phrases yielded about 100 articles each, from which
the annotated bibliography was compiled.
Step 4
•We organized our activities into a write-up session at the end of the project.
•We attempted to synthesize our knowledge and generate a 10000+ word structured literature review
report.
Evidence of research
Theme Discussion
Faris and Cheng, 2012; Finch et al, 2020; Sam, 2020; Waldera,
Cognitive load 2019; Moussiades, 2020
Kim, Lee, & Gweon, 2019; David et al, 2020; Tracey et al,
Ambiguity 2021; Yadav et al, 2019; Raval, 2020
Santos et al, 2020; Augello et al, 2012; Steinbock & Rao, 2020;
Complexity Shawar and Atwell, 2007; Pereira et al, 2016
Evidence of research (Cont’d)
Baudart et al, 2018; Zhang et al, 2016; Cahn, 2017; Wang et al,
Vague 2016; Galitsky, 2021;
Kim, Lee, & Gweon, 2019; Syvanen and Valentini (2020);
Nesting Candello, Pinhanez, and Figueiredo, 2017; Coucke et al, 2018
Thapa, 2019; Samuel et al, 2020; Fensel et al, 2020; Santos et al,
Expression 2020; Pereira & Díaz, 2018; Jain et al, 2018; Liu et al, 2020;
Cheng, 2021