Key Item Type Publication Year Author Title Publication Title ISBN ISSN
DOI Url Abstract Note Date Date Added Date Modified Access Date
Pages Num Pages Issue Volume Number Of Volumes Journal Abbreviation
Short Title Series Series Number Series Text Series Title
Publisher Place Language Rights Type Archive Archive Location
Library Catalog Call Number Extra Notes File Attachments Link Attachments
Manual Tags Automatic Tags Editor Series Editor Translator
Contributor Attorney Agent Book Author Cast Member Commenter Composer
Cosponsor Counsel Interviewer Producer Recipient Reviewed Author
Scriptwriter Words By Guest Number Edition Running Time
Scale Medium Artwork Size Filing Date Application Number
Assignee Issuing Authority Country Meeting Name Conference Name
Court References Reporter Legal Status Priority Numbers Programming
Language Version System Code Code Number Section Session
Committee History Legislative Body
UFWPMC8A conferencePaper 2024 "Kumar, Yulia; Manikandan, Anjana; Li, J.
Jenny; Morreale, Patricia" Optimizing Large Language Models for Auto-Generation
of Programming Quizzes 2024 IEEE Integrated STEM Education Conference (ISEC)
10.1109/ISEC61299.2024.10665141
[Link] "This study analyzes the use
of Large Language Models (LLMs) like ChatGPT in creating quizzes for Java
programming courses, specifically Object-Oriented Programming (CS1) and Data
Structures (CS2). It aims to evaluate the accuracy of LLM-generated assessments,
understand the benefits and drawbacks of using LLMs in CS education from educators'
viewpoints, and identify effective prompt engineering strategies to enhance the
quality of educational materials. The research compares quizzes made by LLMs
against human-created content to assess their consistency with Java programming
principles, alignment with CS1 and CS2 learning goals, and their impact on student
engagement and comprehension, providing insights into LLMs' effectiveness in
academic assessment creation for computer science education." 2024-03
1/11/2025 14:13 1/11/2025 14:13 1/11/2025 14:13 5-Jan
IEEE Xplore ISSN: 2473-7623 C:\Users\Mahnoor\Zotero\storage\
ELJJD9I2\[Link] Accuracy; AI-Supplemental Instructor (AI-SI);
Chatbots; Computer science education; Data structures; Education; Java; Java
programming instruction; Large language models; use of LLMs in CS education
2024 IEEE Integrated STEM Education Conference (ISEC)
SGSGN6DQ preprint 2024 "Chen, Yanxin; He, Ling" Research on the
Application of Large Language Models in Automatic Question Generation: A Case Study
of ChatGLM in the Context of High School Information Technology Curriculum
10.48550/arXiv.2408.11539 [Link] "This
study investigates the application effectiveness of the Large Language Model (LLMs)
ChatGLM in the automated generation of high school information technology exam
questions. Through meticulously designed prompt engineering strategies, the model
is guided to generate diverse questions, which are then comprehensively evaluated
by domain experts. The evaluation dimensions include the Hitting(the degree of
alignment with teaching content), Fitting (the degree of embodiment of core
competencies), Clarity (the explicitness of question descriptions), and Willing to
use (the teacher's willingness to use the question in teaching). The results
indicate that ChatGLM outperforms human-generated questions in terms of clarity and
teachers' willingness to use, although there is no significant difference in hit
rate and fit. This finding suggests that ChatGLM has the potential to enhance the
efficiency of question generation and alleviate the burden on teachers, providing a
new perspective for the future development of educational assessment systems.
Future research could explore further optimizations to the ChatGLM model to
maintain high fit and hit rates while improving the clarity of questions and
teachers' willingness to use them." 8/21/2024 1/11/2025 14:15 1/11/2025 14:15
1/11/2025 14:15 Research on the
Application of Large Language Models in Automatic Question Generation
arXiv [Link]
arXiv:2408.11539 [cs] C:\Users\Mahnoor\Zotero\storage\P4PGZWX7\Chen
and He - 2024 - Research on the Application of Large Language Models in Automatic
Question Generation A Case [Link]; C:\Users\Mahnoor\Zotero\storage\QJWRD66F\
[Link] Computer Science - Computers and Society
arXiv:2408.11539
RSEFUYTD conferencePaper 2024 "Zhang, Zhenhong; Chen, Jiajing; Shi, Weiyan;
Yi, Lingjie; Wang, Chihang; Yu, Qian" Contrastive Learning for Knowledge-Based
Question Generation in Large Language Models 2024 5th International Conference
on Intelligent Computing and Human-Computer Interaction (ICHCI) 979-8-3503-6828-4
10.1109/ICHCI63580.2024.10808104
[Link] "With the rapid development
of artificial intelligence technology, especially the increasingly widespread
application of question-and-answer systems, high-quality question generation has
become a key component in supporting the development of these systems. This article
focuses on knowledge-based question generation technology, which aims to enable
computers to simulate the human questioning process based on understanding specific
texts or knowledge bases. In light of the issues of hallucination and knowledge
gaps present in large-scale language models when applied to knowledge-intensive
tasks, this paper proposes an enhanced question generation method that incorporates
contrastive learning. This method utilizes multiple models to jointly mine domain
knowledge and uses contrastive learning to guide the model in reducing noise and
hallucinations in generation. Experimental results show that by designing prompts
containing contrasting examples, the model's performance in question generation
improves considerably, particularly when contrasting instructions and examples are
used simultaneously, leading to the highest quality of generated questions and
improved accuracy. These results demonstrate that the method proposed in this
study, which combines contrasting context and chain-of-thought prompts, can
effectively improve both the quality and the practicality of question generation."
9/27/2024 1/11/2025 14:30 1/11/2025 14:30 1/11/2025 14:30 583-587
IEEE "Nanchang, China"
en [Link] [Link] (Crossref)
2024 5th International
Conference on Intelligent Computing and Human-Computer Interaction (ICHCI)
IFS52PA5 preprint 2024 "Lohr, Dominic; Berges, Marc; Chugh, Abhishek;
Kohlhase, Michael; Müller, Dennis" Leveraging Large Language Models to Generate
Course-specific Semantically Annotated Learning Objects
10.48550/arXiv.2412.04185 [Link] "Background:
Over the past few decades, the process and methodology of automated question
generation (AQG) have undergone significant transformations. Recent progress in
generative natural language models has opened up new potential in the generation of
educational content. Objectives: This paper explores the potential of large
language models (LLMs) for generating computer science questions that are
sufficiently annotated for automatic learner model updates, are fully situated in
the context of a particular course, and address the cognitive dimension understand.
Methods: Unlike previous attempts that might use basic methods like ChatGPT, our
approach involves more targeted strategies such as retrieval-augmented generation
(RAG) to produce contextually relevant and pedagogically meaningful learning
objects. Results and Conclusions: Our results show that generating structural,
semantic annotations works well. However, this success was not reflected in the
case of relational annotations. The quality of the generated questions often did
not meet educational standards, highlighting that although LLMs can contribute to
the pool of learning materials, their current level of performance requires
significant human intervention to refine and validate the generated content."
12/5/2024 1/11/2025 14:31 1/11/2025 14:31 1/11/2025 14:31
arXiv en
[Link] arXiv:2412.04185 [cs] C:\Users\Mahnoor\Zotero\
storage\4TTGRY64\Lohr et al. - 2024 - Leveraging Large Language Models to Generate
Course-specific Semantically Annotated Learning [Link] Computer
Science - Artificial Intelligence
arXiv:2412.04185
2Y9HRV9W conferencePaper 2021 "Cao, Shuyang; Wang, Lu" Controllable Open-
ended Question Generation with A New Question Type OntologyProceedings of the 59th
Annual Meeting of the Association for Computational Linguistics and the 11th
International Joint Conference on Natural Language Processing (Volume 1: Long
Papers) 10.18653/v1/[Link]-long.502
[Link] "We investigate the less-
explored task of generating open-ended questions that are typically answered by
multiple sentences. We first define a new question type ontology which differentiates
the nuanced nature of questions better than widely used question words. A new
dataset with 4, 959 questions is labeled based on the new ontology. We then propose
a novel question type-aware question generation framework, augmented by a semantic
graph representation, to jointly predict question focuses and produce the question.
Based on this framework, we further use both exemplars and automatically generated
templates to improve controllability and diversity. Experiments on two newly
collected large-scale datasets show that our model improves question quality over
competitive comparisons based on automatic metrics. Human judges also rate our
model outputs highly in answerability, coverage of scope, and overall quality.
Finally, our model variants with templates can produce questions with enhanced
controllability and diversity." 2021 1/11/2025 14:32 1/11/2025 14:32
1/11/2025 14:32 6424-6439
Association for Computational Linguistics Online en
[Link] (Crossref) C:\Users\Mahnoor\Zotero\
storage\6YVULERH\Cao and Wang - 2021 - Controllable Open-ended Question Generation
with A New Question Type [Link]
Proceedings
of the 59th Annual Meeting of the Association for Computational Linguistics and the
11th International Joint Conference on Natural Language Processing (Volume 1: Long
Papers)
3SBAXMK8 bookSection 2024 "Molina, Ivo Lodovico; Švábenský, Valdemar;
Minematsu, Tsubasa; Chen, Li; Okubo, Fumiya; Shimada, Atsushi" Comparison of
Large Language Models for Generating Contextually Relevant Questions
[Link] "This study explores the
effectiveness of Large Language Models (LLMs) for Automatic Question Generation in
educational settings. Three LLMs are compared in their ability to create questions
from university slide text without fine-tuning. Questions were obtained in a two-
step pipeline: first, answer phrases were extracted from slides using Llama 2-Chat
13B; then, the three models generated questions for each answer. To analyze whether
the questions would be suitable in educational applications for students, a survey
was conducted with 46 students who evaluated a total of 246 questions across five
metrics: clarity, relevance, difficulty, slide relation, and question-answer
alignment. Results indicate that GPT-3.5 and Llama 2-Chat 13B outperform Flan T5
XXL by a small margin, particularly in terms of clarity and questionanswer
alignment. GPT-3.5 especially excels at tailoring questions to match the input
answers. The contribution of this research is the analysis of the capacity of LLMs
for Automatic Question Generation in education." 2024 1/11/2025 14:32
1/11/2025 14:32 1/11/2025 14:32 137-143 15160
en [Link]
DOI: 10.1007/978-3-031-72312-4_18 arXiv:2407.20578 [cs] C:\Users\
Mahnoor\Zotero\storage\LWS7V6BM\Molina et al. - 2024 - Comparison of Large Language
Models for Generating Contextually Relevant [Link] Computer
Science - Computers and Society; Computer Science - Artificial Intelligence;
Computer Science - Computation and Language
XZ7XNWJQ journalArticle "Chen, Yanxin; He, Ling" Research on the
Application of Large Language Models in Automatic Question Generation: A Case Study
of ChatGLM in the Context of "This study investigates
the application effectiveness of the Large Language Model (LLMs) ChatGLM in the
automated generation of high school information technology exam questions. Through
meticulously designed prompt engineering strategies, the model is guided to
generate diverse questions, which are then comprehensively evaluated by domain
experts. The evaluation dimensions include the Hitting (the degree of alignment
with teaching content), Fitting (the degree of embodiment of core competencies),
Clarity (the explicitness of question descriptions), and Willing to use (the
teacher's willingness to use the question in teaching). The results indicate that
ChatGLM outperforms human-generated questions in terms of clarity and teachers'
willingness to use, although there is no significant difference in Hitting and
Fitting. This finding suggests that ChatGLM has the potential to enhance the
efficiency of question generation and alleviate the burden on teachers, providing a
new perspective for the future development of educational assessment systems.
Future research could explore further optimizations to the ChatGLM model to
maintain high fit and hit rates while improving the clarity of questions and
teachers' willingness to use them. 1 Introduce 1.1 BackGroud High school
information technology, as a required course for high school students, presents a
rich and extensive coverage of knowledge points, imposing a significant teaching
burden on information technology teachers. In recent years, the rise of Large
Language Models (LLMs), particularly the application of ChatGLM, offers new
possibilities for addressing this issue. Through carefully designed prompts, LLMs
can generate questions that meet the requirements for high school information
technology assessment, thereby providing a more efficient and convenient teaching
tool for information technology teachers." 1/11/2025 14:32 1/11/2025
14:32
en Zotero C:\Users\Mahnoor\
Zotero\storage\RGZZLS6T\Chen and He - Research on the Application of Large Language
Models in Automatic Question Generation A Case [Link]
ZKBXUZXG conferencePaper 2024 "Zhang, Zhenhong; Chen, Jiajing; Shi, Weiyan;
Yi, Lingjie; Wang, Chihang; Yu, Qian" Contrastive Learning for Knowledge-Based
Question Generation in Large Language Models 2024 5th International Conference
on Intelligent Computing and Human-Computer Interaction (ICHCI) 979-8-3503-6828-4
10.1109/ICHCI63580.2024.10808104
[Link] "With the rapid development
of artificial intelligence technology, especially the increasingly widespread
application of question-and-answer systems, high-quality question generation has
become a key component in supporting the development of these systems. This article
focuses on knowledge-based question generation technology, which aims to enable
computers to simulate the human questioning process based on understanding specific
texts or knowledge bases. In light of the issues of hallucination and knowledge
gaps present in large-scale language models when applied to knowledge-intensive
tasks, this paper proposes an enhanced question generation method that incorporates
contrastive learning. This method utilizes multiple models to jointly mine domain
knowledge and uses contrastive learning to guide the model in reducing noise and
hallucinations in generation. Experimental results show that by designing prompts
containing contrasting examples, the model's performance in question generation
improves considerably, particularly when contrasting instructions and examples are
used simultaneously, leading to the highest quality of generated questions and
improved accuracy. These results demonstrate that the method proposed in this
study, which combines contrasting context and chain-of-thought prompts, can
effectively improve both the quality and the practicality of question generation."
9/27/2024 1/11/2025 14:32 1/11/2025 14:32 1/11/2025 14:32 583-587
IEEE "Nanchang, China"
en [Link] [Link] (Crossref)
C:\Users\Mahnoor\Zotero\storage\HTXZNGYH\Zhang et al. -
2024 - Contrastive Learning for Knowledge-Based Question Generation in Large
Language [Link]
2024 5th International
Conference on Intelligent Computing and Human-Computer Interaction (ICHCI)
Y8DCGP9M preprint 2024 "Lohr, Dominic; Berges, Marc; Chugh, Abhishek;
Kohlhase, Michael; Müller, Dennis" Leveraging Large Language Models to Generate
Course-specific Semantically Annotated Learning Objects
10.48550/arXiv.2412.04185 [Link] "Background:
Over the past few decades, the process and methodology of automated question
generation (AQG) have undergone significant transformations. Recent progress in
generative natural language models has opened up new potential in the generation of
educational content. Objectives: This paper explores the potential of large
language models (LLMs) for generating computer science questions that are
sufficiently annotated for automatic learner model updates, are fully situated in
the context of a particular course, and address the cognitive dimension understand.
Methods: Unlike previous attempts that might use basic methods like ChatGPT, our
approach involves more targeted strategies such as retrieval-augmented generation
(RAG) to produce contextually relevant and pedagogically meaningful learning
objects. Results and Conclusions: Our results show that generating structural,
semantic annotations works well. However, this success was not reflected in the
case of relational annotations. The quality of the generated questions often did
not meet educational standards, highlighting that although LLMs can contribute to
the pool of learning materials, their current level of performance requires
significant human intervention to refine and validate the generated content."
12/5/2024 1/11/2025 14:33 1/11/2025 14:33 1/11/2025 14:33
arXiv en
[Link] arXiv:2412.04185 [cs] C:\Users\Mahnoor\Zotero\
storage\V73MJHXA\Lohr et al. - 2024 - Leveraging Large Language Models to Generate
Course-specific Semantically Annotated Learning [Link] Computer
Science - Artificial Intelligence
arXiv:2412.04185
YVKFBA4I journalArticle 2022 "Zhang, Ruqing; Guo, Jiafeng; Chen, Lu; Fan,
Yixing; Cheng, Xueqi" A Review on Question Generation from Natural Language Text
ACM Transactions on Information Systems "1046-8188, 1558-2868"
10.1145/3468889 [Link] "Question
generation is an important yet challenging problem in Artificial Intelligence (AI),
which aims to generate natural and relevant questions from various input formats,
e.g., natural language text, structure database, knowledge base, and image. In this
article, we focus on question generation from natural language text, which has
received tremendous interest in recent years due to the widespread applications
such as data augmentation for question answering systems. During the past decades,
many different question generation models have been proposed, from traditional
rule-based methods to advanced neural network-based methods. Since there have been
a large variety of research works proposed, we believe it is the right time to
summarize the current status, learn from existing methodologies, and gain some
insights for future development. In contrast to existing reviews, in this survey,
we try to provide a more comprehensive taxonomy of question generation tasks from
three different perspectives, i.e., the types of the input context text, the target
answer, and the generated question. We take a deep look into existing models from
different dimensions to analyze their underlying ideas, major design principles,
and training strategies We compare these models through benchmark tasks to obtain
an empirical understanding of the existing techniques. Moreover, we discuss what is
missing in the current literature and what are the promising and desired future
directions." 1/31/2022 1/11/2025 14:33 1/11/2025 14:33 1/11/2025 14:33
Jan-43 1 40 ACM Trans. Inf. Syst.
en [Link] (Crossref)
C:\Users\Mahnoor\Zotero\storage\VZLVAYLI\Zhang et al. - 2022 - A Review
on Question Generation from Natural Language [Link]
X5L9MNGE conferencePaper 2024 "Ramesh, Aninditha; Agarwal, Arav; Doughty,
Jacob Arthur; Ramaneti, Ketan; Savelka, Jaromir; Sakr, Majd" A Benchmark for
Testing the Capabilities of LLMs in Assessing the Quality of Multiple-choice
Questions in Introductory Programming Education Proceedings of the 2024 on ACM
Virtual Global Computing Education Conference V. 1 979-8-4007-0598-4
10.1145/3649165.3690123 [Link] "There
has been a growing interest in utilizing large language models (LLMs) for numerous
educational applications. Recent studies have focused on the use of LLMs for
generating various educational artifacts for programming education, such as
programming exercises, model solutions, or multiple-choice questions (MCQs). The
ability to efficiently and reliably assess the quality of such artifacts, both
automatically and human generated, has become of paramount importance. Hence, there
is a pressing need to develop and make available robust benchmarks. In this paper,
we investigate an example use case of assessing the quality of programming MCQs. To
that end, we carefully curated a data set of 192 MCQs annotated with quality scores
based on a rubric that evaluates crucial aspects such as, e.g., their clarity, the
presence of a single correct answer, and the quality of distractors. The results
show that the task presents a considerable challenge even to the state-of-the-art
LLMs and, hence, further research is needed. To further such research efforts in
this important area we release the dataset as well as the extensible evaluation
pipeline to the public." 12/5/2024 1/11/2025 14:33 1/11/2025 14:33
1/11/2025 14:33 193-199
ACM Virtual Event NC USA en [Link]
(Crossref) C:\Users\Mahnoor\Zotero\storage\DLGHD2KH\Ramesh et
al. - 2024 - A Benchmark for Testing the Capabilities of LLMs in Assessing the
Quality of Multiple-choice [Link]
SIGCSE
Virtual 2024: 1st ACM Virtual Global Computing Education Conference
C28HHKD7 conferencePaper 2024 "Folajimi, Yetunde" From GPT to BERT:
Benchmarking Large Language Models for Automated Quiz Generation Proceedings of the
2024 on ACM Virtual Global Computing Education Conference V. 2 979-8-4007-0604-2
10.1145/3649409.3691090 [Link]
12/5/2024 1/11/2025 14:33 1/11/2025 14:33 1/11/2025 14:33 312-313
From GPT to BERT ACM
Virtual Event NC USA en [Link] (Crossref)
C:\Users\Mahnoor\Zotero\storage\VWQGJ8ZR\Folajimi - 2024 - From
GPT to BERT Benchmarking Large Language Models for Automated Quiz [Link]
SIGCSE Virtual 2024: 1st ACM Virtual Global Computing
Education Conference
323TZHCY conferencePaper 2024 "Pereira Cipriano, Bruno; Silva, Miguel;
Correia, Rodrigo; Alves, Pedro" Towards the Integration of Large Language
Models and Automatic Assessment Tools: Enhancing Student Support in Programming
Assignments Proceedings of the 24th Koli Calling International Conference on
Computing Education Research 979-8-4007-1038-4 10.1145/3699538.3699588
[Link] "The rise of Large Language
Models (LLMs) has sparked discussion in Computer Science Education (CSE) due to
their ability to generate code from text prompts. Students may rely on these tools,
neglecting core skills like computational thinking and program design. Thus, it’s
crucial to responsibly integrate them into computer science courses." 11/12/2024
1/11/2025 14:33 1/11/2025 14:33 1/11/2025 14:33 2-Jan
Towards the Integration of Large Language Models and Automatic
Assessment Tools ACM Koli Finland en
[Link] (Crossref) C:\Users\Mahnoor\Zotero\
storage\EGVAGR5F\Pereira Cipriano et al. - 2024 - Towards the Integration of Large
Language Models and Automatic Assessment Tools Enhancing Student [Link]
Koli Calling '24: 24th Koli Calling International Conference on
Computing Education Research
IAFPN5YR journalArticle 2024 "Hadzhikoleva, Stanka; Rachovski, Todor;
Ivanov, Ivan; Hadzhikolev, Emil; Dimitrov, Georgi" Automated Test Creation Using
Large Language Models: A Practical Application Applied Sciences 2076-3417
10.3390/app14199125 [Link] "The article
presents work on developing a software application for test creation using
artificial intelligence and large language models. Its main goal is to optimize the
educators’ work by automating the process of test generation and evaluation, with
the tests being stored for subsequent analysis and use. The application can
generate test questions based on specified criteria such as difficulty level,
Bloom’s taxonomy level, question type, style and format, feedback inclusion, and
more, thereby providing opportunities to enhance the adaptability and efficiency of
the learning process. It is developed on the Google Firebase platform, utilizing
the ChatGPT API, and also incorporates cloud computing to ensure scalability and
data reliability." 10/9/2024 1/11/2025 14:33 1/11/2025 14:33 1/11/2025
14:33 9125 19 14 Applied Sciences Automated Test Creation Using
Large Language Models en
[Link] [Link]
(Crossref) C:\Users\Mahnoor\Zotero\storage\CQH7UI2Y\Hadzhikoleva
et al. - 2024 - Automated Test Creation Using Large Language Models A Practical
[Link]
EDYH6IUX conferencePaper 2024 "Duong, Ta Nguyen Binh; Meng, Chai Yi"
Automatic Grading of Short Answers Using Large Language Models in Software
Engineering Courses 2024 IEEE Global Engineering Education Conference (EDUCON)
979-8-3503-9402-3 10.1109/EDUCON60312.2024.10578839
[Link] "Short-answer based questions
have been used widely due to their effectiveness in assessing whether the desired
learning outcomes have been attained by students. However, due to their open-ended
nature, many different answers could be considered entirely or partially correct
for the same question. In the context of computer science and software engineering
courses where the enrolment has been increasing recently, manual grading of
shortanswer questions is a time-consuming and tedious process for instructors."
5/8/2024 1/11/2025 14:33 1/11/2025 14:33 1/11/2025 14:33 10-Jan
IEEE "Kos Island,
Greece" en [Link] [Link]
(Crossref) C:\Users\Mahnoor\Zotero\storage\26TRM3ZF\Duong and
Meng - 2024 - Automatic Grading of Short Answers Using Large Language Models in
Software Engineering [Link]
2024 IEEE Global
Engineering Education Conference (EDUCON)
EF6YN9KH journalArticle "Hurani, Muaz; Idris, Hamzeh" "Investigating the
use of LLMs for automated test generation: challenges, benefits, and suitability."
"This thesis investigates the application of Large
Language Models (LLMs) in automated test generation for software development,
focusing on their challenges, benefits, and suitability for businesses. The study
employs a mixed-methods approach, combining a literature review with empirical
evaluations through surveys, interviews, and focus groups involving software
developers and testers. Key findings indicate that LLMs enhance the efficiency and
speed of test case generation, offering substantial improvements in test coverage
and reducing development costs. However, the integration of LLMs poses several
challenges, including technical complexities, the need for extensive customization,
and concerns about the quality and reliability of the generated test cases.
Additionally, ethical issues such as data biases and the potential impact on job
roles were highlighted. The results show that while LLMs excel in generating test
cases for routine tasks, their effectiveness diminishes in complex scenarios
requiring deep domain knowledge and intricate system interactions. The study
concludes that with proper training, continuous feedback, and iterative refinement,
LLMs can be effectively integrated into existing workflows to complement
traditional testing methods." 1/11/2025 14:33 1/11/2025 14:33
en
Zotero "C:\Users\Mahnoor\Zotero\storage\
4QXYWEXW\Hurani and Idris - Investigating the use of LLMs for automated test
generation challenges, benefits, and suitability..pdf"
CH7GKUSG journalArticle "Maity, Subhankar; Deroy, Aniket" The Future
of Learning in the Age of Gen- erative AI: Automated Question Genera- tion and
Assessment with Large Language Models
1/11/2025 14:33 1/11/2025 14:33
en Zotero
C:\Users\Mahnoor\Zotero\storage\G6DNEX3E\Maity and Deroy - The
Future of Learning in the Age of Gen- erative AI Automated Question Genera- tion
and [Link]
U5WITPPI journalArticle "Rakangor, Sheetal; Ghodasara, Dr Y R"
Literature Review of Automatic Question Generation Systems
"NLP is an area of research where many researcher have presented their
work and is still an area under research to achieve higher accuracy. Researcher is
going to use NLP for automatic question generation system. As in any education
system, examination is conducted to judge the caliber of the students. So to
conduct the examination, educator needs to generate the questions manually which is
very time consuming process, so researcher has decided to develop system through
which, automatic questions can be generated, and time and efforts will be reduced.
To develop automatic question generation, many researcher has presented their work
and many algorithms are proposed to generate the automatic question from given
sentence or text. This paper presents review of work to generate questions
automatic from inputted text. Paper reviews methodologies with respect to the
phases of question generation." 1/11/2025 14:33 1/11/2025 14:33
en
Zotero C:\Users\Mahnoor\Zotero\
storage\VQNZR4JG\Rakangor and Ghodasara - Literature Review of Automatic Question
Generation [Link]
CVS7NHH9 journalArticle Programming Question Generation: An
Automated Methodology for Generating Novel Programming Assignments with Varying
Difficulty Levels A comprehensive programming course
requires a clearly delineated collection of programming activities that can help
students improve their programming skills. Instructors devote considerable time to
preparing such problems and their corresponding solutions. They often resort to a
strategy of using previously developed materials from prior semesters and drawing
upon online resources to facilitate the preparation of course materials. We are
presenting a methodology for generating practice exercises that cater to students
with diverse levels of difficulty. The proposed approach is beneficial for
educators in developing instructional resources and personalised assessments
customised to individual students’ programming proficiency. 1/11/2025 14:33
1/11/2025 14:33
en Zotero
C:\Users\Mahnoor\Zotero\storage\T8MFLFPI\Programming Question Generation An
Automated Methodology for Generating Novel Programming [Link]
F8KVEJE3 journalArticle 2024 "Lee, Unggi; Jung, Haewon; Jeon, Younghoon;
Sohn, Younghoon; Hwang, Wonhee; Moon, Jewoong; Kim, Hyeoncheol" Few-shot is
enough: exploring ChatGPT prompt engineering method for automatic question
generation in english education Education and Information Technologies
"1360-2357, 1573-7608" 10.1007/s10639-023-12249-8
[Link] "Through design and
development research (DDR), we aimed to create a validated automatic question
generation (AQG) system using large language models (LLMs) like ChatGPT, enhanced
by prompting engineering techniques. While AQG has become increasingly integral to
online learning for its efficiency in generating questions, issues such as
inconsistent question quality and the absence of transparent and validated
evaluation methods persist. Our research focused on creating a prompt engineering
protocol tailored for AQG. This protocol underwent several iterations of refinement
and validation to improve its performance. By gathering validation scores and
qualitative feedback on the produced questions and the system’s framework, we
examined the effectiveness of the system. The study findings indicate that our
combined use of LLMs and prompt engineering in AQG produces questions with
statistically significant validity. Our research further illuminates academic and
design considerations for AQG design in English education: (a) certain question
types might not be optimal for generation via ChatGPT, (b) ChatGPT sheds light on
the potential for collaborative AI-teacher efforts in question generation,
especially within English education." 2024-06 1/11/2025 14:33 1/11/2025
14:33 1/11/2025 14:33 11483-11515 9 29 Educ Inf Technol Few-
shot is enough en
[Link] (Crossref) C:\Users\Mahnoor\Zotero\storage\
NH2RI6M4\Lee et al. - 2024 - Few-shot is enough exploring ChatGPT prompt
engineering method for automatic question generation [Link]
CZI9HXYX journalArticle 2023 "Mulla, Nikahat; Gharpure, Prachi" "Automatic
question generation: a review of methodologies, datasets, evaluation metrics, and
applications" Progress in Artificial Intelligence "2192-6352, 2192-6360"
10.1007/s13748-023-00295-9 [Link]
00295-9 "Question generation in natural language has a wide variety of
applications. It can be a helpful tool for chatbots for generating interesting
questions as also for automating the process of question generation from a piece of
text. Most modern-day systems, which are conversational, require question
generation ability for identifying the user’s needs and serving customers better.
Generating questions in natural language is now, a more evolved task, which also
includes generating questions for an image or video. In this review, we provide an
overview of the research progress in automatic question generation. We also present
a comprehensive literature review covering the classification of Question Generation
systems by categorizing them into three broad use-cases, namely standalone question
generation, visual question generation, and conversational question generation. We
next discuss the datasets available for the same for each use-case. We further
direct this review towards applications of question generation and discuss the
challenges in this field of research." 2023-03 1/11/2025 14:33 1/11/2025
14:33 1/11/2025 14:33 Jan-32 1 12 Prog Artif Intell
Automatic question generation en
[Link] (Crossref) "C:\Users\Mahnoor\
Zotero\storage\9A4FPFK9\Mulla and Gharpure - 2023 - Automatic question generation a
review of methodologies, datasets, evaluation metrics, and [Link]"
ZFI8SSBR journalArticle 2024 "Meißner, Niklas; Speth, Sandro; Kieslinger,
Julian; Becker, Steffen" EvalQuiz – LLM-based Automated Generation of Self-
Assessment Quizzes in Software Engineering Education
10.18420/SEUH2024_04 [Link] "Self-
assessment quizzes after lectures, educational videos, or chapters are a commonly
used method in software engineering (SE) education to allow students to test the
knowledge they have gained. However, creating these quizzes is time-consuming,
cognitively exhausting, and complex, as an expert in the field needs to create the
quizzes and review the lecture material for validity. Therefore, this paper
presents a concept to automatically generate self-assessment quizzes based on
lecture material using a large language model (LLM) to reduce lecturers’ workload
and simplify the general quiz creation process. The developed prototype was handed
to experts, who subsequently evaluated the approach. The results show that
automatic quiz generation saves time, and the quizzes cover the delivered lecture
material well. However, the generated quizzes often lack originality and
versatility. Therefore, further prompt engineering might be required to achieve
more elaborate results." 2024 1/11/2025 14:33 1/11/2025 14:33 1/11/2025
14:33
en [Link] (Datacite) ISBN:
9783885792550 Publisher: Gesellschaft für Informatik e.V. C:\Users\Mahnoor\
Zotero\storage\N64LKMG8\Meißner et al. - 2024 - EvalQuiz – LLM-based Automated
Generation of Self-Assessment Quizzes in Software Engineering [Link]
Automatic Question Generation; GPT-4; Prompt Engineering; Self-Assessment;
Software Engineering Education