Planned maintenance
A system upgrade is planned for 10/12-2024, at 12:00-13:00. During this time DiVA will be unavailable.
1213141516171815 of 23
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Exploring the Educational Utility of Pretrained Language Models
Stockholm University, Faculty of Social Sciences, Department of Computer and Systems Sciences.ORCID iD: 0000-0002-0945-707x
2024 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

The emergence of pretrained language models has profoundly reshaped natural language processing, serving as foundation models for a wide range of tasks. Over the past decade, pretrained language models have evolved significantly, leading to the development of different types of models and approaches for utilising them. This progression spans from static to contextual models and from smaller models to more powerful, generative large language models. The increasing capabilities of these models have, in turn, led to growing interest in exploring new use cases and applications across various domains, including education, where digitalisation has created opportunities for AI applications that leverage pretrained language models, particularly due to the abundance of text data in educational contexts.

This thesis explores the educational utility of pretrained language models, specifically by investigating how different paradigms of these models can be applied to address tasks in education. These paradigms include various methodologies for leveraging the knowledge embedded in pretrained language models, such as embeddings, fine-tuning, prompt-based learning, and in-context learning. For collaborative learning group formation, a clustering approach based on pretrained embeddings is proposed, enabling the creation of either homogeneous or heterogeneous groups depending on the specific learning situation. For automated essay scoring, a pretrained language model is fine-tuned using both the essay instructions and the essay text as input; the proposed method also highlights key topical sentences that contribute to the predicted essay score. For educational question generation, a method based on prompt-based learning is introduced and shown to be more data-efficient than existing methods. Finally, for educational question answering, certain limitations of the in-context learning (or prompting) paradigm, such as a tendency of large language models to hallucinate or miscalculate, are addressed. Specifically, workflows and prompting strategies based on retrieval-augmented generation and tool-augmented generation are proposed, allowing large language models to ground answers in specific learning materials and to leverage external tools, such as calculators and knowledge bases, within chain-of-thought reasoning processes. These strategies are shown to produce more reliable and transparent answers to complex questions.

Through five empirical studies, methodological innovations within each paradigm of pretrained language models are proposed and evaluated for specific educational use cases. In addition to contributing methodologically to natural language processing, the results demonstrate the potential utility of pretrained language models in educational AI applications, thereby advancing the field of technology enhanced learning. The proposed methods not only improve predictive performance on specific tasks but also aim to enhance the transparency of pretrained language models, which is essential for building reliable and trustworthy educational AI applications.

Place, publisher, year, edition, pages
Stockholm: Department of Computer and Systems Sciences, Stockholm University , 2024.
Series
Report Series / Department of Computer & Systems Sciences, ISSN 1101-8526 ; 24-017
Keywords [en]
Natural Language Processing, Technology Enhanced Learning, Pretrained Language Models, Large Language Models, Generative AI, Collaborative Learning, Automated Essay Scoring, Educational Question Generation, Educational Question Answering
National Category
Computer Sciences Language Technology (Computational Linguistics)
Research subject
Computer and Systems Sciences
Identifiers
URN: urn:nbn:se:su:diva-235084ISBN: 978-91-8107-014-9 (print)ISBN: 978-91-8107-015-6 (electronic)OAI: oai:DiVA.org:su-235084DiVA, id: diva2:1909361
Public defence
2024-12-16, L30, Nodhuset, Borgarfjordsgatan 12, Kista., 09:00 (English)
Opponent
Supervisors
Available from: 2024-11-21 Created: 2024-10-30 Last updated: 2024-11-11Bibliographically approved
List of papers
1. Catching Group Criteria Semantic Information When Forming Collaborative Learning Groups
Open this publication in new window or tab >>Catching Group Criteria Semantic Information When Forming Collaborative Learning Groups
Show others...
2021 (English)In: Technology-Enhanced Learning for a Free, Safe, and Sustainable World: 16th European Conference on Technology Enhanced Learning, EC-TEL 2021, Bolzano, Italy, September 20-24, 2021, Proceedings / [ed] Tinne De Laet; Roland Klemke; Carlos Alario-Hoyos; Isabel Hilliger; Alejandro Ortega-Arranz, Springer , 2021, p. 16-27Conference paper, Published paper (Refereed)
Abstract [en]

Collaborative learning has grown more popular as a form of instruction in recent decades, with a significant number of studies demonstrating its benefits from many perspectives of theory and methodology. However, it has also been demonstrated that effective collaborative learning does not occur spontaneously without orchestrating collaborative learning groups according to the provision of favourable group criteria. Researchers have investigated different foundations and strategies to form such groups. However, the group criteria semantic information, which is essential for classifying groups, has not been explored. To capture the group criteria semantic information, we propose a novel Natural Language Processing (NLP) approach, namely using pre-trained word embedding. Through our approach, we could automatically form homogeneous and heterogeneous collaborative learning groups based on student’s knowledge levels expressed in assessments. Experiments utilising a dataset from a university programming course are used to assess the performance of the proposed approach.

Place, publisher, year, edition, pages
Springer, 2021
Series
Lecture Notes in Computer Science (LNCS), ISSN 0302-9743, E-ISSN 1611-3349 ; 12884
Keywords
Collaborative learning group formation, Word embeddings, Natural Language Processing, Semantic information, Artificial Intelligence
National Category
Computer Sciences
Research subject
Computer and Systems Sciences
Identifiers
urn:nbn:se:su:diva-200604 (URN)10.1007/978-3-030-86436-1_2 (DOI)978-3-030-86436-1 (ISBN)978-3-030-86435-4 (ISBN)
Conference
European Conference on Technology Enhanced Learning, EC-TEL 2021, Bolzano, Italy, September 20-24, 2021
Available from: 2022-01-08 Created: 2022-01-08 Last updated: 2024-10-30Bibliographically approved
2. Beyond Benchmarks: Spotting Key Topical Sentences While Improving Automated Essay Scoring Performance with Topic-Aware BERT
Open this publication in new window or tab >>Beyond Benchmarks: Spotting Key Topical Sentences While Improving Automated Essay Scoring Performance with Topic-Aware BERT
Show others...
2023 (English)In: Electronics, E-ISSN 2079-9292, Vol. 12, no 1, article id 150Article in journal (Refereed) Published
Abstract [en]

Automated Essay Scoring (AES) automatically allocates scores to essays at scale and may help teachers reduce the heavy burden during grading activities. Recently, researchers have deployed neural-based AES approaches to improve upon the state-of-the-art AES performance. These neural-based AES methods mainly take student essays as the sole input and focus on learning the relationship between student essays and essay scores through deep neural networks. However, their only product, the predicted holistic score, is far from providing adequate pedagogical information, such as automated writing evaluation (AWE). In this work, we propose Topic-aware BERT, a new method of learning relations among scores, student essays, as well as topical information in essay instructions. Beyond improving the AES benchmark performance, Topic-aware BERT can automatically retrieve key topical sentences in student essays by probing self-attention maps in intermediate layers. We evaluate the performance of Topic-aware BERT of different variants to (i) perform AES and (ii) retrieve key topical sentences using the open dataset Automated Student Assessment Prize and a manually annotated dataset. Our experiments show that Topic-aware BERT achieves a strong AES performance compared with the previous best neural-based AES methods and demonstrates effectiveness in identifying key topical sentences in argumentative essays.

Keywords
Artificial Intelligence, Natural Language Processing, Automated Essay Scoring, Automated Writing Evaluation, BERT
National Category
Language Technology (Computational Linguistics)
Research subject
Computer and Systems Sciences
Identifiers
urn:nbn:se:su:diva-213557 (URN)10.3390/electronics12010150 (DOI)000910414500001 ()2-s2.0-85145830143 (Scopus ID)
Note

This article belongs to the Special Issue Artificial Intelligence Solutions and Applications for Distributed Systems in Smart Spaces

Available from: 2023-01-09 Created: 2023-01-09 Last updated: 2024-10-30Bibliographically approved
3. Towards Data-effective Educational Question Generation with Prompt-based Learning
Open this publication in new window or tab >>Towards Data-effective Educational Question Generation with Prompt-based Learning
Show others...
2023 (English)Conference paper, Published paper (Refereed)
Abstract [en]

Practice and exam-style questions, as essential educational tools, contribute to educators’ effective teaching. Automatic question generation (QG) is a promising technique that can eliminate the manual effort of constructing questions and boost technology-enhanced education systems. Recently, deep neural network-based question-generation approaches have significantly improved upon state-of-the-art of question generation. Nevertheless, these approaches are often developed based on huge and non-educational datasets consisting of over 100,000 examples, which negatively affect the scalability and reliability of the educational QG systems. This study proposes a prompt-based learning QG approach that could generate questions in a data-effective way. The proposed prompt-based learning QG approach is trained and evaluated on a general dataset SQuAD, and an educational dataset SciQ. Experiment results demonstrate that our approach outperforms existing best QG models by a vast margin in data-effective scenarios and could generate high-quality educational questions with as few as 1,000 training examples.

Place, publisher, year, edition, pages
Springer Nature, 2023
Series
Lecture Notes in Networks and Systems, ISSN 2367-3370, E-ISSN 2367-3389 ; 711
Keywords
Question Generation, Natual Language Processing, Artificial Intelligence, Prompt-based Learning
National Category
Information Systems
Research subject
Computer and Systems Sciences
Identifiers
urn:nbn:se:su:diva-224808 (URN)10.1007/978-3-031-37717-4_11 (DOI)2-s2.0-85174674904 (Scopus ID)978-3-031-37716-7 (ISBN)
Conference
Computing Conference 2023
Available from: 2023-12-27 Created: 2023-12-27 Last updated: 2024-10-30Bibliographically approved
4. Towards Improving the Reliability and Transparency of ChatGPT for Educational Question Answering
Open this publication in new window or tab >>Towards Improving the Reliability and Transparency of ChatGPT for Educational Question Answering
2023 (English)In: LNCS Springer Conference Proceedings, Springer, 2023Conference paper, Published paper (Refereed)
Abstract [en]

Large language models (LLMs), such as ChatGPT, have shown remarkable performance on various natural language processing (NLP) tasks, including educational question answering (EQA). However, LLMs generate text entirely based on knowledge obtained during pre-training, which means they struggle with recent information or domain-specific knowledge bases. Moreover, only providing answers to questions posed to LLMs without any grounding materials makes it difficult for students to judge their validity.

We therefore propose a method for integrating information retrieval systems with LLMs when developing EQA systems, which in addition to improving EQA performance grounds the answers in the educational context. Our experiments show that the proposed system outperforms vanilla ChatGPT with a vast margin of 110.9%, 67.8%, and 43.3% on BLEU, ROUGE, and METEOR scores. In addition, we argue that the use of the retrieved educational context enhances the transparency and reliability of the EQA process, making it easier to determine the correctness of the answers.

Place, publisher, year, edition, pages
Springer, 2023
Series
Lecture Notes in Computer Science (LNCS), ISSN 0302-9743, E-ISSN 1611-3349 ; 14200
Keywords
AI NLP ChatGPT LLMs Educational Question Answering
National Category
Computer Sciences
Research subject
Computer and Systems Sciences
Identifiers
urn:nbn:se:su:diva-224809 (URN)10.1007/978-3-031-42682-7_32 (DOI)2-s2.0-85171972992 (Scopus ID)978-3-031-42681-0 (ISBN)
Conference
Eighteenth European Conference on Technology Enhanced Learning EC-TEL, 2023
Available from: 2023-12-27 Created: 2023-12-27 Last updated: 2024-10-30Bibliographically approved
5. Selecting from Multiple Strategies Improves the Foreseeable Reasoning of Tool-Augmented Large Language Models
Open this publication in new window or tab >>Selecting from Multiple Strategies Improves the Foreseeable Reasoning of Tool-Augmented Large Language Models
2024 (English)In: Joint European Conference on Machine Learning and Knowledge Discovery in Databases, 2024Conference paper, Published paper (Refereed)
Abstract [en]

Large language models (LLMs) can be augmented by interacting with external tools and knowledge bases, allowing them to overcome some of their known limitations, such as not having access to up-to-date information or struggling to solve math problems, thereby going beyond the knowledge and capabilities obtained during pre-training. Recent prompting techniques have enabled tool-augmented LLMs to combine reasoning and action to solve complex problems with the help of tools. This is essential for allowing LLMs to strategically determine the timing and nature of tool-calling actions in order to enhance their decision-making process and improve their outputs. However, the reliance of current prompting techniques on a single reasoning path or their limited ability to adjust plans within that path can adversely impact the performance of tool-augmented LLMs. In this paper, we introduce a novel prompting method, whereby an LLM agent selects and executes one among multiple candidate strategies. We assess the effectiveness of our method on three question answering datasets, on which it outperforms state-of-the-art methods like ReWOO, while also being a competitive and more cost-efficient alternative to ReAct. We also investigate the impact of selecting a reasoning trajectory from different strategy pool sizes, further highlighting the risks in only considering a single strategy.

Keywords
Large language models, Tool-augmented language models, Chain-of-thought prompting, Question answering.
National Category
Computer Systems
Research subject
Computer and Systems Sciences; Computational Linguistics
Identifiers
urn:nbn:se:su:diva-235082 (URN)10.1007/978-3-031-70352-2_12 (DOI)978-3-031-70351-5 (ISBN)
Conference
Machine Learning and Knowledge Discovery in Databases. Research Track. ECML PKDD 2024.
Available from: 2024-10-30 Created: 2024-10-30 Last updated: 2024-10-30

Open Access in DiVA

Exploring the Educational Utility of Pretrained Language Models(3509 kB)48 downloads
File information
File name FULLTEXT01.pdfFile size 3509 kBChecksum SHA-512
c7038fd10c092dc3b3cbf000aa5eebbb146b5a18b2a5e3ee3557781153dc28a792eb9e4ec2a7fb7daaa70e31420c755fb60281d9eee5714385573f940c4bc181
Type fulltextMimetype application/pdf

Authority records

Wu, Yongchao

Search in DiVA

By author/editor
Wu, Yongchao
By organisation
Department of Computer and Systems Sciences
Computer SciencesLanguage Technology (Computational Linguistics)

Search outside of DiVA

GoogleGoogle Scholar
Total: 48 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

isbn
urn-nbn

Altmetric score

isbn
urn-nbn
Total: 262 hits
1213141516171815 of 23
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf