University Links: Home Page | Site Map
Covenant University Repository

ENHANCED IN-CONTEXT LEARNING FOR CODE ANALYSIS WITH COMPACT LARGE LANGUAGE MODELS AND MONTE CARLO TREE SEARCH

AJIBOLA, Ifedayo Olabode and Covenant University, Theses Masters (2024) ENHANCED IN-CONTEXT LEARNING FOR CODE ANALYSIS WITH COMPACT LARGE LANGUAGE MODELS AND MONTE CARLO TREE SEARCH. Masters thesis, Covenant University.

[img] PDF
Download (174kB)

Abstract

Large Language Models (LLMs) demonstrate impressive reasoning abilities, yet their performance can falter when dealing with extensive context lengths. Techniques like Retrieval Augmented Generation (RAG) and Chain-of-Thought prompting seek to bridge this gap, but they face limitations when applied to large code-based contexts due to the complexity of representing inter-object relationships. Monte Carlo Tree Search (MCTS), a heuristic search algorithm, offers a potential solution by aiding LLMs in identifying crucial code repository aspects, thus facilitating downstream tasks. This research focuses on applying MCTS to enhance the performance of "Compact LLMs" - models small enough to run inference on consumer-grade GPUs. Our findings confirm that MCTS indeed boosts performance compared to the baseline Compact LLM. However, these compact models, even with MCTS, still lag behind larger models in performance.

Item Type: Thesis (Masters)
Uncontrolled Keywords: large language model, compact LLM, chain-of-thought, monte carlo tree search, in-context learning, code analysis
Subjects: Q Science > QA Mathematics > QA75 Electronic computers. Computer science
Q Science > QA Mathematics > QA76 Computer software
Divisions: Faculty of Engineering, Science and Mathematics > School of Electronics and Computer Science
Depositing User: Patricia Nwokealisi
Date Deposited: 30 Oct 2024 11:54
Last Modified: 30 Oct 2024 11:54
URI: http://eprints.covenantuniversity.edu.ng/id/eprint/18544

Actions (login required)

View Item View Item