Enhancing Trustworthiness in LLM-Generated Code: A Reinforcement Learning and Domain-Knowledge Constrained Approach
| dc.contributor.author | Piplai, Aritran | |
| dc.contributor.author | Kotal, Anantaa | |
| dc.contributor.author | Mittal, Sudip | |
| dc.contributor.author | Joshi, Karuna | |
| dc.contributor.author | Finin, Tim | |
| dc.contributor.author | Joshi, Anupam | |
| dc.date.accessioned | 2026-03-26T14:26:55Z | |
| dc.date.issued | 2025-02 | |
| dc.description | UMBC CODEBOT '25 Workshop, Columbia, MD , 25-26 February 2025 | |
| dc.description.abstract | Imagine analyzing a piece of code that uses the function ConnectToServer() withan encrypted string as its argument. A large language model (LLM), trained onextensive programming data, might flag the use of encryption as suspicious andgenerate an explanation suggesting that the function likely connects to a maliciousserver. While this explanation might seem plausible, it can often be unfaithful—itovergeneralizes based on statistical patterns from its training data without trulyunderstanding the context or validating its claims [8]. A REACT (Reasoning andActing) framework, which combines reasoning with action steps, is likely a betterapproach because it allows the LLM to propose actions—such as decrypting the stringor examining server connections—while reasoning about the results [7]. However,REACT still lacks a feedback mechanism to evaluate the effectiveness of thoseactions or iteratively refine the sequence based on empirical observations. Without such feedback, it risks falling short in dynamic scenarios, where the validation of predictions and adaptation to new evidence are critical [10]. | |
| dc.description.uri | https://ebiquity.umbc.edu/_file_directory_/papers/1428.pdf | |
| dc.format.extent | 4 pages | |
| dc.genre | conference papers and proceedings | |
| dc.genre | preprints | |
| dc.identifier.uri | http://hdl.handle.net/11603/42288 | |
| dc.language.iso | en | |
| dc.relation.isAvailableAt | The University of Maryland, Baltimore County (UMBC) | |
| dc.relation.ispartof | UMBC Computer Science and Electrical Engineering Department | |
| dc.relation.ispartof | UMBC Faculty Collection | |
| dc.relation.ispartof | UMBC Center for Cybersecurity | |
| dc.relation.ispartof | UMBC Information Systems Department | |
| dc.rights | This item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author. | |
| dc.subject | UMBC Ebiquity Researh Group | |
| dc.subject | UMBC Accelerated Cognitive Cybersecurity Laboratory | |
| dc.subject | UMBC Cybersecurity Institute | |
| dc.title | Enhancing Trustworthiness in LLM-Generated Code: A Reinforcement Learning and Domain-Knowledge Constrained Approach | |
| dc.type | Text | |
| dcterms.creator | https://orcid.org/0000-0002-6354-1686 | |
| dcterms.creator | https://orcid.org/0000-0002-6593-1792 | |
| dcterms.creator | https://orcid.org/0000-0002-8641-3193 |
Files
Original bundle
1 - 1 of 1
