Mostrar el registro sencillo del ítem

dc.contributor.authorPulari, Sini Raj
dc.contributor.authorUmadevi, Maramreddy
dc.contributor.authorVasudevan, Shriram K.
dc.date2025-03-01
dc.date.accessioned2026-03-11T09:47:13Z
dc.date.available2026-03-11T09:47:13Z
dc.identifier.citationS. R. Pulari, M. Umadevi, S. K. Vasudevan. Improved Fine-Tuned Reinforcement Learning From Human Feedback Using Prompting Methods for News Summarization, International Journal of Interactive Multimedia and Artificial Intelligence, vol. 9, no. 2, pp. 59-67, 2025, http://dx.doi.org/10.9781/ijimai.2025.02.001es_ES
dc.identifier.urihttps://reunir.unir.net/handle/123456789/19232
dc.description.abstractChatGPT uses a generative pretrained transformer neural network model, which is under the larger umbrella of generative models. One major boom after ChatGPT is the advent of prompt engineering, which is the most critical part of ChatGPT that utilizes Large Language Models (LLM) and helps ChatGPT provide the desired outputs based on the style and tone of interactions carried out with it. Reinforcement learning from human feedback (RLHF) was used as the major aspect for fine-tuning LLM-based models. This work proposes a human selection strategy that is incorporated in the RLHF process to prevent undesirable consequences of the rightful choice of human reviewers for feedback. H-Rouge is a new metric proposed for humanized AI systems. A detailed evaluation of State-of-the-art summarization algorithms and prompt-based methods have been provided as part of the article. The proposed methods have introduced a strategy for human selection of RLHF models which employs multi-objective optimization to balance various goals encountered during the process with H-Rouge. This article will help nuance readers conduct research in the field of text summarization to start with prompt engineering in the summarization field, and future work will help them proceed in the right direction of research.es_ES
dc.language.isoenges_ES
dc.publisherUNIRes_ES
dc.relation.urihttps://www.ijimai.org/index.php/ijimai/article/view/259es_ES
dc.rightsopenAccesses_ES
dc.subjectAbstractive Summarizationes_ES
dc.subjectExtractive Summarizationes_ES
dc.subjectNatural Language Processinges_ES
dc.subjectNews Summarizationes_ES
dc.subjectPrompt Engineeringes_ES
dc.subjectReinforcement Learning From Human Feedback (RLHF)es_ES
dc.titleImproved Fine-Tuned Reinforcement Learning From Human Feedback Using Prompting Methods for News Summarizationes_ES
dc.typearticlees_ES
reunir.tag~IJIMAIes_ES
dc.identifier.doihttps://doi.org/10.9781/ijimai.2025.02.001


Ficheros en el ítem

Thumbnail

Este ítem aparece en la(s) siguiente(s) colección(ones)

Mostrar el registro sencillo del ítem