Por favor, use este identificador para citar o enlazar este ítem: https://doi.org/10.1109/ACCESS.2024.3372853

Registro completo de metadatos
Campo DCValorLengua/Idioma
dc.contributor.authorMartínez Sánchez, Pablo Antonio-
dc.contributor.authorBernabé García, Gregorio-
dc.contributor.authorGarcía Carrasco, José Manuel-
dc.contributor.otherFacultades, Departamentos, Servicios y Escuelas::Departamentos de la UMU::Ingeniería y Tecnología de Computadoreses
dc.date.accessioned2024-03-26T09:19:19Z-
dc.date.available2024-03-26T09:19:19Z-
dc.date.issued2024-03-01-
dc.identifier.citationIEEE Access. Volumen 12, 2024es
dc.identifier.issnElectronic: 2169-3536-
dc.identifier.urihttp://hdl.handle.net/10201/140462-
dc.description©2024. This manuscript version is made available under the CC-BY-NC-ND 4.0 license http://creativecommons.org/licenses/by-nc-nd/4.0/ This document is the published, version of a Published Work that appeared in final form in IEEE Access. To access the final edited and published work see https://doi.org/10.1109/ACCESS.2024.3372853es
dc.description.abstractLarge language models (LLMs) have been massively applied to many tasks, often surpassing state-of-the-art approaches. While their effectiveness in code generation has been extensively studied (e.g., AlphaCode), their potential for code detection remains unexplored. This work presents the first analysis of code detection using LLMs. Our study examines essential kernels, including matrix multiplication, convolution, fast-fourier transform and LU factorization, implemented in C/C++. We propose both a preliminary, naive prompt and a novel prompting strategy for code detection. Results reveal that conventional prompting achieves great precision but poor accuracy (67.5%, 22.5%, 79.5% and 64% for GEMM, convolution, FFT and LU factorization, respectively) due to a high number of false positives. Our novel prompting strategy substantially reduces false positives, resulting in excellent overall accuracy (91.2%, 98%, 99.7% and 99.7%, respectively). These results pose a considerable challenge to existing state-of-the-art code detection methods.es
dc.formatapplication/pdfes
dc.format.extent11es
dc.languageenges
dc.relationThis work was supported in part by the Ministerio de Ciencia e Innovación (MCIN)/Agencia Estatal de Investigación (AEI)/10.13039/501100011033 under Grant TED2021-129221B-I00 and Grant PID2022-136315OB-I00; in part by ‘‘European Union (EU) NextGenerationEU/Plan de Recuperación, Transformación y Resiliencia (PRTR);’’ and in part by ‘‘European Regional Development Fund (ERDF) A way of making Europe,’’ EU.es
dc.rightsinfo:eu-repo/semantics/openAccesses
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/4.0/*
dc.subjectCode detectiones
dc.subjectCompilerses
dc.subjectHeterogeneous computinges
dc.subjectHigh-performance computinges
dc.subjectLarge language modeles
dc.titleCode Detection for Hardware Acceleration Using Large Language Modelses
dc.typeinfo:eu-repo/semantics/articlees
dc.identifier.doihttps://doi.org/10.1109/ACCESS.2024.3372853-
Aparece en las colecciones:Artículos: Ingeniería y Tecnología de Computadores

Ficheros en este ítem:
Fichero Descripción TamañoFormato 
IEEE24.pdf887,05 kBAdobe PDFVista previa
Visualizar/Abrir


Este ítem está sujeto a una licencia Creative Commons Licencia Creative Commons Creative Commons