Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (464)

Search Parameters:
Keywords = GPT model

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 37868 KiB  
Article
3D Character Animation and Asset Generation Using Deep Learning
by Vlad-Constantin Lungu-Stan and Irina Georgiana Mocanu
Appl. Sci. 2024, 14(16), 7234; https://doi.org/10.3390/app14167234 (registering DOI) - 16 Aug 2024
Viewed by 379
Abstract
Besides video content, a significant part of entertainment is represented by computer games and animations such as cartoons. Creating such entertainment is based on two fundamental steps: asset generation and character animation. The main problem stems from its repetitive nature and the needed [...] Read more.
Besides video content, a significant part of entertainment is represented by computer games and animations such as cartoons. Creating such entertainment is based on two fundamental steps: asset generation and character animation. The main problem stems from its repetitive nature and the needed amounts of concentration and skill. The latest advances in deep learning and generative techniques have provided a set of powerful tools which can be used to alleviate these problems by facilitating the tasks of artists and engineers and providing a better workflow. In this work we explore practical solutions for facilitating and hastening the creative process: character animation and asset generation. In character animation, the task is to either move the joints of a subject manually or to correct the noisy data coming out of motion capture. The main difficulties of these tasks are their repetitive nature and the needed amounts of concentration and skill. For the animation case, we propose two decoder-only transformer based solutions, inspired by the current success of GPT. The first, AnimGPT, targets the original animation workflow by predicting the next pose of an animation based on a set of previous poses, while the second, DenoiseAnimGPT, tackles the motion capture case by predicting the clean current pose based on all previous poses and the current noisy pose. Both models obtained good performances on the CMU motion dataset, with the generated results being imperceptible to the untrained human eye. Quantitative evaluation was performed using mean absolute error between the ground truth motion vectors and the predicted motion vector. For both networks AnimGPT and DenoiseAnimGPT errors were 0.345, respectively 0.2513 (for 50 frames) that indicates better performances compared with other solutions. For asset generation, diffusion models were used. Using image generation and outpainting, we created a method that generates good backgrounds by combining the idea of text conditioned generation and text conditioned image editing. A time coherent algorithm that creates animated effects for characters was obtained. Full article
(This article belongs to the Special Issue Applications of Artificial Intelligence and Machine Learning in Games)
Show Figures

Figure 1

13 pages, 1004 KiB  
Article
Assessment Study of ChatGPT-3.5’s Performance on the Final Polish Medical Examination: Accuracy in Answering 980 Questions
by Julia Siebielec, Michal Ordak, Agata Oskroba, Anna Dworakowska and Magdalena Bujalska-Zadrozny
Healthcare 2024, 12(16), 1637; https://doi.org/10.3390/healthcare12161637 - 16 Aug 2024
Viewed by 289
Abstract
Background/Objectives: The use of artificial intelligence (AI) in education is dynamically growing, and models such as ChatGPT show potential in enhancing medical education. In Poland, to obtain a medical diploma, candidates must pass the Medical Final Examination, which consists of 200 questions with [...] Read more.
Background/Objectives: The use of artificial intelligence (AI) in education is dynamically growing, and models such as ChatGPT show potential in enhancing medical education. In Poland, to obtain a medical diploma, candidates must pass the Medical Final Examination, which consists of 200 questions with one correct answer per question, is administered in Polish, and assesses students’ comprehensive medical knowledge and readiness for clinical practice. The aim of this study was to determine how ChatGPT-3.5 handles questions included in this exam. Methods: This study considered 980 questions from five examination sessions of the Medical Final Examination conducted by the Medical Examination Center in the years 2022–2024. The analysis included the field of medicine, the difficulty index of the questions, and their type, namely theoretical versus case-study questions. Results: The average correct answer rate achieved by ChatGPT for the five examination sessions hovered around 60% and was lower (p < 0.001) than the average score achieved by the examinees. The lowest percentage of correct answers was in hematology (42.1%), while the highest was in endocrinology (78.6%). The difficulty index of the questions showed a statistically significant correlation with the correctness of the answers (p = 0.04). Questions for which ChatGPT-3.5 provided incorrect answers had a lower (p < 0.001) percentage of correct responses. The type of questions analyzed did not significantly affect the correctness of the answers (p = 0.46). Conclusions: This study indicates that ChatGPT-3.5 can be an effective tool for assisting in passing the final medical exam, but the results should be interpreted cautiously. It is recommended to further verify the correctness of the answers using various AI tools. Full article
(This article belongs to the Special Issue Artificial Intelligence in Healthcare: Opportunities and Challenges)
Show Figures

Figure 1

14 pages, 3790 KiB  
Article
Accuracy Evaluation of GPT-Assisted Differential Diagnosis in Emergency Department
by Fatemeh Shah-Mohammadi and Joseph Finkelstein
Diagnostics 2024, 14(16), 1779; https://doi.org/10.3390/diagnostics14161779 - 15 Aug 2024
Viewed by 242
Abstract
In emergency department (ED) settings, rapid and precise diagnostic evaluations are critical to ensure better patient outcomes and efficient healthcare delivery. This study assesses the accuracy of differential diagnosis lists generated by the third-generation ChatGPT (ChatGPT-3.5) and the fourth-generation ChatGPT (ChatGPT-4) based on [...] Read more.
In emergency department (ED) settings, rapid and precise diagnostic evaluations are critical to ensure better patient outcomes and efficient healthcare delivery. This study assesses the accuracy of differential diagnosis lists generated by the third-generation ChatGPT (ChatGPT-3.5) and the fourth-generation ChatGPT (ChatGPT-4) based on electronic health record notes recorded within the first 24 h of ED admission. These models process unstructured text to formulate a ranked list of potential diagnoses. The accuracy of these models was benchmarked against actual discharge diagnoses to evaluate their utility as diagnostic aids. Results indicated that both GPT-3.5 and GPT-4 reasonably accurately predicted diagnoses at the body system level, with GPT-4 slightly outperforming its predecessor. However, their performance at the more granular category level was inconsistent, often showing decreased precision. Notably, GPT-4 demonstrated improved accuracy in several critical categories that underscores its advanced capabilities in managing complex clinical scenarios. Full article
(This article belongs to the Special Issue AI-Assisted Diagnostics in Telemedicine and Digital Health)
Show Figures

Figure 1

26 pages, 10272 KiB  
Article
Utilizing Large Language Models to Illustrate Constraints for Construction Planning
by Chuanni He, Bei Yu, Min Liu, Lu Guo, Li Tian and Jianfeng Huang
Buildings 2024, 14(8), 2511; https://doi.org/10.3390/buildings14082511 - 14 Aug 2024
Viewed by 529
Abstract
Effective construction project planning relies on addressing constraints related to materials, labor, equipment, and others. Planning meetings are typical venues for stakeholders to identify, communicate, and remove constraints. However, a critical gap exists in lacking an automated approach to identify, classify, analyze, and [...] Read more.
Effective construction project planning relies on addressing constraints related to materials, labor, equipment, and others. Planning meetings are typical venues for stakeholders to identify, communicate, and remove constraints. However, a critical gap exists in lacking an automated approach to identify, classify, analyze, and track constraint discussions during onsite planning meetings. Therefore, this research aims to 1. develop a natural language processing model to classify constraints in meeting discussions; 2. uncover the discussion patterns of managers and foremen regarding various constraints; and 3. extract the root causes for constraints, evaluate their impacts, and prepare managers to develop practical solutions for constraint removal. This research collected meeting transcripts from 94 onsite planning meetings of a building project, spanning 263,836 words. Next, this research leveraged a general pretrained transformer (GPT) to segment discussion dialogs into topics. A Bidirectional Encoder Representations from Transformers (BERT)-based model was developed to categorize constraint types for each topic. The constraint patterns among meeting attendees were assessed. Furthermore, a GPT-based tool was devised to track root causes, impacts, and solutions for various constraints. Test results revealed an 8.8% improvement in constraint classification accuracy compared with the traditional classification model. An occupational characteristic in constraint discussion was observed in that the management team tended to balance their focus on various constraints, while foremen concentrated on more practical issues. This research contributes to the body of knowledge by leveraging language models to analyze construction planning meetings. The findings facilitate project managers in establishing constraint logs for diagnosing and prognosticating planning issues. Full article
(This article belongs to the Section Construction Management, and Computers & Digitization)
Show Figures

Figure 1

30 pages, 7039 KiB  
Article
A GPT-Powered Assistant for Real-Time Interaction with Building Information Models
by David Fernandes, Sahej Garg, Matthew Nikkel and Gursans Guven
Buildings 2024, 14(8), 2499; https://doi.org/10.3390/buildings14082499 - 13 Aug 2024
Viewed by 574
Abstract
This study introduces DAVE (Digital Assistant for Virtual Engineering), a Generative Pre-trained Transformer (GPT)-powered digital assistant prototype, designed to enable real-time, multimodal interactions within Building Information Modeling (BIM) environments for updating and querying BIM models using text or voice commands. DAVE integrates directly [...] Read more.
This study introduces DAVE (Digital Assistant for Virtual Engineering), a Generative Pre-trained Transformer (GPT)-powered digital assistant prototype, designed to enable real-time, multimodal interactions within Building Information Modeling (BIM) environments for updating and querying BIM models using text or voice commands. DAVE integrates directly with Autodesk Revit through Python scripts, the Revit API, and the OpenAI API and utilizes Natural Language Processing (NLP). This study presents (1) the development of a practical AI chatbot application that leverages conversational AI and BIM for dynamic actions within BIM models (e.g., updates and queries) at any stage of a construction project and (2) the demonstration of real-time, multimodal BIM model management through voice or text, which aims to reduce the complexity and technical barriers typically associated with BIM processes. The details of DAVE’s development and system architecture are outlined in this paper. Additionally, the comprehensive process of prototype testing and evaluation including the response time analysis and error analysis, which investigated the issues encountered during system validation, are detailed. The prototype demonstrated 94% success in accurately processing and executing single-function user queries. By enabling conversational interactions with BIM models, DAVE represents a significant contribution to the current body of knowledge. Full article
Show Figures

Graphical abstract

15 pages, 382 KiB  
Article
Domain Adaptation for Arabic Machine Translation: Financial Texts as a Case Study
by Emad A. Alghamdi, Jezia Zakraoui and Fares A. Abanmy
Appl. Sci. 2024, 14(16), 7088; https://doi.org/10.3390/app14167088 - 13 Aug 2024
Viewed by 385
Abstract
Neural machine translation (NMT) has shown impressive performance when trained on large-scale corpora. However, generic NMT systems have demonstrated poor performance on out-of-domain translation. To mitigate this issue, several domain adaptation methods have recently been proposed which often lead to better translation quality [...] Read more.
Neural machine translation (NMT) has shown impressive performance when trained on large-scale corpora. However, generic NMT systems have demonstrated poor performance on out-of-domain translation. To mitigate this issue, several domain adaptation methods have recently been proposed which often lead to better translation quality than genetic NMT systems. While there has been some continuous progress in NMT for English and other European languages, domain adaption in Arabic has received little attention in the literature. The current study, therefore, aims to explore the effectiveness of domain-specific adaptation for Arabic MT (AMT), in yet unexplored domain, financial news articles. To this end, we developed a parallel corpus for Arabic-English (AR-EN) translation in the financial domain to benchmark different domain adaptation methods. We then fine-tuned several pre-trained NMT and Large Language models including ChatGPT-3.5 Turbo on our dataset. The results showed that fine-tuning pre-trained NMT models on a few well-aligned in-domain AR-EN segments led to noticeable improvement. The quality of ChatGPT translation was superior to other models based on automatic and human evaluations. To the best of our knowledge, this is the first work on fine-tuning ChatGPT towards financial domain transfer learning. To contribute to research in domain translation, we made our datasets and fine-tuned models available. Full article
Show Figures

Figure 1

14 pages, 5936 KiB  
Article
GeoLocator: A Location-Integrated Large Multimodal Model (LMM) for Inferring Geo-Privacy
by Yifan Yang, Siqin Wang, Daoyang Li, Shuju Sun and Qingyang Wu
Appl. Sci. 2024, 14(16), 7091; https://doi.org/10.3390/app14167091 - 13 Aug 2024
Viewed by 439
Abstract
To ensure the sustainable development of artificial intelligence (AI) application in urban and geospatial science, it is important to protect the geographic privacy, or geo-privacy, which refers to an individual’s geographic location details. As a crucial aspect of personal security, geo-privacy plays a [...] Read more.
To ensure the sustainable development of artificial intelligence (AI) application in urban and geospatial science, it is important to protect the geographic privacy, or geo-privacy, which refers to an individual’s geographic location details. As a crucial aspect of personal security, geo-privacy plays a key role not only in individual protection but also in maintaining ethical standards in geoscientific practices. Despite its importance, geo-privacy is often not sufficiently addressed in daily activities. With the increasing use of large multimodal models (LMMs) such as GPT-4 for open-source intelligence (OSINT), the risks related to geo-privacy breaches have significantly escalated. This study introduces a novel GPT-4-based model, GeoLocator, integrated with location capabilities, and conducts four experiments to evaluate its ability to accurately infer location information from images and social media content. The results demonstrate that GeoLocator can generate specific geographic details with high precision, thereby increasing the potential for inadvertent exposure of sensitive geospatial information. This highlights the dual challenges posed by online data-sharing and information-gathering technologies in the context of geo-privacy. We conclude with a discussion on the broader impacts of GeoLocator and our findings on individuals and communities, emphasizing the urgent need for increased awareness and protective measures against geo-privacy breaches in the era of advancing AI and widespread social media usage. This contribution thus advocates for sustainable and responsible geoscientific practices. Full article
(This article belongs to the Special Issue Artificial Intelligence and the Future of Smart Cities)
Show Figures

Figure 1

15 pages, 265 KiB  
Article
Leveraging Large Language Models for Precision Monitoring of Chemotherapy-Induced Toxicities: A Pilot Study with Expert Comparisons and Future Directions
by Oskitz Ruiz Sarrias, María Purificación Martínez del Prado, María Ángeles Sala Gonzalez, Josune Azcuna Sagarduy, Pablo Casado Cuesta, Covadonga Figaredo Berjano, Elena Galve-Calvo, Borja López de San Vicente Hernández, María López-Santillán, Maitane Nuño Escolástico, Laura Sánchez Togneri, Laura Sande Sardina, María Teresa Pérez Hoyos, María Teresa Abad Villar, Maialen Zabalza Zudaire and Onintza Sayar Beristain
Cancers 2024, 16(16), 2830; https://doi.org/10.3390/cancers16162830 - 12 Aug 2024
Viewed by 419
Abstract
Introduction: Large Language Models (LLMs), such as the GPT model family from OpenAI, have demonstrated transformative potential across various fields, especially in medicine. These models can understand and generate contextual text, adapting to new tasks without specific training. This versatility can revolutionize clinical [...] Read more.
Introduction: Large Language Models (LLMs), such as the GPT model family from OpenAI, have demonstrated transformative potential across various fields, especially in medicine. These models can understand and generate contextual text, adapting to new tasks without specific training. This versatility can revolutionize clinical practices by enhancing documentation, patient interaction, and decision-making processes. In oncology, LLMs offer the potential to significantly improve patient care through the continuous monitoring of chemotherapy-induced toxicities, which is a task that is often unmanageable for human resources alone. However, existing research has not sufficiently explored the accuracy of LLMs in identifying and assessing subjective toxicities based on patient descriptions. This study aims to fill this gap by evaluating the ability of LLMs to accurately classify these toxicities, facilitating personalized and continuous patient care. Methods: This comparative pilot study assessed the ability of an LLM to classify subjective toxicities from chemotherapy. Thirteen oncologists evaluated 30 fictitious cases created using expert knowledge and OpenAI’s GPT-4. These evaluations, based on the CTCAE v.5 criteria, were compared to those of a contextualized LLM model. Metrics such as mode and mean of responses were used to gauge consensus. The accuracy of the LLM was analyzed in both general and specific toxicity categories, considering types of errors and false alarms. The study’s results are intended to justify further research involving real patients. Results: The study revealed significant variability in oncologists’ evaluations due to the lack of interaction with fictitious patients. The LLM model achieved an accuracy of 85.7% in general categories and 64.6% in specific categories using mean evaluations with mild errors at 96.4% and severe errors at 3.6%. False alarms occurred in 3% of cases. When comparing the LLM’s performance to that of expert oncologists, individual accuracy ranged from 66.7% to 89.2% for general categories and 57.0% to 76.0% for specific categories. The 95% confidence intervals for the median accuracy of oncologists were 81.9% to 86.9% for general categories and 67.6% to 75.6% for specific categories. These benchmarks highlight the LLM’s potential to achieve expert-level performance in classifying chemotherapy-induced toxicities. Discussion: The findings demonstrate that LLMs can classify subjective toxicities from chemotherapy with accuracy comparable to expert oncologists. The LLM achieved 85.7% accuracy in general categories and 64.6% in specific categories. While the model’s general category performance falls within expert ranges, specific category accuracy requires improvement. The study’s limitations include the use of fictitious cases, lack of patient interaction, and reliance on audio transcriptions. Nevertheless, LLMs show significant potential for enhancing patient monitoring and reducing oncologists’ workload. Future research should focus on the specific training of LLMs for medical tasks, conducting studies with real patients, implementing interactive evaluations, expanding sample sizes, and ensuring robustness and generalization in diverse clinical settings. Conclusions: This study concludes that LLMs can classify subjective toxicities from chemotherapy with accuracy comparable to expert oncologists. The LLM’s performance in general toxicity categories is within the expert range, but there is room for improvement in specific categories. LLMs have the potential to enhance patient monitoring, enable early interventions, and reduce severe complications, improving care quality and efficiency. Future research should involve specific training of LLMs, validation with real patients, and the incorporation of interactive capabilities for real-time patient interactions. Ethical considerations, including data accuracy, transparency, and privacy, are crucial for the safe integration of LLMs into clinical practice. Full article
(This article belongs to the Section Methods and Technologies Development)
17 pages, 507 KiB  
Article
Streamlining Distribution Routes Using the Language Model of Artificial Intelligence
by Kristína Kleinová and Martin Straka
Sustainability 2024, 16(16), 6890; https://doi.org/10.3390/su16166890 (registering DOI) - 11 Aug 2024
Viewed by 666
Abstract
This article addresses the use of artificial intelligence for the needs of effective, sustainable development in logistics and its components. The subject of this article is to highlight the possibility of processing optimization methods using an artificial intelligence module. The goal is to [...] Read more.
This article addresses the use of artificial intelligence for the needs of effective, sustainable development in logistics and its components. The subject of this article is to highlight the possibility of processing optimization methods using an artificial intelligence module. The goal is to determine whether the AI module can replicate the same, or at least have a similar result, as the traditional optimization methods used in practice. The challenge involves constantly identifying reserves in already highly sophisticated micro-logistics systems using modern commercial means of artificial intelligence. Applying artificial intelligence to elements of a company’s micro-logistics model is a new approach. This article aims to determine whether artificial intelligence can reduce costs through calculations in a specific area defined for it. By optimizing distribution routes using ChatGPT-3.5, we significantly reduced the total distance traveled, leading to substantial savings in transportation costs. This optimization led to a significant improvement in the efficiency of logistic processes and considerable cost savings. This result demonstrates that artificial intelligence can be an effective tool for solving complex logistic tasks. The possibilities of effectively sustainable logistics development with the help of artificial intelligence lie not only in the quality of the achieved outputs but also in the speed of the calculations and the procedures for solving defined project tasks. It follows from this definition that artificial intelligence will continue to play an essential role in the defined field of logistics in the future. Full article
(This article belongs to the Special Issue Sustainability in Logistics and Supply Chain Management)
Show Figures

Figure 1

13 pages, 791 KiB  
Article
ChatGPT as an Information Source for Patients with Migraines: A Qualitative Case Study
by Pascal Schütz, Sina Lob, Hiba Chahed, Lisa Dathe, Maren Löwer, Hannah Reiß, Alina Weigel, Joanna Albrecht, Pinar Tokgöz and Christoph Dockweiler
Healthcare 2024, 12(16), 1594; https://doi.org/10.3390/healthcare12161594 - 10 Aug 2024
Viewed by 548
Abstract
Migraines are one of the most common and expensive neurological diseases worldwide. Non-pharmacological and digitally delivered treatment options have long been used in the treatment of migraines. For instance, migraine management tools, online migraine diagnosis or digitally networked patients have been used. Recently, [...] Read more.
Migraines are one of the most common and expensive neurological diseases worldwide. Non-pharmacological and digitally delivered treatment options have long been used in the treatment of migraines. For instance, migraine management tools, online migraine diagnosis or digitally networked patients have been used. Recently, applications of ChatGPT are used in fields of healthcare ranging from identifying potential research topics to assisting professionals in clinical diagnosis and helping patients in managing their health. Despite advances in migraine management, only a minority of patients are adequately informed and treated. It is important to provide these patients with information to help them manage the symptoms and their daily activities. The primary aim of this case study was to examine the appropriateness of ChatGPT to handle symptom descriptions responsibly, suggest supplementary assistance from credible sources, provide valuable perspectives on treatment options, and exhibit potential influences on daily life for patients with migraines. Using a deductive, qualitative study, ten interactions with ChatGPT on different migraine types were analyzed through semi-structured interviews. ChatGPT provided relevant information aligned with common scientific patient resources. Responses were generally intelligible and situationally appropriate, providing personalized insights despite occasional discrepancies in interaction. ChatGPT’s empathetic tone and linguistic clarity encouraged user engagement. However, source citations were found to be inconsistent and, in some cases, not comprehensible, which affected the overall comprehensibility of the information. ChatGPT might be promising for patients seeking information on migraine conditions. Its user-specific responses demonstrate potential benefits over static web-based sources. However, reproducibility and accuracy issues highlight the need for digital health literacy. The findings underscore the necessity for continuously evaluating AI systems and their broader societal implications in health communication. Full article
18 pages, 3249 KiB  
Article
New Approach for Automated Explanation of Material Phenomena (AA6082) Using Artificial Neural Networks and ChatGPT
by Tomaž Goričan, Milan Terčelj and Iztok Peruš
Appl. Sci. 2024, 14(16), 7015; https://doi.org/10.3390/app14167015 - 9 Aug 2024
Viewed by 458
Abstract
Artificial intelligence methods, especially artificial neural networks (ANNs), have increasingly been utilized for the mathematical description of physical phenomena in (metallic) material processing. Traditional methods often fall short in explaining the complex, real-world data observed in production. While ANN models, typically functioning as [...] Read more.
Artificial intelligence methods, especially artificial neural networks (ANNs), have increasingly been utilized for the mathematical description of physical phenomena in (metallic) material processing. Traditional methods often fall short in explaining the complex, real-world data observed in production. While ANN models, typically functioning as “black boxes”, improve production efficiency, a deeper understanding of the phenomena, akin to that provided by explicit mathematical formulas, could enhance this efficiency further. This article proposes a general framework that leverages ANNs (i.e., Conditional Average Estimator—CAE) to explain predicted results alongside their graphical presentation, marking a significant improvement over previous approaches and those relying on expert assessments. Unlike existing Explainable AI (XAI) methods, the proposed framework mimics the standard scientific methodology, utilizing minimal parameters for the mathematical representation of physical phenomena and their derivatives. Additionally, it analyzes the reliability and accuracy of the predictions using well-known statistical metrics, transitioning from deterministic to probabilistic descriptions for better handling of real-world phenomena. The proposed approach addresses both aleatory and epistemic uncertainties inherent in the data. The concept is demonstrated through the hot extrusion of aluminum alloy 6082, where CAE ANN models and predicts key parameters, and ChatGPT explains the results, enabling researchers and/or engineers to better understand the phenomena and outcomes obtained by ANNs. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

24 pages, 637 KiB  
Article
Testing Stimulus Equivalence in Transformer-Based Agents
by Alexis Carrillo and Moisés Betancort
Future Internet 2024, 16(8), 289; https://doi.org/10.3390/fi16080289 - 9 Aug 2024
Viewed by 333
Abstract
This study investigates the ability of transformer-based models (TBMs) to form stimulus equivalence (SE) classes. We employ BERT and GPT as TBM agents in SE tasks, evaluating their performance across training structures (linear series, one-to-many and many-to-one) and relation types (select–reject, select-only). Our [...] Read more.
This study investigates the ability of transformer-based models (TBMs) to form stimulus equivalence (SE) classes. We employ BERT and GPT as TBM agents in SE tasks, evaluating their performance across training structures (linear series, one-to-many and many-to-one) and relation types (select–reject, select-only). Our findings demonstrate that both models performed above mastery criterion in the baseline phase across all simulations (n = 12). However, they exhibit limited success in reflexivity, transitivity, and symmetry tests. Notably, both models achieved success only in the linear series structure with select–reject relations, failing in one-to-many and many-to-one structures, and all select-only conditions. These results suggest that TBM may be forming decision rules based on learned discriminations and reject relations, rather than responding according to equivalence class formation. The absence of reject relations appears to influence their responses and the occurrence of hallucinations. This research highlights the potential of SE simulations for: (a) comparative analysis of learning mechanisms, (b) explainability techniques for TBM decision-making, and (c) TBM bench-marking independent of pre-training or fine-tuning. Future investigations can explore upscaling simulations and utilize SE tasks within a reinforcement learning framework. Full article
Show Figures

Figure 1

16 pages, 1963 KiB  
Article
Cross-Domain Fake News Detection Using a Prompt-Based Approach
by Jawaher Alghamdi, Yuqing Lin and Suhuai Luo
Future Internet 2024, 16(8), 286; https://doi.org/10.3390/fi16080286 - 8 Aug 2024
Viewed by 392
Abstract
The proliferation of fake news poses a significant challenge in today’s information landscape, spanning diverse domains and topics and undermining traditional detection methods confined to specific domains. In response, there is a growing interest in strategies for detecting cross-domain misinformation. However, traditional machine [...] Read more.
The proliferation of fake news poses a significant challenge in today’s information landscape, spanning diverse domains and topics and undermining traditional detection methods confined to specific domains. In response, there is a growing interest in strategies for detecting cross-domain misinformation. However, traditional machine learning (ML) approaches often struggle with the nuanced contextual understanding required for accurate news classification. To address these challenges, we propose a novel contextualized cross-domain prompt-based zero-shot approach utilizing a pre-trained Generative Pre-trained Transformer (GPT) model for fake news detection (FND). In contrast to conventional fine-tuning methods reliant on extensive labeled datasets, our approach places particular emphasis on refining prompt integration and classification logic within the model’s framework. This refinement enhances the model’s ability to accurately classify fake news across diverse domains. Additionally, the adaptability of our approach allows for customization across diverse tasks by modifying prompt placeholders. Our research significantly advances zero-shot learning by demonstrating the efficacy of prompt-based methodologies in text classification, particularly in scenarios with limited training data. Through extensive experimentation, we illustrate that our method effectively captures domain-specific features and generalizes well to other domains, surpassing existing models in terms of performance. These findings contribute significantly to the ongoing efforts to combat fake news dissemination, particularly in environments with severely limited training data, such as online platforms. Full article
(This article belongs to the Special Issue Embracing Artificial Intelligence (AI) for Network and Service)
Show Figures

Figure 1

18 pages, 946 KiB  
Article
SqliGPT: Evaluating and Utilizing Large Language Models for Automated SQL Injection Black-Box Detection
by Zhiwen Gui, Enze Wang, Binbin Deng, Mingyuan Zhang, Yitao Chen, Shengfei Wei, Wei Xie and Baosheng Wang
Appl. Sci. 2024, 14(16), 6929; https://doi.org/10.3390/app14166929 - 7 Aug 2024
Viewed by 914
Abstract
SQL injection (SQLI) black-box detection, which simulates external attack scenarios, is crucial for assessing vulnerabilities in real-world web applications. However, existing black-box detection methods rely on predefined rules to cover the most common SQLI cases, lacking diversity in vulnerability detection scheduling and payload, [...] Read more.
SQL injection (SQLI) black-box detection, which simulates external attack scenarios, is crucial for assessing vulnerabilities in real-world web applications. However, existing black-box detection methods rely on predefined rules to cover the most common SQLI cases, lacking diversity in vulnerability detection scheduling and payload, suffering from limited efficiency and accuracy. Large Language Models (LLMs) have shown significant advancements in several domains, so we developed SqliGPT, an LLM-powered SQLI black-box scanner that leverages the advanced contextual understanding and reasoning abilities of LLMs. Our approach introduces the Strategy Selection Module to improve detection efficiency and the Defense Bypass Module to address insufficient defense mechanisms. We evaluated SqliGPT against six state-of-the-art scanners using our SqliMicroBenchmark. Our evaluation results indicate that SqliGPT successfully detected all 45 targets, outperforming other scanners, particularly on targets with insufficient defenses. Additionally, SqliGPT demonstrated excellent efficiency in executing detection tasks, slightly underperforming Arachni and SQIRL on 27 targets but besting them on the other 18 targets. This study highlights the potential of LLMs in SQLI black-box detection and demonstrates the feasibility and effectiveness of LLMs in enhancing detection efficiency and accuracy. Full article
Show Figures

Figure 1

23 pages, 2024 KiB  
Review
Large Language Models in Healthcare and Medical Domain: A Review
by Zabir Al Nazi and Wei Peng
Informatics 2024, 11(3), 57; https://doi.org/10.3390/informatics11030057 - 7 Aug 2024
Viewed by 1596
Abstract
The deployment of large language models (LLMs) within the healthcare sector has sparked both enthusiasm and apprehension. These models exhibit the remarkable ability to provide proficient responses to free-text queries, demonstrating a nuanced understanding of professional medical knowledge. This comprehensive survey delves into [...] Read more.
The deployment of large language models (LLMs) within the healthcare sector has sparked both enthusiasm and apprehension. These models exhibit the remarkable ability to provide proficient responses to free-text queries, demonstrating a nuanced understanding of professional medical knowledge. This comprehensive survey delves into the functionalities of existing LLMs designed for healthcare applications and elucidates the trajectory of their development, starting with traditional Pretrained Language Models (PLMs) and then moving to the present state of LLMs in the healthcare sector. First, we explore the potential of LLMs to amplify the efficiency and effectiveness of diverse healthcare applications, particularly focusing on clinical language understanding tasks. These tasks encompass a wide spectrum, ranging from named entity recognition and relation extraction to natural language inference, multimodal medical applications, document classification, and question-answering. Additionally, we conduct an extensive comparison of the most recent state-of-the-art LLMs in the healthcare domain, while also assessing the utilization of various open-source LLMs and highlighting their significance in healthcare applications. Furthermore, we present the essential performance metrics employed to evaluate LLMs in the biomedical domain, shedding light on their effectiveness and limitations. Finally, we summarize the prominent challenges and constraints faced by large language models in the healthcare sector by offering a holistic perspective on their potential benefits and shortcomings. This review provides a comprehensive exploration of the current landscape of LLMs in healthcare, addressing their role in transforming medical applications and the areas that warrant further research and development. Full article
Show Figures

Figure 1

Back to TopTop