Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (129)

Search Parameters:
Keywords = text tagging

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
9 pages, 177 KiB  
Essay
Curious Travellers: New Journeys for the Home Tour
by Mary-Ann Constantine
Humanities 2025, 14(2), 34; https://doi.org/10.3390/h14020034 - 17 Feb 2025
Viewed by 131
Abstract
This short concluding chapter reflects on the work of an ongoing collaborative academic project focused on the C18th home tour. Curious Travellers could be described as a ‘crucible’ project—a space in which different media, different perspectives, and different research skills combine and collide. [...] Read more.
This short concluding chapter reflects on the work of an ongoing collaborative academic project focused on the C18th home tour. Curious Travellers could be described as a ‘crucible’ project—a space in which different media, different perspectives, and different research skills combine and collide. Currently funded by the AHRC, it is a digital humanities project involving TEI tagging and crowd-sourcing, but its foundation is archival research into manuscripts. It is focused through the influential Tours of Wales and Scotland published by the naturalist and antiquarian Thomas Pennant, yet it seeks to unpick the multiple voices and collaborations behind his texts and to explore their legacy in the journeys and texts of others. The creation of new editions continues to generate new topics and research questions, including Anglophone/Celtic-language interactions; the writings of women tourists; the role of material objects (specimens and souvenirs) and of visual culture in knowledge exchange and production. Increasingly, project researchers are relating their work to broader global contexts of colonialism and environmental history. The diversity of the genre has proved hugely stimulating for a range of audiences beyond academia: community engagement and creative practices have been a key feature from the start. There are, of course, challenges—practical, methodological, financial. This reflective piece will acknowledge the constraints, as well as the possibilities, of being multi-stranded, cross-disciplinary—and intermittently funded. Full article
(This article belongs to the Special Issue Eighteenth-Century Travel Writing: New Directions)
39 pages, 23368 KiB  
Article
Vision-Based Localization in Urban Areas for Mobile Robots
by Erdal Alimovski, Gokhan Erdemir and Ahmet Emin Kuzucuoglu
Sensors 2025, 25(4), 1178; https://doi.org/10.3390/s25041178 - 14 Feb 2025
Viewed by 340
Abstract
Robust autonomous navigation systems rely on mapping, locomotion, path planning, and localization factors. Localization, one of the most essential factors of navigation, is a crucial requirement for a mobile robot because it needs the capability to localize itself in the environment. Global Positioning [...] Read more.
Robust autonomous navigation systems rely on mapping, locomotion, path planning, and localization factors. Localization, one of the most essential factors of navigation, is a crucial requirement for a mobile robot because it needs the capability to localize itself in the environment. Global Positioning Systems (GPSs) are commonly used for outdoor mobile robot localization tasks. However, various environmental circumstances, such as high-rise buildings and trees, affect GPS signal quality, which leads to reduced precision or complete signal blockage. This study proposes a visual-based localization system for outdoor mobile robots in crowded urban environments. The proposed system comprises three steps. The first step is to detect the text in urban areas using the “Efficient and Accurate Scene Text Detector (EAST)” algorithm. Then, EasyOCR was applied to the detected text for the recognition phase to extract text from images that were obtained from EAST. The results from text detection and recognition algorithms were enhanced by applying post-processing and word similarity algorithms. In the second step, once the text detection and recognition process is completed, the recognized word (label/tag) is sent to the Places API in order to return the recognized word’s coordinates that are passed within the specified radius. Parallely, points of interest (POI) data are collected for a defined area by a certain radius while the robot has an accurate internet connection. The proposed system was tested in three distinct urban areas by creating five scenarios under different lighting conditions, such as morning and evening, using the outdoor delivery robot utilized in this study. In the case studies, it has been shown that the proposed system provides a low error of around 4 m for localization tasks. Compared to existing works, the proposed system consistently outperforms all other approaches using just one sensor. The results indicate the efficacy of the proposed system for localization tasks in environments where GPS signals are limited or completely blocked. Full article
(This article belongs to the Special Issue Mobile Robots: Navigation, Control and Sensing—2nd Edition)
Show Figures

Figure 1

18 pages, 799 KiB  
Article
Quantifying Interdisciplinarity in Scientific Articles Using Deep Learning Toward a TRIZ-Based Framework for Cross-Disciplinary Innovation
by Nicolas Douard, Ahmed Samet, George Giakos and Denis Cavallucci
Mach. Learn. Knowl. Extr. 2025, 7(1), 7; https://doi.org/10.3390/make7010007 - 12 Jan 2025
Viewed by 609
Abstract
Interdisciplinary research (IDR) is essential for addressing complex global challenges that surpass the capabilities of any single discipline. However, measuring interdisciplinarity remains challenging due to conceptual ambiguities and inconsistent methodologies. To overcome these challenges, we propose a deep learning approach that quantifies interdisciplinarity [...] Read more.
Interdisciplinary research (IDR) is essential for addressing complex global challenges that surpass the capabilities of any single discipline. However, measuring interdisciplinarity remains challenging due to conceptual ambiguities and inconsistent methodologies. To overcome these challenges, we propose a deep learning approach that quantifies interdisciplinarity in scientific articles through semantic analysis of titles and abstracts. Utilizing the Semantic Scholar Open Research Corpus (S2ORC), we leveraged metadata field tags to categorize papers as either interdisciplinary or monodisciplinary, establishing the foundation for supervised learning in our model. Specifically, we preprocessed the textual data and employed a Text Convolutional Neural Network (Text CNN) architecture to identify semantic patterns indicative of interdisciplinarity. Our model achieved an F1 score of 0.82, surpassing baseline machine learning models. By directly analyzing semantic content and incorporating metadata for training, our method addresses the limitations of previous approaches that rely solely on bibliometric features such as citations and co-authorship. Furthermore, our large-scale analysis of 136 million abstracts revealed that approximately 25% of the literature within the specified disciplines is interdisciplinary. Additionally, we outline how our quantification method can be integrated into a TRIZ-based (Theory of Inventive Problem Solving) methodological framework for cross-disciplinary innovation, providing a foundation for systematic knowledge transfer and inventive problem solving across domains. Overall, this approach not only offers a scalable measurement of interdisciplinarity but also contributes to a framework for facilitating innovation through structured cross-domain knowledge integration. Full article
(This article belongs to the Section Learning)
Show Figures

Figure 1

28 pages, 2857 KiB  
Article
IndoGovBERT: A Domain-Specific Language Model for Processing Indonesian Government SDG Documents
by Agus Riyadi, Mate Kovacs, Uwe Serdült and Victor Kryssanov
Big Data Cogn. Comput. 2024, 8(11), 153; https://doi.org/10.3390/bdcc8110153 - 9 Nov 2024
Viewed by 1671
Abstract
Achieving the Sustainable Development Goals (SDGs) requires collaboration among various stakeholders, particularly governments and non-state actors (NSAs). This collaboration results in but is also based on a continually growing volume of documents that needs to be analyzed and processed in a systematic way [...] Read more.
Achieving the Sustainable Development Goals (SDGs) requires collaboration among various stakeholders, particularly governments and non-state actors (NSAs). This collaboration results in but is also based on a continually growing volume of documents that needs to be analyzed and processed in a systematic way by government officials. Artificial Intelligence and Natural Language Processing (NLP) could, thus, offer valuable support for progressing towards SDG targets, including automating the government budget tagging and classifying NSA requests and initiatives, as well as helping uncover the possibilities for matching these two categories of activities. Many non-English speaking countries, including Indonesia, however, face limited NLP resources, such as, for instance, domain-specific pre-trained language models (PTLMs). This circumstance makes it difficult to automate document processing and improve the efficacy of SDG-related government efforts. The presented study introduces IndoGovBERT, a Bidirectional Encoder Representations from Transformers (BERT)-based PTLM built with domain-specific corpora, leveraging the Indonesian government’s public and internal documents. The model is intended to automate various laborious tasks of SDG document processing by the Indonesian government. Different approaches to PTLM development known from the literature are examined in the context of typical government settings. The most effective, in terms of the resultant model performance, but also most efficient, in terms of the computational resources required, methodology is determined and deployed for the development of the IndoGovBERT model. The developed model is then scrutinized in several text classification and similarity assessment experiments, where it is compared with four Indonesian general-purpose language models, a non-transformer approach of the Multilabel Topic Model (MLTM), as well as with a Multilingual BERT model. Results obtained in all experiments highlight the superior capability of the IndoGovBERT model for Indonesian government SDG document processing. The latter suggests that the proposed PTLM development methodology could be adopted to build high-performance specialized PTLMs for governments around the globe which face SDG document processing and other NLP challenges similar to the ones dealt with in the presented study. Full article
(This article belongs to the Special Issue Artificial Intelligence and Natural Language Processing)
Show Figures

Figure 1

34 pages, 4479 KiB  
Article
Development of a Children’s Educational Dictionary for a Low-Resource Language Using AI Tools
by Diana Rakhimova, Aidana Karibayeva, Vladislav Karyukin, Assem Turarbek, Zhansaya Duisenbekkyzy and Rashid Aliyev
Computers 2024, 13(10), 253; https://doi.org/10.3390/computers13100253 - 2 Oct 2024
Cited by 1 | Viewed by 1388
Abstract
Today, various interactive tools or partially available artificial intelligence applications are actively used in educational processes to solve multiple problems for resource-rich languages, such as English, Spanish, French, etc. Unfortunately, the situation is different and more complex for low-resource languages, like Kazakh, Uzbek, [...] Read more.
Today, various interactive tools or partially available artificial intelligence applications are actively used in educational processes to solve multiple problems for resource-rich languages, such as English, Spanish, French, etc. Unfortunately, the situation is different and more complex for low-resource languages, like Kazakh, Uzbek, Mongolian, and others, due to the lack of qualitative and accessible resources, morphological complexity, and the semantics of agglutinative languages. This article presents research on early childhood learning resources for the low-resource Kazakh language. Generally, a dictionary for children differs from classical educational dictionaries. The difference between dictionaries for children and adults lies in their purpose and methods of presenting information. A themed dictionary will make learning and remembering new words easier for children because they will be presented in a specific context. This article discusses developing an approach to creating a thematic children’s dictionary of the low-resource Kazakh language using artificial intelligence. The proposed approach is based on several important stages: the initial formation of a list of English words with the use of ChatGPT; identification of their semantic weights; generation of phrases and sentences with the use of the list of semantically related words; translation of obtained phrases and sentences from English to Kazakh, dividing them into bigrams and trigrams; and processing with Kazakh language POS pattern tag templates to adapt them for children. When the dictionary was formed, the semantic proximity of words and phrases to the given theme and age restrictions for children were taken into account. The formed dictionary phrases were evaluated using the cosine similarity, Euclidean similarity, and Manhattan distance metrics. Moreover, the dictionary was extended with video and audio data by implementing models like DALL-E 3, Midjourney, and Stable Diffusion to illustrate the dictionary data and TTS (Text to Speech) technology for the Kazakh language for voice synthesis. The developed thematic dictionary approach was tested, and a SUS (System Usability Scale) assessment of the application was conducted. The experimental results demonstrate the proposed approach’s high efficiency and its potential for wide use in educational purposes. Full article
(This article belongs to the Special Issue Smart Learning Environments)
Show Figures

Figure 1

19 pages, 2296 KiB  
Article
A Hybrid Approach to Ontology Construction for the Badini Kurdish Language
by Media Azzat, Karwan Jacksi and Ismael Ali
Information 2024, 15(9), 578; https://doi.org/10.3390/info15090578 - 19 Sep 2024
Viewed by 1563
Abstract
Semantic ontologies have been widely utilized as crucial tools within natural language processing, underpinning applications such as knowledge extraction, question answering, machine translation, text comprehension, information retrieval, and text summarization. While the Kurdish language, a low-resource language, has been the subject of some [...] Read more.
Semantic ontologies have been widely utilized as crucial tools within natural language processing, underpinning applications such as knowledge extraction, question answering, machine translation, text comprehension, information retrieval, and text summarization. While the Kurdish language, a low-resource language, has been the subject of some ontological research in other dialects, a semantic web ontology for the Badini dialect remains conspicuously absent. This paper addresses this gap by presenting a methodology for constructing and utilizing a semantic web ontology for the Badini dialect of the Kurdish language. A Badini annotated corpus (UOZBDN) was created and manually annotated with part-of-speech (POS) tags. Subsequently, an HMM-based POS tagger model was developed using the UOZBDN corpus and applied to annotate additional text for ontology extraction. Ontology extraction was performed by employing predefined rules to identify nouns and verbs from the model-annotated corpus and subsequently forming semantic predicates. Robust methodologies were adopted for ontology development, resulting in a high degree of precision. The POS tagging model attained an accuracy of 95.04% when applied to the UOZBDN corpus. Furthermore, a manual evaluation conducted by Badini Kurdish language experts yielded a 97.42% accuracy rate for the extracted ontology. Full article
(This article belongs to the Special Issue Knowledge Representation and Ontology-Based Data Management)
Show Figures

Figure 1

23 pages, 7789 KiB  
Article
Design and Implementation of an Interactive Question-Answering System with Retrieval-Augmented Generation for Personalized Databases
by Jaeyeon Byun, Bokyeong Kim, Kyung-Ae Cha and Eunhyung Lee
Appl. Sci. 2024, 14(17), 7995; https://doi.org/10.3390/app14177995 - 6 Sep 2024
Cited by 1 | Viewed by 2445
Abstract
This study introduces a novel approach to personalized information retrieval by integrating retrieval augmentation generation (RAG) with a personalized database system. Recent advancements in large language models (LLMs) have shown impressive text generation capabilities but face limitations in knowledge accuracy and hallucinations. Our [...] Read more.
This study introduces a novel approach to personalized information retrieval by integrating retrieval augmentation generation (RAG) with a personalized database system. Recent advancements in large language models (LLMs) have shown impressive text generation capabilities but face limitations in knowledge accuracy and hallucinations. Our research addresses these challenges by combining LLMs with structured, personalized data to enhance search precision and relevance. By tagging keywords within personal documents and organizing information into context-based categories, users can conduct efficient searches within their data repositories. We conducted experiments using the GPT-3.5 and text-embedding-ada-002 models and evaluated the RAG assessment framework with five different language models and two embedding models. Our results indicate that the combination of GPT-3.5 and text-embedding-ada-002 is effective for a personalized database question-answering system, with potential for various language models depending on the application. Our approach offers improved accuracy, real-time data updates, and enhanced user experience, making a significant contribution to information retrieval by LLMs and impacting various artificial intelligence applications. Full article
Show Figures

Figure 1

17 pages, 2268 KiB  
Article
The Sustainable Innovation of AI: Text Mining the Core Capabilities of Researchers in the Digital Age of Industry 4.0
by Yajun Ji, Shengtai Zhang, Fang Han, Ran Cui and Tao Jiang
Sustainability 2024, 16(17), 7767; https://doi.org/10.3390/su16177767 - 6 Sep 2024
Viewed by 1106
Abstract
Sustainable innovation in the field of artificial intelligence (AI) is essential for the development of Industry 4.0. Recognizing the innovation abilities of researchers is fundamental to achieving sustainable innovation within organizations. This study proposes a method for identifying the core innovative competency field [...] Read more.
Sustainable innovation in the field of artificial intelligence (AI) is essential for the development of Industry 4.0. Recognizing the innovation abilities of researchers is fundamental to achieving sustainable innovation within organizations. This study proposes a method for identifying the core innovative competency field of researchers through text mining, which involves the extraction of core competency tags, topic clustering, and calculating the relevance between researchers and topics. Using AI as a case study, the research identifies the core innovative competency field of researchers, uncovers opportunities for sustainable innovation, and highlights key innovators. This approach offers deeper insights for AI R&D activities, providing effective support for promoting sustainable innovation. Compared to traditional expertise identification methods, this approach provides a more in-depth and detailed portrayal of researchers’ expertise, particularly highlighting potential innovation domains with finer granularity. It is less influenced by subjective factors and can be conveniently applied to identify the core innovative competency field of researchers in any other research field, making it especially suitable for interdisciplinary areas. By offering a precise and comprehensive understanding of researchers’ capability fields, this method enhances the strategic planning and execution of innovative projects, ensuring that organizations can effectively leverage the expertise of their researchers to drive forward sustainable innovation. Full article
(This article belongs to the Special Issue Industry 4.0, Digitization and Opportunities for Sustainability)
Show Figures

Figure 1

19 pages, 1464 KiB  
Article
TAG2G: A Diffusion-Based Approach to Interlocutor-Aware Co-Speech Gesture Generation
by Filippo Favali, Viktor Schmuck, Valeria Villani and Oya Celiktutan
Electronics 2024, 13(17), 3364; https://doi.org/10.3390/electronics13173364 - 24 Aug 2024
Viewed by 1139
Abstract
Extended reality (XR) systems are about to be integrated into our daily lives and will provide support in a variety of fields such as education and coaching. Enhancing user experience demands agents that are capable of displaying realistic affective and social behaviors within [...] Read more.
Extended reality (XR) systems are about to be integrated into our daily lives and will provide support in a variety of fields such as education and coaching. Enhancing user experience demands agents that are capable of displaying realistic affective and social behaviors within these systems, and, as a prerequisite, with the capability of understanding their interaction partner and responding appropriately. Based on our literature review of recent works published in the field of co-speech gesture generation, researchers have developed complex models capable of generating gestures characterized by a high level of human-likeness and speaker appropriateness. Nevertheless, this is only true in settings where the agent has an active status (i.e., the agent acts as the speaker), or it is delivering a monologue in a non-interactive setting. However, as illustrated in multiple works and competitions like the GENEA Challenge, these models remain inadequate in generating interlocutor-aware gestures. We consider interlocutor-aware gesture generation the process of displaying gestures that take into account the conversation partner’s behavior. Moreover, in settings where the agent is the listener, generated gestures lack the level of naturalness that we expect from a face-to-face conversation. To overcome these issues, we have designed a pipeline, called TAG2G, composed of a diffusion model, which was demonstrated to be a stable and powerful tool in gesture generation, and a vector-quantized variational auto-encoder (VQVAE), widely employed to produce meaningful gesture embeddings. Refocusing from monadic to dyadic multimodal input settings (i.e., taking into account text, audio, and previous gestures of both participants of a conversation) allows us to explore and infer the complex interaction mechanisms that lie in a balanced two-sided conversation. As per our results, a multi-agent conversational input setup improves the generated gestures’ appropriateness with respect to the conversational counterparts. Conversely, when the agent is speaking, a monadic approach performs better in terms of the generated gestures’ appropriateness in relation to the speech. Full article
Show Figures

Figure 1

31 pages, 6859 KiB  
Article
Multimodal Analysis of the Spanish Linguistic Landscape in Alabama
by Alicia Cipria and Erin O’Rourke
Languages 2024, 9(8), 264; https://doi.org/10.3390/languages9080264 - 30 Jul 2024
Cited by 1 | Viewed by 2198
Abstract
The study of linguistic landscapes (LL) examines the use of signage in public spaces, often with a focus on the use of non-majority languages. The main goals of this project are to map, quantify, and analyze signage in Spanish within Tuscaloosa County, AL, [...] Read more.
The study of linguistic landscapes (LL) examines the use of signage in public spaces, often with a focus on the use of non-majority languages. The main goals of this project are to map, quantify, and analyze signage in Spanish within Tuscaloosa County, AL, an emerging site of Spanish language use which differs from the large urban places often studied in the LL literature. Photographs of public signage in Spanish were taken and uploaded to an ArcGIS Field Maps app to allow for both geolocation of the image and tagging of the image for specific linguistic and visual characteristics, which are subsumed under multimodality. Multimodality refers to the interaction of the linguistic code with other modes of communication such as images, colors, flags, and other cultural objects to make meaning in a given LL text. Within the multimodality framework, we examine the use of Spanish by itself or with English, location of the signage, communicative functions (symbolic, informative), and the combination of multimodal resources to index the actors originating the text and their intended audience. Full article
(This article belongs to the Special Issue Spanish in the US: A Sociolinguistic Approach)
Show Figures

Figure 1

13 pages, 1132 KiB  
Article
MM-Transformer: A Transformer-Based Knowledge Graph Link Prediction Model That Fuses Multimodal Features
by Dongsheng Wang, Kangjie Tang, Jun Zeng, Yue Pan, Yun Dai, Huige Li and Bin Han
Symmetry 2024, 16(8), 961; https://doi.org/10.3390/sym16080961 - 29 Jul 2024
Cited by 1 | Viewed by 1778
Abstract
Multimodal knowledge graph completion necessitates the integration of information from multiple modalities (such as images and text) into the structural representation of entities to improve link prediction. However, most existing studies have overlooked the interaction between different modalities and the symmetry in the [...] Read more.
Multimodal knowledge graph completion necessitates the integration of information from multiple modalities (such as images and text) into the structural representation of entities to improve link prediction. However, most existing studies have overlooked the interaction between different modalities and the symmetry in the modal fusion process. To address this issue, this paper proposed a Transformer-based knowledge graph link prediction model (MM-Transformer) that fuses multimodal features. Different modal encoders are employed to extract structural, visual, and textual features, and symmetrical hybrid key-value calculations are performed on features from different modalities based on the Transformer architecture. The similarities of textual tags to structural tags and visual tags are calculated and aggregated, respectively, and multimodal entity representations are modeled and optimized to reduce the heterogeneity of the representations. The experimental results show that compared with the current multimodal SOTA method, MKGformer, MM-Transformer improves the Hits@1 and Hits@10 evaluation indicators by 1.17% and 1.39%, respectively, proving that the proposed method can effectively solve the problem of multimodal feature fusion in the knowledge graph link prediction task. Full article
Show Figures

Figure 1

26 pages, 6811 KiB  
Article
Fd-CasBGRel: A Joint Entity–Relationship Extraction Model for Aquatic Disease Domains
by Hongbao Ye, Lijian Lv, Chengquan Zhou and Dawei Sun
Appl. Sci. 2024, 14(14), 6147; https://doi.org/10.3390/app14146147 - 15 Jul 2024
Viewed by 1094
Abstract
Entity–relationship extraction plays a pivotal role in the construction of domain knowledge graphs. For the aquatic disease domain, however, this relationship extraction is a formidable task because of overlapping relationships, data specialization, limited feature fusion, and imbalanced data samples, which significantly weaken the [...] Read more.
Entity–relationship extraction plays a pivotal role in the construction of domain knowledge graphs. For the aquatic disease domain, however, this relationship extraction is a formidable task because of overlapping relationships, data specialization, limited feature fusion, and imbalanced data samples, which significantly weaken the extraction’s performance. To tackle these challenges, this study leverages published books and aquatic disease websites as data sources to compile a text corpus, establish datasets, and then propose the Fd-CasBGRel model specifically tailored to the aquatic disease domain. The model uses the Casrel cascading binary tagging framework to address relationship overlap; utilizes task fine-tuning for better performance on aquatic disease data; trains on specialized aquatic disease corpora to improve adaptability; and integrates the BRC feature fusion module—which incorporates self-attention mechanisms, BiLSTM, relative position encoding, and conditional layer normalization—to leverage entity position and context for enhanced fusion. Further, it replaces the traditional cross-entropy loss function with the GHM loss function to mitigate category imbalance issues. The experimental results indicate that the F1 score of the Fd-CasBGRel on the aquatic disease dataset reached 84.71%, significantly outperforming several benchmark models. This model effectively addresses the challenges of ternary extraction’s low performance caused by high data specialization, insufficient feature integration, and data imbalances. The model achieved the highest F1 score of 86.52% on the overlapping relationship category dataset, demonstrating its robust capability in extracting overlapping data. Furthermore, We also conducted comparative experiments on the publicly available dataset WebNLG, and the model in this paper obtained the best performance metrics compared to the rest of the comparative models, indicating that the model has good generalization ability. Full article
Show Figures

Figure 1

15 pages, 1736 KiB  
Article
Multi-Level Attention with 2D Table-Filling for Joint Entity-Relation Extraction
by Zhenyu Zhang, Lin Shi, Yang Yuan, Huanyue Zhou and Shoukun Xu
Information 2024, 15(7), 407; https://doi.org/10.3390/info15070407 - 14 Jul 2024
Viewed by 1063
Abstract
Joint entity-relation extraction is a fundamental task in the construction of large-scale knowledge graphs. This task relies not only on the semantics of the text span but also on its intricate connections, including classification and structural details that most previous models overlook. In [...] Read more.
Joint entity-relation extraction is a fundamental task in the construction of large-scale knowledge graphs. This task relies not only on the semantics of the text span but also on its intricate connections, including classification and structural details that most previous models overlook. In this paper, we propose the incorporation of this information into the learning process. Specifically, we design a novel two-dimensional word-pair tagging method to define the task of entity and relation extraction. This allows type markers to focus on text tokens, gathering information for their corresponding spans. Additionally, we introduce a multi-level attention neural network to enhance its capacity to perceive structure-aware features. Our experiments show that our approach can overcome the limitations of earlier tagging methods and yield more accurate results. We evaluate our model using three different datasets: SciERC, ADE, and CoNLL04. Our model demonstrates competitive performance compared to the state-of-the-art, surpassing other approaches across the majority of evaluated metrics. Full article
Show Figures

Figure 1

15 pages, 248 KiB  
Article
Beyond Conversational Dialogue: Constructing a Catholic Dialogical Ethical Model for Multi-Religious Nigeria
by Ilesanmi G. Ajibola
Religions 2024, 15(7), 823; https://doi.org/10.3390/rel15070823 - 8 Jul 2024
Viewed by 1024
Abstract
This article argues that the Catholic Church in Nigeria needs to move beyond interreligious dialogue that dwells more on Councils’ interactions and discourses to develop a dialogical ethical framework that engages religious multiplicity in a more active capacity. Although Nigeria’s religious diversity necessitates [...] Read more.
This article argues that the Catholic Church in Nigeria needs to move beyond interreligious dialogue that dwells more on Councils’ interactions and discourses to develop a dialogical ethical framework that engages religious multiplicity in a more active capacity. Although Nigeria’s religious diversity necessitates interreligious dialogue, that alone is insufficient for constructing concrete ethics of dialogue. The article thus argued for an ethical framework tagged one family, many flavours. The theological sense of the proposal is rooted in Catholic social teachings but open to engagement with other religious traditions for mutual respect and social justice. The religious scope of the article is on Roman Catholicism and the Nigeria Muslim Ummah. The article addressed related ethical challenges confronting Nigeria’s interreligious landscape as a diverse religious community. Primary and secondary sources were used in gathering information for the article; thus, scriptural texts and traditions in Islam, as well as sources in Roman Catholicism, were theologically engaged. The suggested model acknowledges the importance of retaining one’s religious identity while also recognising the importance of interreligious dialogue and the right of the religious other in ethical matters. The article is envisioned to promote conversations about translating dialogical frameworks into practice. Full article
(This article belongs to the Special Issue Reimagining Catholic Ethics Today)
12 pages, 391 KiB  
Article
SCC-GPT: Source Code Classification Based on Generative Pre-Trained Transformers
by Mohammad D. Alahmadi, Moayad Alshangiti and Jumana Alsubhi
Mathematics 2024, 12(13), 2128; https://doi.org/10.3390/math12132128 - 7 Jul 2024
Viewed by 1046
Abstract
Developers often rely on online resources, such as Stack Overflow (SO), to seek assistance for programming tasks. To facilitate effective search and resource discovery, manual tagging of questions and posts with the appropriate programming language is essential. However, accurate tagging is not consistently [...] Read more.
Developers often rely on online resources, such as Stack Overflow (SO), to seek assistance for programming tasks. To facilitate effective search and resource discovery, manual tagging of questions and posts with the appropriate programming language is essential. However, accurate tagging is not consistently achieved, leading to the need for the automated classification of code snippets into the correct programming language as a tag. In this study, we introduce a novel approach to automated classification of code snippets from Stack Overflow (SO) posts into programming languages using generative pre-trained transformers (GPT). Our method, which does not require additional training on labeled data or dependency on pre-existing labels, classifies 224,107 code snippets into 19 programming languages. We employ the text-davinci-003 model of ChatGPT-3.5 and postprocess its responses to accurately identify the programming language. Our empirical evaluation demonstrates that our GPT-based model (SCC-GPT) significantly outperforms existing methods, achieving a median F1-score improvement that ranges from +6% to +31%. These findings underscore the effectiveness of SCC-GPT in enhancing code snippet classification, offering a cost-effective and efficient solution for developers who rely on SO for programming assistance. Full article
(This article belongs to the Special Issue AI-Augmented Software Engineering)
Show Figures

Figure 1

Back to TopTop