Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,132)

Search Parameters:
Keywords = text classification

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
30 pages, 1318 KiB  
Article
Malware Classification Using Dynamically Extracted API Call Embeddings
by Sahil Aggarwal and Fabio Di Troia
Appl. Sci. 2024, 14(13), 5731; https://doi.org/10.3390/app14135731 (registering DOI) - 30 Jun 2024
Abstract
Malware classification stands as a crucial element in establishing robust computer security protocols, encompassing the segmentation of malware into discrete groupings. Recently, the emergence of machine learning has presented itself as an apt approach for addressing this challenge. Models can undergo training employing [...] Read more.
Malware classification stands as a crucial element in establishing robust computer security protocols, encompassing the segmentation of malware into discrete groupings. Recently, the emergence of machine learning has presented itself as an apt approach for addressing this challenge. Models can undergo training employing diverse malware attributes, such as opcodes and API calls, to distill valuable insights for effective classification. Within the realm of natural language processing, word embeddings assume a pivotal role by representing text in a manner that aligns closely with the proximity of similar words. These embeddings facilitate the quantification of word resemblances. This research embarks on a series of experiments that harness hybrid machine learning methodologies. We derive word vectors from dynamic API call logs associated with malware and integrate them as features in collaboration with diverse classifiers. Our methodology involves the utilization of Hidden Markov Models and Word2Vec to generate embeddings from API call logs. Additionally, we amalgamate renowned models like BERT and ELMo, noted for their capacity to yield contextualized embeddings. The resultant vectors are channeled into our classifiers, namely Support Vector Machines (SVMs), Random Forest (RF), k-Nearest Neighbors (kNNs), and Convolutional Neural Networks (CNNs). Through two distinct sets of experiments, our objective revolves around the classification of both malware families and categories. The outcomes achieved illuminate the efficacy of API call embeddings as a potent instrument in the domain of malware classification, particularly in the realm of identifying malware families. The best combination was RF and word embeddings generated by Word2Vec, ELMo, and BERT, achieving an accuracy between 0.91 and 0.93. This result underscores the potential of our approach in effectively classifying malware. Full article
(This article belongs to the Collection Innovation in Information Security)
Show Figures

Figure 1

14 pages, 917 KiB  
Article
Improving Text Classification with Large Language Model-Based Data Augmentation
by Huanhuan Zhao, Haihua Chen, Thomas A. Ruggles, Yunhe Feng, Debjani Singh and Hong-Jun Yoon
Electronics 2024, 13(13), 2535; https://doi.org/10.3390/electronics13132535 - 28 Jun 2024
Viewed by 276
Abstract
Large Language Models (LLMs) such as ChatGPT possess advanced capabilities in understanding and generating text. These capabilities enable ChatGPT to create text based on specific instructions, which can serve as augmented data for text classification tasks. Previous studies have approached data augmentation (DA) [...] Read more.
Large Language Models (LLMs) such as ChatGPT possess advanced capabilities in understanding and generating text. These capabilities enable ChatGPT to create text based on specific instructions, which can serve as augmented data for text classification tasks. Previous studies have approached data augmentation (DA) by either rewriting the existing dataset with ChatGPT or generating entirely new data from scratch. However, it is unclear which method is better without comparing their effectiveness. This study investigates the application of both methods to two datasets: a general-topic dataset (Reuters news data) and a domain-specific dataset (Mitigation dataset). Our findings indicate that: 1. ChatGPT generated new data consistently enhanced model’s classification results for both datasets. 2. Generating new data generally outperforms rewriting existing data, though crafting the prompts carefully is crucial to extract the most valuable information from ChatGPT, particularly for domain-specific data. 3. The augmentation data size affects the effectiveness of DA; however, we observed a plateau after incorporating 10 samples. 4. Combining the rewritten sample with new generated sample can potentially further improve the model’s performance. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

18 pages, 1089 KiB  
Article
Research on Feature Fusion Method Based on Graph Convolutional Networks
by Dong Wang and Xuelin Chen
Appl. Sci. 2024, 14(13), 5612; https://doi.org/10.3390/app14135612 - 27 Jun 2024
Viewed by 194
Abstract
This paper proposes an enhanced BertGCN-Fusion (BGF) model aimed at addressing the limitations of Graph Convolutional Networks (GCN) in processing global text features for text categorization tasks. While traditional GCN effectively capture local structural features, they face challenges when integrating global semantic features. [...] Read more.
This paper proposes an enhanced BertGCN-Fusion (BGF) model aimed at addressing the limitations of Graph Convolutional Networks (GCN) in processing global text features for text categorization tasks. While traditional GCN effectively capture local structural features, they face challenges when integrating global semantic features. Issues such as the potential loss of global semantic information due to local feature fusion and limited depth of information propagation are prevalent. To overcome these challenges, the BGF model introduces improvements based on the BertGCN framework: (1) Feature fusion mechanism: Introducing a linear layer to fuse BERT outputs with traditional features facilitates the integration of fine-grained local semantic features from BERT with traditional global features. (2) Multilayer fusion approach: Employing a multilayer fusion technique enhances the integration of textual semantic features, thereby comprehensively and accurately capturing text semantic information. Experimental results demonstrate that the BGF model achieves notable performance improvements across multiple datasets. On the R8 and R52 datasets, the BGF model achieves accuracies of 98.45% and 93.77%, respectively, marking improvements of 0.28% to 0.90% compared to the BertGCN model. These findings highlight the BGF model’s efficacy in overcoming the deficiencies of traditional GCN in processing global semantic features, presenting an efficient approach for handling text data. Full article
31 pages, 4733 KiB  
Article
Enhanced Network Intrusion Detection System for Internet of Things Security Using Multimodal Big Data Representation with Transfer Learning and Game Theory
by Farhan Ullah, Ali Turab, Shamsher Ullah, Diletta Cacciagrano and Yue Zhao
Sensors 2024, 24(13), 4152; https://doi.org/10.3390/s24134152 - 26 Jun 2024
Viewed by 305
Abstract
Internet of Things (IoT) applications and resources are highly vulnerable to flood attacks, including Distributed Denial of Service (DDoS) attacks. These attacks overwhelm the targeted device with numerous network packets, making its resources inaccessible to authorized users. Such attacks may comprise attack references, [...] Read more.
Internet of Things (IoT) applications and resources are highly vulnerable to flood attacks, including Distributed Denial of Service (DDoS) attacks. These attacks overwhelm the targeted device with numerous network packets, making its resources inaccessible to authorized users. Such attacks may comprise attack references, attack types, sub-categories, host information, malicious scripts, etc. These details assist security professionals in identifying weaknesses, tailoring defense measures, and responding rapidly to possible threats, thereby improving the overall security posture of IoT devices. Developing an intelligent Intrusion Detection System (IDS) is highly complex due to its numerous network features. This study presents an improved IDS for IoT security that employs multimodal big data representation and transfer learning. First, the Packet Capture (PCAP) files are crawled to retrieve the necessary attacks and bytes. Second, Spark-based big data optimization algorithms handle huge volumes of data. Second, a transfer learning approach such as word2vec retrieves semantically-based observed features. Third, an algorithm is developed to convert network bytes into images, and texture features are extracted by configuring an attention-based Residual Network (ResNet). Finally, the trained text and texture features are combined and used as multimodal features to classify various attacks. The proposed method is thoroughly evaluated on three widely used IoT-based datasets: CIC-IoT 2022, CIC-IoT 2023, and Edge-IIoT. The proposed method achieves excellent classification performance, with an accuracy of 98.2%. In addition, we present a game theory-based process to validate the proposed approach formally. Full article
(This article belongs to the Section Internet of Things)
Show Figures

Figure 1

20 pages, 502 KiB  
Article
Reversal of the Word Sense Disambiguation Task Using a Deep Learning Model
by Algirdas Laukaitis
Appl. Sci. 2024, 14(13), 5550; https://doi.org/10.3390/app14135550 - 26 Jun 2024
Viewed by 224
Abstract
Word sense disambiguation (WSD) remains a persistent challenge in the natural language processing (NLP) community. While various NLP packages exist, the Lesk algorithm in the NLTK library demonstrates suboptimal accuracy. In this research article, we propose an innovative methodology and an open-source framework [...] Read more.
Word sense disambiguation (WSD) remains a persistent challenge in the natural language processing (NLP) community. While various NLP packages exist, the Lesk algorithm in the NLTK library demonstrates suboptimal accuracy. In this research article, we propose an innovative methodology and an open-source framework that effectively addresses the challenges of WSD by optimizing memory usage without compromising accuracy. Our system seamlessly integrates WSD into NLP tasks, offering functionality similar to that provided by the NLTK library. However, we go beyond the existing approaches by introducing a novel idea related to WSD. Specifically, we leverage deep neural networks and consider the language patterns learned by these models as the new gold standard. This approach suggests modifying existing semantic dictionaries, such as WordNet, to align with these patterns. Empirical validation through a series of experiments confirmed the effectiveness of our proposed method, achieving state-of-the-art performance across multiple WSD datasets. Notably, our system does not require the installation of additional software beyond the well-known Python libraries. The classification model is saved in a readily usable text format, and the entire framework (model and data) is publicly available on GitHub for the NLP research community. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
24 pages, 2050 KiB  
Article
Comprehensive Neural Cryptanalysis on Block Ciphers Using Different Encryption Methods
by Ongee Jeong, Ezat Ahmadzadeh and Inkyu Moon
Mathematics 2024, 12(13), 1936; https://doi.org/10.3390/math12131936 - 22 Jun 2024
Viewed by 206
Abstract
In this paper, we perform neural cryptanalysis on five block ciphers: Data Encryption Standard (DES), Simplified DES (SDES), Advanced Encryption Standard (AES), Simplified AES (SAES), and SPECK. The block ciphers are investigated on three different deep learning-based attacks, Encryption Emulation (EE), Plaintext Recovery [...] Read more.
In this paper, we perform neural cryptanalysis on five block ciphers: Data Encryption Standard (DES), Simplified DES (SDES), Advanced Encryption Standard (AES), Simplified AES (SAES), and SPECK. The block ciphers are investigated on three different deep learning-based attacks, Encryption Emulation (EE), Plaintext Recovery (PR), Key Recovery (KR), and Ciphertext Classification (CC) attacks. The attacks attempt to break the block ciphers in various cases, such as different types of plaintexts (i.e., block-sized bit arrays and texts), different numbers of round functions and quantity of training data, different text encryption methods (i.e., Word-based Text Encryption (WTE) and Sentence-based Text Encryption (STE)), and different deep learning model architectures. As a result, the block ciphers can be vulnerable to EE and PR attacks using a large amount of training data, and STE can improve the strength of the block ciphers, unlike WTE, which shows almost the same classification accuracy as the plaintexts, especially in a CC attack. Moreover, especially in the KR attack, the Recurrent Neural Network (RNN)-based deep learning model shows higher average Bit Accuracy Probability than the fully connected-based deep learning model. Furthermore, the RNN-based deep learning model is more suitable than the transformer-based deep learning model in the CC attack. Besides, when the keys are the same as the plaintexts, the KR attack can perfectly break the block ciphers, even if the plaintexts are randomly generated. Additionally, we identify that DES and SPECK32/64 applying two round functions are more vulnerable than those applying the single round function by performing the KR attack with randomly generated keys and randomly generated single plaintext. Full article
17 pages, 3591 KiB  
Article
Building Materials Classification Model Based on Text Data Enhancement and Semantic Feature Extraction
by Qiao Yan, Fei Jiao and Wei Peng
Buildings 2024, 14(6), 1859; https://doi.org/10.3390/buildings14061859 - 19 Jun 2024
Viewed by 278
Abstract
In order to accurately extract and match carbon emission factors from the Chinese textual building materials list and construct a precise carbon emission factor database, it is crucial to accurately classify the textual building materials. In this study, a novel classification model based [...] Read more.
In order to accurately extract and match carbon emission factors from the Chinese textual building materials list and construct a precise carbon emission factor database, it is crucial to accurately classify the textual building materials. In this study, a novel classification model based on text data enhancement and semantic feature extraction is proposed and applied for building materials classification. Firstly, the explanatory information on the building materials is collected and normalized to construct the original dataset. Then, the Latent Dirichlet Allocation and statistical-language-model-based hybrid ensemble data enhancement methods are explained in detail, and the semantic features closely related to the carbon emission factor are extracted by constructed composite convolutional networks and the transformed word vectors. Finally, the ensemble classification model is designed, constructed, and applied to match the carbon emission factor from the textual building materials. The experimental results show that the proposed model improves the F1Macro score by 4–12% compared to traditional machine learning and deep learning models. Full article
(This article belongs to the Section Building Materials, and Repair & Renovation)
Show Figures

Figure 1

9 pages, 562 KiB  
Article
Coding Diagnoses from the Electronic Death Certificate with the 11th Revision of the International Statistical Classification of Diseases and Related Health Problems: An Exploratory Study from Germany
by Jürgen Stausberg and Ulrich Vogel
Healthcare 2024, 12(12), 1214; https://doi.org/10.3390/healthcare12121214 - 18 Jun 2024
Viewed by 275
Abstract
The 11th Revision of the International Statistical Classification of Diseases, Injuries, and Causes of Death (ICD-11) will replace its predecessor as international standard for cause-of death-statistics. The digitization of healthcare is a main motivation for its introduction. In parallel, the replacement of the [...] Read more.
The 11th Revision of the International Statistical Classification of Diseases, Injuries, and Causes of Death (ICD-11) will replace its predecessor as international standard for cause-of death-statistics. The digitization of healthcare is a main motivation for its introduction. In parallel, the replacement of the paper-based death certificate with an electronic format is under evaluation. At the moment, the death certificate is used in paper-based format with ICD-10 for coding in Germany. To be prepared for the switch to ICD-11, the compatibility between ICD-11 and the electronic certificate should be assured. Objectives were to check the appropriateness of diagnosis-related information found on death certificates for an ICD-11 coding and to describe enhancements to the certificate’s structure needed to fully utilize the strengths of ICD-11. As part of an exploratory test of a respective application, information from 453 electronic death certificates were provided by one local health authority. From a sample of 200 certificates, 433 diagnosis texts were coded into the German version of ICD-11. The appropriateness of the results as well as the further requirements of ICD-11, particularly with regard to post-coordination, were checked. For 430 diagnosis texts, 649 ICD-11 codes were used. Three hundred and sixty two diagnosis texts were rated as appropriately represented through the coding result. Almost all certificates contained diagnosis texts that lacked details required by ICD-11 for a precise coding. The distribution of diseases was very similar between ICD-10 and ICD-11 coding. A few gaps in ICD-11 were identified. Information requested by ICD-11 for a mandatory post-coordination were almost entirely absent from the death certificates. The structure and content of the death certificate are currently not well prepared for an ICD-11 coding. Necessary information was frequently missing. The line-oriented structure of death certificates has to be supplemented with a more flexible approach. Then, the semantic knowledge base of ICD-11 should better guide the content related input fields of a future electronic death certificate. Full article
Show Figures

Figure 1

17 pages, 2025 KiB  
Systematic Review
Generative Adversarial Networks (GANs) in the Field of Head and Neck Surgery: Current Evidence and Prospects for the Future—A Systematic Review
by Luca Michelutti, Alessandro Tel, Marco Zeppieri, Tamara Ius, Edoardo Agosti, Salvatore Sembronio and Massimo Robiony
J. Clin. Med. 2024, 13(12), 3556; https://doi.org/10.3390/jcm13123556 - 18 Jun 2024
Viewed by 443
Abstract
Background: Generative Adversarial Networks (GANs) are a class of artificial neural networks capable of generating content such as images, text, and sound. For several years already, artificial intelligence algorithms have shown promise as tools in the medical field, particularly in oncology. Generative Adversarial [...] Read more.
Background: Generative Adversarial Networks (GANs) are a class of artificial neural networks capable of generating content such as images, text, and sound. For several years already, artificial intelligence algorithms have shown promise as tools in the medical field, particularly in oncology. Generative Adversarial Networks (GANs) represent a new frontier of innovation, as they are revolutionizing artificial content generation, opening opportunities in artificial intelligence and deep learning. Purpose: This systematic review aims to investigate what the stage of development of such technology is in the field of head and neck surgery, offering a general overview of the applications of such algorithms, how they work, and the potential limitations to be overcome in the future. Methods: The Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines were followed in conducting this study, and the PICOS framework was used to formulate the research question. The following databases were evaluated: MEDLINE, Embase, Cochrane Central Register of Controlled Trials (CENTRAL), Scopus, ClinicalTrials.gov, ScienceDirect, and CINAHL. Results: Out of 700 studies, only 9 were included. Eight applications of GANs in the head and neck region were summarized, including the classification of craniosynostosis, recognition of the presence of chronic sinusitis, diagnosis of radicular cysts in panoramic X-rays, segmentation of craniomaxillofacial bones, reconstruction of bone defects, removal of metal artifacts from CT scans, prediction of the postoperative face, and improvement of the resolution of panoramic X-rays. Conclusions: Generative Adversarial Networks may represent a new evolutionary step in the study of pathology, oncological and otherwise, making the approach to the disease much more precise and personalized. Full article
Show Figures

Figure 1

18 pages, 3460 KiB  
Article
An Enhanced Feature Extraction Framework for Cross-Modal Image–Text Retrieval
by Jinzhi Zhang, Luyao Wang, Fuzhong Zheng, Xu Wang and Haisu Zhang
Remote Sens. 2024, 16(12), 2201; https://doi.org/10.3390/rs16122201 - 17 Jun 2024
Viewed by 360
Abstract
In general, remote sensing images depict intricate scenes. In cross-modal retrieval tasks involving remote sensing images, the accompanying text includes numerus information with an emphasis on mainly large objects due to higher attention, and the features from small targets are often omitted naturally. [...] Read more.
In general, remote sensing images depict intricate scenes. In cross-modal retrieval tasks involving remote sensing images, the accompanying text includes numerus information with an emphasis on mainly large objects due to higher attention, and the features from small targets are often omitted naturally. While the conventional vision transformer (ViT) method adeptly captures information regarding large global targets, its capability to extract features of small targets is limited. This limitation stems from the constrained receptive field in ViT’s self-attention layer, which hinders the extraction of information pertaining to small targets due to interference from large targets. To address this concern, this study introduces a patch classification framework based on feature similarity, which establishes distinct receptive fields in the feature space to mitigate interference from large targets on small ones, thereby enhancing the ability of traditional ViT to extract features from small targets. We conducted evaluation experiments on two popular datasets—the Remote Sensing Image–Text Match Dataset (RSITMD) and the Remote Sensing Image Captioning Dataset (RSICD)—resulting in mR indices of 35.6% and 19.47%, respectively. The proposed approach contributes to improving the detection accuracy of small targets and can be applied to more complex image–text retrieval tasks involving multi-scale ground objects. Full article
Show Figures

Figure 1

16 pages, 4140 KiB  
Article
MFSC: A Multimodal Aspect-Level Sentiment Classification Framework with Multi-Image Gate and Fusion Networks
by Lingling Zi, Xiangkai Pan and Xin Cong
Electronics 2024, 13(12), 2349; https://doi.org/10.3390/electronics13122349 - 15 Jun 2024
Viewed by 249
Abstract
Currently, there is a great deal of interest in multimodal aspect-level sentiment classification using both textual and visual information, which changes the traditional use of only single-modal to identify sentiment polarity. Considering that existing methods could be strengthened in terms of classification accuracy, [...] Read more.
Currently, there is a great deal of interest in multimodal aspect-level sentiment classification using both textual and visual information, which changes the traditional use of only single-modal to identify sentiment polarity. Considering that existing methods could be strengthened in terms of classification accuracy, we conducted a study on aspect-level multimodal sentiment classification with the aim of exploring the interaction between textual and visual features. Specifically, we construct a multimodal aspect-level sentiment classification framework with multi-image gate and fusion networks called MFSC. MFSC consists of four parts, i.e., text feature extraction, visual feature extraction, text feature enhancement, and multi-feature fusion. Firstly, a bidirectional long short-term memory network is adopted to extract the initial text feature. Based on this, a text feature enhancement strategy is designed, which uses text memory network and adaptive weights to extract the final text features. Meanwhile, a multi-image gate method is proposed for fusing features from multiple images and filtering out irrelevant noise. Finally, a text-visual feature fusion method based on an attention mechanism is proposed to better improve the classification performance by capturing the association between text and images. Experimental results show that MFSC has advantages in classification accuracy and macro-F1. Full article
Show Figures

Figure 1

26 pages, 3411 KiB  
Article
Predicting the Direction of NEPSE Index Movement with News Headlines Using Machine Learning
by Keshab Raj Dahal, Ankrit Gupta and Nawa Raj Pokhrel
Econometrics 2024, 12(2), 16; https://doi.org/10.3390/econometrics12020016 - 11 Jun 2024
Viewed by 1095
Abstract
Predicting stock market movement direction is a challenging task due to its fuzzy, chaotic, volatile, nonlinear, and complex nature. However, with advancements in artificial intelligence, abundant data availability, and improved computational capabilities, creating robust models capable of accurately predicting stock market movement is [...] Read more.
Predicting stock market movement direction is a challenging task due to its fuzzy, chaotic, volatile, nonlinear, and complex nature. However, with advancements in artificial intelligence, abundant data availability, and improved computational capabilities, creating robust models capable of accurately predicting stock market movement is now feasible. This study aims to construct a predictive model using news headlines to predict stock market movement direction. It conducts a comparative analysis of five supervised classification machine learning algorithms—logistic regression (LR), support vector machine (SVM), random forest (RF), extreme gradient boosting (XGBoost), and artificial neural network (ANN)—to predict the next day’s movement direction of the close price of the Nepal Stock Exchange (NEPSE) index. Sentiment scores from news headlines are computed using the Valence Aware Dictionary for Sentiment Reasoning (VADER) and TextBlob sentiment analyzer. The models’ performance is evaluated based on sensitivity, specificity, accuracy, and the area under the receiver operating characteristic (ROC) curve (AUC). Experimental results reveal that all five models perform equally well when using sentiment scores from the TextBlob analyzer. Similarly, all models exhibit almost identical performance when using sentiment scores from the VADER analyzer, except for minor variations in AUC in SVM vs. LR and SVM vs. ANN. Moreover, models perform relatively better when using sentiment scores from the TextBlob analyzer compared to the VADER analyzer. These findings are further validated through statistical tests. Full article
Show Figures

Figure 1

19 pages, 1876 KiB  
Article
Improving Training Dataset Balance with ChatGPT Prompt Engineering
by Mateusz Kochanek, Igor Cichecki, Oliwier Kaszyca, Dominika Szydło, Michał Madej, Dawid Jędrzejewski, Przemysław Kazienko and Jan Kocoń
Electronics 2024, 13(12), 2255; https://doi.org/10.3390/electronics13122255 - 8 Jun 2024
Viewed by 477
Abstract
The rapid evolution of large language models, in particular OpenAI’s GPT-3.5-turbo and GPT-4, indicates a growing interest in advanced computational methodologies. This paper proposes a novel approach to synthetic data generation and knowledge distillation through prompt engineering. The potential of large language models [...] Read more.
The rapid evolution of large language models, in particular OpenAI’s GPT-3.5-turbo and GPT-4, indicates a growing interest in advanced computational methodologies. This paper proposes a novel approach to synthetic data generation and knowledge distillation through prompt engineering. The potential of large language models (LLMs) is used to address the problem of unbalanced training datasets for other machine learning models. This is not only a common issue but also a crucial determinant of the final model quality and performance. Three prompting strategies have been considered: basic, composite, and similarity prompts. Although the initial results do not match the performance of comprehensive datasets, the similarity prompts method exhibits considerable promise, thus outperforming other methods. The investigation of our rebalancing methods opens pathways for future research on leveraging continuously developed LLMs for the enhanced generation of high-quality synthetic data. This could have an impact on many large-scale engineering applications. Full article
Show Figures

Figure 1

30 pages, 1001 KiB  
Article
Genre Classification of Books in Russian with Stylometric Features: A Case Study
by Natalia Vanetik, Margarita Tiamanova, Genady Kogan and Marina Litvak
Information 2024, 15(6), 340; https://doi.org/10.3390/info15060340 - 7 Jun 2024
Viewed by 431
Abstract
Within the literary domain, genres function as fundamental organizing concepts that provide readers, publishers, and academics with a unified framework. Genres are discrete categories that are distinguished by common stylistic, thematic, and structural components. They facilitate the categorization process and improve our understanding [...] Read more.
Within the literary domain, genres function as fundamental organizing concepts that provide readers, publishers, and academics with a unified framework. Genres are discrete categories that are distinguished by common stylistic, thematic, and structural components. They facilitate the categorization process and improve our understanding of a wide range of literary expressions. In this paper, we introduce a new dataset for genre classification of Russian books, covering 11 literary genres. We also perform dataset evaluation for the tasks of binary and multi-class genre identification. Through extensive experimentation and analysis, we explore the effectiveness of different text representations, including stylometric features, in genre classification. Our findings clarify the challenges present in classifying Russian literature by genre, revealing insights into the performance of different models across various genres. Furthermore, we address several research questions regarding the difficulty of multi-class classification compared to binary classification, and the impact of stylometric features on classification accuracy. Full article
(This article belongs to the Special Issue Text Mining: Challenges, Algorithms, Tools and Applications)
Show Figures

Figure 1

18 pages, 1821 KiB  
Article
Enhancing Medical Image Retrieval with UMLS-Integrated CNN-Based Text Indexing
by Karim Gasmi, Hajer Ayadi and Mouna Torjmen
Diagnostics 2024, 14(11), 1204; https://doi.org/10.3390/diagnostics14111204 - 6 Jun 2024
Viewed by 342
Abstract
In recent years, Convolutional Neural Network (CNN) models have demonstrated notable advancements in various domains such as image classification and Natural Language Processing (NLP). Despite their success in image classification tasks, their potential impact on medical image retrieval, particularly in text-based medical image [...] Read more.
In recent years, Convolutional Neural Network (CNN) models have demonstrated notable advancements in various domains such as image classification and Natural Language Processing (NLP). Despite their success in image classification tasks, their potential impact on medical image retrieval, particularly in text-based medical image retrieval (TBMIR) tasks, has not yet been fully realized. This could be attributed to the complexity of the ranking process, as there is ambiguity in treating TBMIR as an image retrieval task rather than a traditional information retrieval or NLP task. To address this gap, our paper proposes a novel approach to re-ranking medical images using a Deep Matching Model (DMM) and Medical-Dependent Features (MDF). These features incorporate categorical attributes such as medical terminologies and imaging modalities. Specifically, our DMM aims to generate effective representations for query and image metadata using a personalized CNN, facilitating matching between these representations. By using MDF, a semantic similarity matrix based on Unified Medical Language System (UMLS) meta-thesaurus, and a set of personalized filters taking into account some ranking features, our deep matching model can effectively consider the TBMIR task as an image retrieval task, as previously mentioned. To evaluate our approach, we performed experiments on the medical ImageCLEF datasets from 2009 to 2012. The experimental results show that the proposed model significantly enhances image retrieval performance compared to the baseline and state-of-the-art approaches. Full article
(This article belongs to the Special Issue Medical Data Processing and Analysis—2nd Edition)
Show Figures

Figure 1

Back to TopTop