Book Volume 2
Preface
Page: i-i (1)
Author: Pankaj Kumar Mishra and Satya Prakash Yadav
DOI: 10.2174/9789815305395125020001
A Method Based on Machine Learning to Classify Text for the Field of Cybersecurity
Page: 1-12 (12)
Author: Siddharth Sriram*
DOI: 10.2174/9789815305395125020003
PDF Price: $30
Abstract
Rapid advancements in networks and computer systems have opened a new door for immoral acts like cybercrime, which threaten public safety, and security, as well as the global economy. The purpose of this proposal is to analyse IP fraud and cyberbullying as two distinct types of cybercrime. The primary goals of this study are to use instances of cybercrime to provide a short examination of cybercrime activities, and the family member principles, and propose a pairing schema. Using the Naive Bayes (NB) & Support Vector Machine (SVM) artificial intelligence techniques, cybercrime instances are categorised according to their ideal qualities. The Twitter data in the Kaggle database has been clustered using K-means. User ID, sign-up date, referral, browser, gender, and age as well as IP address are just a few of the most useful information used to educate the computer. Total 151,113 datasets were used for experimental analysis of the suggested algorithm's performance. The accuracy of the suggested approach, 97%, is higher than that of the current method (NB). The challenge of regression may be easily surmounted with the use of the random forest method for the categorization of the resultant cybercrimes. The planned study uses age categories as the foundation for identifying the different offenses.
A Practicable E-commerce-Based TextClassification System
Page: 13-22 (10)
Author: Sidhant Das*
DOI: 10.2174/9789815305395125020004
PDF Price: $30
Abstract
This article examines the features of the dealer's brush list evaluation material in light of research findings on misleading assessment and identification of online purchasing. A Gated Recurrent Unit (GRU) model using keyword weighting is presented as a solution to the issue that it is challenging for the DL model to collect the feature data of the whole assessment text in a false evaluation identifying job. The TFIDF technique is first used to generate the list of keywords, and then that list's weight is applied to the word vector. Finally, a weighted vector of words is categorised using this method of the model to finish the recognition job of erroneous evaluation, replacing the pooling component of the GRU model with a constrained Boltzmann machine. By using a variety of text categorization algorithms and comparing their results in terms of correctness and performance, this research aspires to represent the practical benefits of applications that use machine learning in the real world. We built a system that can run several text classification algorithms, and we used that system to create models that were educated using actual data taken from E-Commerce, a virtual fashion e-commerce platform. The Convolutional Neural Network technique achieved the greatest mean accuracy of 96.08% (with a range of 85.44% to 99.99%) with an average deviation of 5.65%.
AI Model for Text Classification Using FastText
Page: 23-32 (10)
Author: Sorabh Sharma*
DOI: 10.2174/9789815305395125020005
PDF Price: $30
Abstract
The purpose of text categorization, a machine learning technique, is to automatically assign tags or categories to texts. Natural language processing (NLP)- based text classifiers can quickly analyse vast volumes of text and classify it based on emotions, themes, and human intent. FastText was created by Facebook's AI Research team and is available to the public as a free library. Its primary goal is the efficient and accurate processing of big datasets in order to provide scalable remedies for the problems of text categorization and representation. Traditional machine learning techniques used in most text categorization models suffer from issues including the curse of dimensionality and subpar performance. This research offers a fastText-based AI text classification model to address the aforementioned issues. The fastText approach allows our model to create a low-dimensional, continuous, and high-quality representation of text by mining the text for relevant information through feature engineering. The experiment uses Python to define the text dataset, and the results demonstrate that our model outperforms the baseline model trained using classic ML methods in terms of accuracy, recall, and F values.
An Algorithm for Textual Classification of News Utilizing Artificial Intelligence Technology
Page: 33-43 (11)
Author: Rahul Mishra*
DOI: 10.2174/9789815305395125020006
PDF Price: $30
Abstract
The rate at which technology is improving is increasing all throughout the world. Every day, a tremendous amount of textual data is produced as a result of the Internet, websites, business data, medical information, and the media. Extraction of interesting patterns from text data with varied lengths such as views, summaries, and facts is a challenging issue. This work provides a deep learning (DL) algorithm-based approach to news text classification to address the issues of large amounts of text data and cumbersome features obtaining value in news. Although the relationship among words as well as categories has a significant impact on the categorization of news text, previous approaches to text classification relied solely on the knowledge of the connections between words to make their classification decisions. This research uses the idea of a tailored algorithm to provide a CNN, LSTM, and MLP-based customizable ensemble framework for categorising news text data. The proposed model is based on a parallel representation of word vectors and word dispersion. We feed the term vector to the CNN module to convey the relationship between words, as well as nourish the discrete vector corresponding to the relationship between words and categories into the MLP module to achieve deep learning of the spatial data on features, the time-series feature information, and the connection words and categories in news texts. Extensive experimental study confirmed the dependability and efficacy of the proposed approach. The experimental results demonstrated that the proposed method improved the most - in terms of precision, recall rate, and comprehensive value, while also addressing the problems of text length, extraction of features issues with the news text, and classification of news text.
Analysis of the Sentiment of Tweets Regarding COVID-19 Vaccines Using Natural Language Processing and Machine Learning Sectionification Algorithms
Page: 44-56 (13)
Author: Sukhman Ghumman*
DOI: 10.2174/9789815305395125020007
PDF Price: $30
Abstract
The unique Coronavirus pandemic of 2019 (called COVID-19 by the globe Health Organisation) has exposed several governments throughout the globe to risk. The Covid-19 epidemic, which had previously only affected the Chinese population, is now a major worry for countries all over the globe. Additionally to the obvious health effects of COVID-19 epidemic, this study reveals its repercussions on the worldwide economy. The research went on to talk about how they analysed public opinion and learned new things about Covid-19 vaccinations by using content Analytics and sentiment evaluation in Natural Language Processing (NLP) using content from Twitter. To categorise and analyse the outcomes, researchers used two machine learning algorithms: logistic regression (LR), random forest, decision tree, and convolutional neural networks (CNNs). To better identify public opinion, several preprocessing methods were used and categorised responses into neutral, positive, and negative categories. The public's opinion on Covid-19 vaccinations is 31% favourable, 22% negative, and 47% neutral, according to the results of the emotion section distribution. CNN achieved 98% accuracy, according to the tested machine learning algorithms.
Classification of Medical Text using ML and DL Techniques
Page: 57-68 (12)
Author: Sulabh Mahajan*
DOI: 10.2174/9789815305395125020008
PDF Price: $30
Abstract
The use of sarcasm in everyday conversation has recently risen to prominence. All the kids of our age utilise sarcasm to convey a negative message in a more nuanced manner. With the advancement of AI and machine learning techniques in the area of natural language processing (NLP), it has become more difficult to reliably and effectively identify sarcasm. This research provides a new method for sarcasm detection using machine studying and deep learning, hoping to make a meaningful contribution to this expanding area of study. In order to prepare the phrase for a hybrid deep learning model for training and classification, this method employs bidirectional encoder representations from transformers (BERT). The combination of CNN and LSTM creates the hybrid model used here. The suggested model has been tested on two datasets, with the goal of identifying sarcasm from nonsarcastic words. The trained model obtained a 99.63% accuracy rate, a 99.33% precision rate, a 99.83% recall rate, and an F1-score of 99.56%. These findings are based on 10 rounds of crossvalidation performed on the model that was suggested using the medical datasets.
Evaluation of ML and Advanced Deep Learning Text Classification Systems
Page: 69-80 (12)
Author: Tarun Kapoor*
DOI: 10.2174/9789815305395125020009
PDF Price: $30
Abstract
Classifying texts into groups determined by their content is called text classification. In this process, automatic labelling of documents written in natural languages is carried out according to predetermined labels. Both text comprehension systems, which perform transformations on texts such as creating summaries, answering queries, and extracting data, and text-retrieving systems, which obtain texts in fulfillment of a user query, rely heavily on text categorization. In order to learn effectively, current algorithms for supervised learning for text classification need a large enough training set. This research introduces a novel text categorization algorithm that uses artificial intelligence techniques (machine studying and deep learning techniques) and needs fewer documents for training than previous methods. To generate a feature set from already-classified text documents, we resort to “word relation,” or association rules based on these words. To classify the data, we use the idea of a Convolutional Neural Network with Deep Convolution to the extracted features and then employ a single genetic algorithm approach. The suggested method has been built and thoroughly tested in a working system. The results of the experiments show that the suggested system is an effective text classifier.
Machine Learning Method Employed for the Objective of Identifying Text on Tweet Dataset
Page: 81-91 (11)
Author: Sakshi Pandey*
DOI: 10.2174/9789815305395125020010
PDF Price: $30
Abstract
When it comes to training ML systems, internet-based data is invaluable. Despite the difficulty in collecting this information, teams of experts from academic institutions and research labs have created publicly accessible databases. Twitter and other social media platforms provided large quantities of useful information throughout the pandemic, which was used to evaluate healthcare decisions. In order to forecast illness prevalence and offer early warnings, we suggest analysing user attitudes by using efficient supervised machine learning algorithms. The gathered tweets were sorted into positive, negative, and neutral categories for preprocessing. Hybrid feature extraction is the innovative aspect of our work; we used it to correctly describe posts by combining syntactic features (TF-IDF) and semantic elements (FastText and Glove), which in turn improved classification. The experimental findings suggest that when using Naive Bayes, the combination of FastText and TF-IDF achieves the best results.
Textual Classification Utilizing the Integration of Semantics and Statistical Methodology
Page: 92-102 (11)
Author: Ayush Gandhi*
DOI: 10.2174/9789815305395125020011
PDF Price: $30
Abstract
Effectively classifying texts is possible using several classification techniques. Machine learning constructs a classifier by studying and memorising the characteristics of several classes. For text categorization, deep learning provides similar advantages since it can function with great precision using simpler architecture and processing. In order to categorise textual information, this research makes use of machine learning and deep learning methods. There is a great deal of extraneous details in textual data that must be removed during pre-processing. To prepare it for analysis, we remove duplicate columns and impute missing data. In the next step, we use deep learning techniques for classification, including long short-term memory (LSTM), artificial neural network (ANN), as well as gated recurrent unit (GRU). According to the findings, GRU obtains 92% accuracy, which is higher than that of any other model or baseline investigation.
The Use of Machine Learning Techniques to Classify Content on the Web
Page: 103-114 (12)
Author: Dikshit Sharma*
DOI: 10.2174/9789815305395125020012
PDF Price: $30
Abstract
In text categorization, texts are sorted into groups according to their content. It is the process of automatically classifying texts written in natural languages according to a set of guidelines. Both text comprehension systems, which perform transformations on text such as creating summaries, answering queries, and extracting data, and retrieval of text systems, which collect texts in reaction to a user query on the internet content, rely heavily on text categorization. In order to learn effectively, current algorithms for supervised learning for text classification need a large enough training set. This research introduces a novel text categorization system that makes use of an AI approach and needs fewer articles for training over information found on the web. To generate a feature set from already categorised texts, we resort to “word relation,” or association rules based on these terms. The obtained characteristics are then processed by a Support Vector Machine, and ultimately, a single genetic algorithm idea is introduced for classification. The suggested approach has been developed and validated in a working system. The experimental results verify the effectiveness of the proposed system as a text classifier.
Lexical Methods for Identifying Emotions in Text Based on Machine Learning
Page: 115-126 (12)
Author: Mridula Gupta*
DOI: 10.2174/9789815305395125020013
PDF Price: $30
Abstract
The study of emotions has emerged as an important area of research because of the wealth of information it can provide. Emotions can be expressed in a variety of ways including words, facial expressions, written material, and movements. Natural language processing (NLP) & deep learning concepts are essential to solving the content-based classification problem that is emotion detection in a text document. Therefore, in this research, we suggest using deep learning to aid semantic text analysis in the task of identifying human emotions from transcripts of spoken language. Visual forms of expression, such as makeover jargon, may be used to convey the feeling. Datasets of recorded voices from people with Autism Spectrum Disorder (ASD) are transcribed for analysis. However, in this paper, we specialize in detecting emotions from all of the textual dataset and using the semantic data enhancement process to fill a few of the phrases, or half-broken speech, as patients with Autism Spectrum Disorder (ASD) lack social contact skills due to the patient not very well articulating their communication.
Identification of Websites Using an Efficient Method Employing Text Mining Methods
Page: 127-138 (12)
Author: Madhur Taneja*
DOI: 10.2174/9789815305395125020014
PDF Price: $30
Abstract
Herein, we introduce a method for website classification using deep neural networks and mixed data extractors. We use iterative training as well as supervised learning approaches to use a gradient descent methodology to simulate the website categorization. This modern model is comprised of a webpage encoder, a convolutional neural network (CNN) feature extraction, a bidirectional long short-term memory (LSTM) feature extractor, as well as a fully connected classifier. It may retrieve various website features at various granularities. Our model may quickly select a suitable website class by concatenating mixed features obtained from mixed feature extractors. On the realistic website dataset that has been obtained, we conduct in-depth tests. The dataset is compiled using domains that were taken from the telecom operator's DNS records. The proposed categorization schema outperforms state-of-the-art models in comparison to our fresh model as well as a slew of popular machine learning algorithms in terms of accuracy, recall, F1, and precision. Other web apps may benefit from all of this as well, such as detecting fake websites as well as ads.
Machine Learning-based High-Dimensional Text Document Classification and Clustering
Page: 139-149 (11)
Author: Ansh Kataria*
DOI: 10.2174/9789815305395125020015
PDF Price: $30
Abstract
Text classification is a difficult technique. Many techniques have been developed to decrease the dimension of feature vectors for use in text classification due to their enormous size. This work provides a detailed discussion of unique parameters utilising an optic clustering strategy, as well as a review of some of the most essential text categorization algorithms. In this case, the words are clustered according to their level of similarity. Each cluster's membership function is based on the mean along with the standard deviation of its data. Finally, characteristics are chosen from each grouping. Each cluster's extracted feature is the weighted sum of its words. There's also no need to guess or use trial-and-error approaches to determine the optimal number of clusters.
The Application of an N-Gram Machine Learning Method to the Text Classification of Healthcare Transcriptions
Page: 150-159 (10)
Author: Pratibha Sharma*
DOI: 10.2174/9789815305395125020016
PDF Price: $30
Abstract
An integral aspect of natural language processing is text categorization, the goal of which is to assign a predetermined category to a given text. Feature selection and categorization models come in a wide variety of forms. Most researchers, however, would rather utilise the prepackaged functions of existing libraries. In the field of natural language processing (NLP), automated medical text categorization is very helpful for decoding the information hidden in clinical descriptions. Machine learning approaches seem to be fairly successful for medical text categorization problems; nevertheless, substantial human work is required in order to provide labelled training data. Clinical and translational research has benefited greatly from the computerised collection of vast amounts of precise patient information, including illness status, blood tests, medications taken, and side effects, along with therapy results. As a result, the medical literature contains a massive amount of information on individual patients, making it very difficult to digest. In this research, we suggest using N-grams and a Support Vector Machine (SVM) to classify healthcare-related texts. We conduct experiments to determine the viability of our code and analyse it across a variety of categorization methods.
Method for Adaptive Combination of Multiple Features for Text Classification in Agriculture
Page: 160-170 (11)
Author: Jaskirat Singh*
DOI: 10.2174/9789815305395125020017
PDF Price: $30
Abstract
When applying conventional text classification techniques, the values in agricultural text are converted into characters, which destroys the original semantic representation of numerical aspects. A unique text classification approach is suggested, based on the dynamic fusion of several characteristics, to completely mine the deep latent semantic characteristics in agricultural literature. The global key semantic characteristics of the text were extracted using the Bi-directional Gated Recurrent Neural Networks (GRU) model with attention mechanism, while the local semantic data about the text at various levels was extracted using the multiple windows Convolution Neural Network. Finally, the number that features essential semantic expressions was obtained using a machine learning approach for creating the quantitative value feature vector. To further enhance the deep semantic expression found in agricultural text and successfully improve the impact of farm text categorization with phenotypic numerical type, we use a focus technique to dynamically fuse the derived numerous semantic characteristics.
Deep Learning-based Text-Retrieval System with Relevance Feedback
Page: 171-180 (10)
Author: Simran Kalra*
DOI: 10.2174/9789815305395125020018
PDF Price: $30
Abstract
We presented an Information Retrieval (IR) system that learns from existing information and creates a single pertinent document that, we presume, has all its indexed pertinent details for a query. Deep learning makes such a system viable. We then asked people to score the query plus word-cloud representation of three randomly selected relevant texts and our new synthetic document. The synthetic document topped all inquiries and users. We then trained a CNN using query-relevant data. We performed ”deep learn” function on a synthetic, relevant material using the CNN. We used crowdsourcing to compare the “deep-learned” material to related documents. Users can see a query and four-word cloud (three relevant documents and our deep learning synthetic document). The synthetic document provides the the most relevant feedback.
Domain Knowledge-based BERT Model with Deep Learning for Text Classification
Page: 181-189 (9)
Author: Akhilesh Kalia*
DOI: 10.2174/9789815305395125020019
PDF Price: $30
Abstract
Lexical model BERT already trained on BookCorpus and Wikipedia works well on two NLP tasks after downstream fine-tuning. The requirements of BERT model include strategy analyses and task-specific and domain-related data. The problems of task awareness as well as instruction data in BERT-DL, a BERT-based text-classification model, are addressed through auxiliary sentences. The pre-training, training, and post-training steps for BERT4TC's domain challenges are all provided. Learning speed, sequencing duration, and secret state vectors that select fine-tuning are all investigated in extended trials over 7 public datasets. The BERT4TC model is then contrasted using a variety of auxiliary terms and post-training goals. On multiple-class datasets, BERT4TC with the ideal auxiliary phrase outperforms previous state-of-theart feature-based algorithms and fine-tuning approaches. Our domain-related corpustrained BERT4TC beats BERT on binary sentiment categorization datasets.
Applying Deep Learning to Classify Massive Amounts of Text Using Convolutional Neural Systems
Page: 190-201 (12)
Author: Shubhansh Bansal*
DOI: 10.2174/9789815305395125020020
PDF Price: $30
Abstract
Supervised learning based on deep learning is often used for mass-scale picture categorization. However, it takes a lot of computing effort and energy to retrain these vast networks to accept new, unknown data. When retraining, it is possible that training samples used before would not be accessible. We present a scalable, gradually expanding CNN that can learn new jobs while reusing some of the base networks and an efficient training mechanism. Our approach takes cues from transfer learning methods, but unlike other approaches, it retains knowledge of previously mastered tasks. Convolutional layers from the early section of the base network are reused in the updated network, and a few more convolutional kernels are added to the later layers to facilitate learning a new set of classes. On the task of categorising texts, we tested the suggested method. Our method achieves comparable classification accuracy to the standard incremental learning method in which networks are updated solely with new training samples, without any network sharing), while also being more resourcefriendly and taking less time and space to train.
An Algorithm for Categorizing Opinions in Text from Various Social Media Platforms
Page: 202-212 (11)
Author: Pavas Saini*
DOI: 10.2174/9789815305395125020021
PDF Price: $30
Abstract
More individuals are sharing their thoughts and feelings through internet videos as social media platforms proliferate. While successful emotional fusion in multimodal data is a key component of multimodal sentiment analysis, most existing research falls short in this area. Predicting users' emotional inclinations through their expressions of language is made easier by multi-modal sentiment detection. As a result, the field of multi-modal sentiment detection has grown rapidly in recent years. As a result, multimodal sentiment analysis is quickly rising to the forefront of academic interest. However, in actual social media, visuals and sentences do not always work together to represent emotional polarity. Additionally, there are several information modalities, each contributing in its unique way to the overall emotional polarity. A multimodal approach to sentiment analysis that takes into account contextual knowledge is presented as a solution to these issues. The approach begins by mining social media texts for subject information that comprehensively describes the comment material. Additionally, we use cutting-edge pre-training models to identify emotional qualities that span several domains. Then, we provide methods for merging features at different levels, such as cross-modal global fusion as well as cross-modal high-level semantics fusing. At long last, we run our tests on a multimodal dataset that really exists in the real world. Results show that the proposed approach can correctly classify the tone of heterogeneous online reviews, and it also outperforms the standard approach in many other ways as well.
Text Classification Method for Tracking Rare Events on Twitter
Page: 213-225 (13)
Author: Prabhjot Kaur*
DOI: 10.2174/9789815305395125020022
PDF Price: $30
Abstract
A natural catastrophe is an example of a rare occurrence that does not happen often but may have devastating effects on people and their environment when it occurs. People now have a quick and easy outlet for voicing their ideas thanks to social media. Thus, it may be used by researchers to learn about how individuals react to and think about a wide variety of extremely unusual occurrences. Many research works use social media data to investigate how people's reactions to unusual occurrences in the real world translate to their online personas, thoughts, feelings, and actions. In this piece of work, we offer a method for extracting features and classifying tweets on unusual events like Hurricane Sandy. To begin, a new approach to feature extraction is presented, one that may be used to extract relevant features from each communication. The next step is to offer a Score-based categorization system for differentiating between communications about events and those that are unrelated. Finally, the development of a rare event is analyzed using our suggested approach and the popular keyword search method. The findings show that the suggested method is effective in distinguishing between text messages connected to unusual events and those that are unrelated.
Text Document Preprocessing and Classification Using SVM and Improved CNN
Page: 226-237 (12)
Author: Jaspreet Sidhu*
DOI: 10.2174/9789815305395125020023
PDF Price: $30
Abstract
Text categorization is a crucial technology in data mining as well as data retrieval that has been extensively investigated and is developing at a rapid pace. Convolutional neural networks (CNNs) are a kind of deep learning modeling that may reduce the complexity of the model while accurately extracting characteristics from input text. Support vector machine (SVM) results have always been more trustworthy and superior to those of other traditional artificial intelligence approaches. Using enhanced convolutional neural network (CNNs) as well as support vector machines (SVMs), we offer a novel approach to online text categorization in this study. Our approach begins with text attribute identification and prediction using a model based on CNN with a five-layer network structure. Databases including both text and images will find it to be a major factor in the long run.
Identification of Text Emotions Through the Use of Convolutional Neural Network Models
Page: 238-248 (11)
Author: Vaibhav Kaushik*
DOI: 10.2174/9789815305395125020024
PDF Price: $30
Abstract
Increasing numbers of people are using the Internet to share their feelings and communicate with one another, and the vast majority of these expressions of emotion take the form of text. Using sentiment dictionaries, machine learning, and deep learning are the three most common approaches to text sentiment categorization studies. Due to the exponential growth of textual data, it is crucial to create models that can automatically analyse this material. Labels like gender, age, nationality, emotion, etc., may be included in the texts. Numerous investigations of text categorization have emerged because the use of such labels may be useful in various commercial sectors. The Convolutional Neural Network, also referred to as CNN, was recently utilised to the problem of text categorization, with promising results. In this study, we advocate for the use of convolutional neural network networks for the job of classifying emotions. Using three popular datasets, we demonstrate that our networks outperform existing cutting-edge deep learning models by using successive convolutional layers to process substantially longer sentences.
Classification & Clustering of Text Based on Doc2Vec & K-means Clustering based Similarity Measurements
Page: 249-260 (12)
Author: Prakriti Kapoor*
DOI: 10.2174/9789815305395125020025
PDF Price: $30
Abstract
One crucial task in text processing is determining how similar two papers are to one another. A novel similarity metric is suggested in this study. Finding a suitable similarity metric for written materials that permits the development of coherent groupings is a significant difficulty for document clustering. After that, we use TFIDF to build a vector space, and then we use the ward's approach and the K-means algorithm to accomplish clustering. WordNet is additionally employed in the process of semantic document clustering. Visualisations and an interactive website illustrating the connections between all clusters illustrate the findings. The existence (and quantity) of words in texts are all that are taken into account while utilising the traditional bag-ofwords paradigm. This process might lead to texts with identical meanings but distinct vocabulary being placed in various groups. The findings acquired using the suggested approach are analysed for their correctness using the F-measure. Comparisons using the sentence vectors model (Doc2vec) and the bag-of-words model are made to confirm the edge of the suggested strategy. The suggested methodology may be used to decipher web chat logs and client feedback posted online. We evaluate our method on a variety of real-world data sets including examples of text classification and clustering problems. The findings prove that the proposed measure outperforms competing strategies.
Categorization of COVID-19 Twitter Data Based on an Aspect-Oriented Sentiment Analysis and Fuzzy Logic
Page: 261-271 (11)
Author: Tarang Bhatnagar*
DOI: 10.2174/9789815305395125020026
PDF Price: $30
Abstract
During times of disaster or epidemic, social media has emerged as a vital means of communication. It is difficult to examine the complete situational awareness via many elements and emotions to aid authorities due to the unpredictability of these calamities. Currently, systems for aspect recognition and sentiment analysis rely heavily on labelled data and require human curation of aspect categories. To analyze public opinion from a variety of angles, this study suggested a hybrid text analytical approach. Using the popular Latent Dirichlet Allocation (LDA) topic modeling, we first extracted and clustered the elements from the data. We then used the linguistic inquiry and word count (LIWC) lexicon to extract the sentiments and label the dataset. Finally, in the third layer of our structure, we mapped the elements into emotions, and the sentiments were classified using well-known machine learning classifiers. The comparison of our technique with other aspect-oriented sentiment analysis approaches shows encouraging results in experiments with actual datasets, and our method with several variants of classifiers surpasses current methods with top F1 scores of 91%.
Feature-Level Sentiment Analysis of Data Collected through Electronic Commerce
Page: 272-281 (10)
Author: Preetjot Singh*
DOI: 10.2174/9789815305395125020027
PDF Price: $30
Abstract
When dealing with data in the form of text, the most popular method for doing analysis and determining sentiment content is called “Sentiment Analysis.” Sentiment analysis is also known as Opinion Mining. Suggestions, feedback, tweets, and comments are all examples of the various types of text data that are being created. Customer feedback on e-commerce sites is a constant source of new information. Online stores may better meet client needs, improve their services, and boost sales by analyzing E-Commerce data. Positive, negative, and neutral feedback from customers may be separated using sentiment analysis. Numerous methods for Sentiment Analysis have been developed by academics. Typically, only a single machine learning algorithm is used for sentiment analysis. The purpose of this study, which makes use of Amazon review data, is to extract positive, negative, and neutral review ratings by locating aspect phrases, identifying the Parts-of-Speech, and applying classification algorithms to the collected data.
Classification Algorithms for Evaluating Customer Opinions using AI
Page: 282-293 (12)
Author: Saniya Khurana*
DOI: 10.2174/9789815305395125020028
PDF Price: $30
Abstract
There are now a great deal of consumer reviews of items that are written entirely in text. The reviews express their opinions. Opinion mining is another name for sentiment analysis. A common way for businesses to keep tabs on how customers feel about their brands and goods is through the use of sentiment analysis on textual data. Naive Bayes, Random Forest, Decision Tree, and Support Vector Machine classifiers are all used and compared in this study. In this study, we evaluate the efficacy of several classifiers by measuring their ability to correctly categorize mobile product data sets of varying sizes. Data were collected from popular online retailers like Amazon, Flipkart, and Snapdeal and analyzed to determine categorization accuracy. Naive Bayes, Random Forest, Decision Tree, and Support Vector Machines are some of the categorization algorithms compared here.
Analysis of Sentiment Employing the Word2vec with CNN-LSTM Classification System
Page: 294-305 (12)
Author: Rajat Saini*
DOI: 10.2174/9789815305395125020029
PDF Price: $30
Abstract
The identification of problems has become easier in sentiment categorization using conventional neural network–based short text classification methods . Word2vec, a convolutional neural network (CNN), and Bidirectional Long-term and Short-term Memory networks (LSTM) are used incombination to overcome this issue. Using Word2vec word embeddings, the CNN-LSTM model was able to attain an accuracy of 91.48%, as demonstrated experimentally. This demonstrates that the hybrid network model outperforms the single-structure neural network when dealing with relatively brief texts.
Hadoop-based Twitter Sentiment Analysis Using Deep Learning
Page: 306-315 (10)
Author: Manpreet Singh*
DOI: 10.2174/9789815305395125020030
PDF Price: $30
Abstract
Sentiment analysis, the practice of classifying and identifying opinions displayed in audio, words, database reports, and tweets to ascertain if the opinion is positive, neutral, or negative, is of great interest to many individuals in the microblogging service arena. It could be challenging to extract sentiment from Twitter data due to its quirks. The research suggests a way to analyze sentiment using a Hadoop infrastructure with a deep learning classifier. In order to gather characteristics, data is distributed across Hadoop nodes. After that, the data from Twitter is parsed for the most crucial parts. The input data from Twitter is sorted into two categories, positive review along with negative review, via a deep learning algorithm, including a deep recurrent neural networks classifier. Some of the metrics used to measure performance include classification accuracy, specificity, and sensitivity. With a sensitivity of 0.9404 and a generality of 0.9157, the proposed technique surpassed traditional methods in classification with a precision of 0.9302.
A Contrast Between Bert and Word2vec's Approaches to Text Sentiment Analysis
Page: 316-326 (11)
Author: Manish Nagpal*
DOI: 10.2174/9789815305395125020031
PDF Price: $30
Abstract
A novel approach to creating a dataset to train a neural network that can analyze the tone of social media postings is proposed in this study. The paper goes on to detail how the word2vec and BERT algorithms may be used in a neural network to analyze social media messages and evaluate their emotional tone. A hybrid of cosine similarity as well as ontological mappings based on a tweaked version of the Term Frequency-Inverse Document Frequency (TFIDF) characteristics is used by the algorithm for semantic searching. The execution includes sentiment analysis, phrase extraction process, textual belief indicator, keyword-based search, as well as text summary, among other things. Additionally, trials were carried out proving the efficacy of the methods presented. The efficiency of stemming and lemmatization of the text in creating a training set for sentiment analysis was also tested experimentally.
Text Emotion Categorization Using a Convolutional Recurrent Neural Network Enhanced by an Attention Mechanism-based Skip Gram Method
Page: 327-338 (12)
Author: Madhur Grover*
DOI: 10.2174/9789815305395125020032
PDF Price: $30
Abstract
Text-based web archives have become increasingly common as technology has advanced. For many text classification applications, classic machine learning classifiers like support vector machines (SVMs) and naïve Bayes (NBayes) perform well. Since short texts have fewer words and convolutional and pooling layers have their limits, these classifiers suffer from sparsity and lack long-term dependencies. In this study, we present a convolutional recurrent neural network architecture that makes use of a modified skip-gram method. For the adversarial training of the skip-gram algorithm, we employ the L2 regularization technique. It can boost the model's performance in text sentiment classification tasks and increase its robustness and generalizability. To extract information from the entire text while dampening the influence of irrelevant words, we deployed a convolutional neural network equipped with attention mechanisms. The CNN-based categorization of text emotion is complete. When compared to other classifiers used on the Twitter dataset, our model and algorithm were shown to be more efficient and accurate.
Multimodal Sentiment Analysis in Text, Images, and GIFs Using Deep Learning
Page: 339-349 (11)
Author: Deepak Minhas*
DOI: 10.2174/9789815305395125020033
PDF Price: $30
Abstract
More and more people are inclined to use GIFs, videos, and photographs on social media as a way to convey their feelings and thoughts. We developed a Pythonbased multimodal sentiment analysis tool for various Twitter formats, taking into account not just the text of a tweet but also its accompanying GIFs and pictures, for more precise sentiment scoring. We employ fine-tuned CNN for image sentiment analysis, VADER for text analysis, and image sentiment and facial expression analysis for GIFs, with each frame individually analyzed. Our research shows that combining textual and picture data yields superior outcomes compared to models that depend only on either images or text. The output scores from our text, picture, and GIF modules will be aggregated to get the final sentiment score for the incoming tweets.
Public Opinion Regarding COVID-19 Analyzed for Emotion Using Deep Learning Techniques
Page: 350-362 (13)
Author: Abhinav Mishra*
DOI: 10.2174/9789815305395125020034
PDF Price: $30
Abstract
As a result of the COVID-19 epidemic, many individuals are experiencing extreme worry, dread, and other difficult emotions. Since coronavirus immunizations were first introduced, people's reactions have gotten more nuanced and varied. In this study, we will use deep learning methods to decode their emotions. Twitter provides a glimpse into what is popular and what is on people's minds at any given time, and social media is presently the finest means to convey sentiments and emotions. Our goal while conducting this study was to have a better grasp of how different groups of individuals feel about vaccinations. The research period for this research's tweet was from December 21st to July 21st. Of the most talked-about vaccines that have recently been available in various regions of the world were the subject of several tweets. The term Valence Aware Sentiment Dictionary An NLP program called Believed (VADER) was used to examine people's sentiments on certain vaccines. We were better able to see the big picture after categorizing the collected attitudes into positive (33.96 percent), negative (17.55 percent), and neutral (48.49 percent) camps. We also included into our study an examination of the tweets' chronology, given that attitudes changed over time. The performance of the forecasting models was evaluated using an RNN-oriented design that included bidirectional LSTM (Bi-LSTM) as well as long short-term memory (LSTM); LSTM attained an accuracy of 90.59% as well as BiLSTM of 90.83%. Additional performance metrics, such as Precision, F1-score, as well as a matrix of confusion, were used to confirm our hypotheses as well as outcomes. The findings of this research provide credence to efforts to eradicate coronavirus across the globe by expanding our knowledge of public opinion on COVID-19 vaccines.
CNN-based Deep Learning Techniques for Movie Review Analysis of Sentiments
Page: 363-373 (11)
Author: Prateek Garg*
DOI: 10.2174/9789815305395125020035
PDF Price: $30
Abstract
Twitter, Facebook, Instagram, etc. are just a few of the many online discussion platforms that have sprung up as a result of the explosion in internet use and popularity, giving individuals a place to air their views on current events. Films get both acclamation and criticism from the general public. As a major form of entertainment, they inspire user evaluations of film and television on websites like IMDB and Amazon. Scientists and researchers give careful thought to these critiques and comments in order to extract useful information from the data. This data lacks organisation but is of critical importance nevertheless. Opinion mining, also known as sentiment classification, is a growing field that uses machine learning and deep learning to analyse the polarity of the feelings expressed in a review. Since text typically carries rich semantics useful for analysis, sentiment analysis has grown into the most active investigation in NLP (natural language processing). The continuous progress of deep learning has substantially increased the capacity to analyse this content. Convolutional Neural Networks (CNN) are commonly utilised for natural language processing since they are one of the most successful deep learning methodologies. This paper elaborates on the methods, datasets, outcomes, and limits of CNN-based sentiment analysis of film critics' reviews.
Machine Learning and Deep Learning Models for Sentiment Analysis of Product Reviews
Page: 374-385 (12)
Author: Saket Mishra*
DOI: 10.2174/9789815305395125020036
PDF Price: $30
Abstract
In our research, we use sentiment analysis to determine how well ratings and reviews are compared on Amazon.com. The process of determining whether a text's tone is favorable or negative and labelling it as such is known as sentiment analysis. Consumers may write evaluations on e-commerce sites like Amazon.com and indicate the polarity of their opinion. There is a discrepancy between the review and the rating in certain cases. We used deep learning to analyze the sentiment of Amazon.com product reviews in order to find reviews with inconsistent star ratings. A paragraph vector was utilized to transform textual product evaluations into numeric data that was then fed into a neural network with recurrent equipped with a gated recurrent unit for training. We built a model that takes into account the review text's semantic connections to the product data. Additionally, we built a web service that uses the trained model to predict the rating score of a submitted review and gives feedback to a reviewer if the anticipated and submitted ratings do not line up.
Sentiment Analysis of Hotel Reviews Based on Deep Learning
Page: 386-397 (12)
Author: Jagmeet Sohal*
DOI: 10.2174/9789815305395125020037
PDF Price: $30
Abstract
Many hotel evaluations have been written and shared online these days. Machine learning sentiment classification requires complicated artificial design features and feature extraction technique, whereas emotion dictionary-based sentiment classification requires a large amount of emotional database resources. In this study, we present the idea of long short-term memory. The text categorization method is used to determine the general tone. First, the brief comment text is processed into the LSTM network using word2vec and word segmentation technology; next, a dropout technique is implemented to avoid overfitting in order to get the final rating model. By using the LSTM network's superior short-term memory, a positive impact has been realized on sentiment categorization of reviews of hotels, with a precision of more than 95%.
Utilizing Machine Learning for Natural Language Processing to Conduct Sentiment Analysis on Twitter Data in Multiple Languages
Page: 398-408 (11)
Author: Rahul Mishra*
DOI: 10.2174/9789815305395125020038
PDF Price: $30
Abstract
Retailers, market analysts, and other users of the web are greatly influenced by user views. Arranging the unstructured data gathered from various social media networks correctly is necessary for doing relevant analysis. Emotional evaluation as a method for cross-lingual data classification has received considerable attention. Textual organization is a subfield of natural language processing, or NLP, that may be used to classify an individual's emotional or mental condition as positive, negative, beneficial, or detrimental, like a thumbs up or thumbs down, etc. A combination of sentiment analysis as well as deep learning techniques might be the key to solving this kind of problem. Deep learning models, which are capable of machine learning, are particularly useful for this. One of the most widely used deep learning architectures for analyzing sentiment in text is called Long Short Term Memory. These frameworks have potential applications in NLP. In this study, we provide algorithms to solve the problem of multilingual sentiment analysis, and we evaluate their precision factors to determine which one is the most effective.
The Use of Machine Learning to Analyze the Sentiment for Social Media Networks
Page: 409-419 (11)
Author: Darleen Grover*
DOI: 10.2174/9789815305395125020039
PDF Price: $30
Abstract
The amount of textual information on the internet has increased significantly with the debut of social media platforms like Twitter, including news stories and historical records. This is due to the growth of Web 2.0. More individuals are using the internet and different forms of social media to share their thoughts and feelings with the world. As a result, more phrases with emotional nuance were created by the general public. It is natural that researchers will look into new approaches to understanding people's emotions and reactions. In addition to providing a novel hybrid system that combines text mining and neural networks for sentiment categorization, this study evaluates the efficacy of many machine learning and deep learning techniques. More than a million tweets from across five different topics were utilized to create this dataset. Seventy-five percent of the dataset was used for training, while the remaining twenty-five percent was used for testing. When compared to traditional supervised learning methods, the system's hybrid approach displays a maximum accuracy of 83.7%.
Sentiment Classification of Textual Content using Hybrid DNN and SVM Models
Page: 420-432 (13)
Author: Abhishek Singla*
DOI: 10.2174/9789815305395125020040
PDF Price: $30
Abstract
The proliferation of Web 2.0 has resulted in a deluge of real-time, unstructured data such as user comments, opinions, and likes. The lack of structure in the data makes it difficult to create a reliable prediction model for sentiment analysis. There have been promising applications of several DNN architectures to sentiment analysis, however, these methods tend to treat all features identically and struggle with high-dimensional feature spaces. In addition, existing techniques fail to effectively combine semantic and sentiment knowledge for the purpose of extracting meaningful relevant contextual sentiment characteristics. This paper proposes an integrated convolutional neural network, or CNN, architecture that takes sentiment as well as context into consideration as a means of intelligently developing and highlighting significant components of relevant sentiment contextual in the text. To start, we use transformers' bidirectional encoder representations to create sentiment-enhanced embeddings of words for text semantic extraction using integrated emotion lexicons with broad coverage for feature identification. The proposed approach then adjusts the CNN in a way that it can detect both word order/contextual text semantics data as well as the long-dependency relationship in the phrase sequence. Our approach also employs a system to prioritize the most important portions of the phrase sequence. One last step in sentiment analysis is the use of support vector machines (SVMs) to reduce the complexity of the space of features and identify locally significant characteristics. The accuracy of existing text sentiment categorization is greatly improved by the use of the proposed model, as shown by an evaluation of real-world benchmark datasets.
Big Data Analysis and Information Quality: Challenges, Solutions, and Open Problems
Page: 433-447 (15)
Author: Sahil Suri*
DOI: 10.2174/9789815305395125020041
PDF Price: $30
Abstract
Big Data is here, thanks to the proliferation of social media and IoT devices. As a result of the immense benefit that Big Data has brought to the public as well as businesses, the question of how to handle and utilise it more effectively has captivated people from every walk of life. Big Data processing has been plagued by difficulties due to the 4V features of Big Data. Solving the data quality issue is crucial to Big Data processing, as is ensuring data quality, which is a precondition for Big Data to play its worth. Two examples of where Big Data has been put to good use are in recommendation and prediction systems. In this research, we investigate Big Data recommendation and prediction systems and endeavour to ascertain the quality of data at each stage of the process: collection, preprocessing, storage, and analysis. The proposed remedy follows from a thorough description and examination of the issue at hand. We have left a few questions unanswered at the conclusion of the report.
Using Deep Learning Techniques to Detect Traffic Information in Social Media Texts
Page: 448-459 (12)
Author: Sourav Rampal*
DOI: 10.2174/9789815305395125020042
PDF Price: $30
Abstract
Given the real-time and pervasive nature of social media, data mining for traffic-related insights is a newly emerging area of study. In this work, we discuss the challenge of mining social media for traffic-generating microblogs on Sina Weibo, the Chinese equivalent of Twitter. It is recast as a classification issue in short texts for machine learning. In the first step, we use a dataset of three billion microblogs and the continuous bag-of-word model to learn word embedding representations. Word embedding, as opposed to the standard one-hot vector representation of words, has been shown to be useful in natural language processing tasks because of its ability to capture semantic similarity between words. Then, we suggest feeding the learned word embeddings into convolutional neural networks (CNNs), long short-term memory (LSTM) models, and their combination LSTM-CNN to extract traffic-related microblogs. We evaluate the suggested techniques against state-of-the-art methods such as the support vector machine (SVM) model using a bag of n-gram features, the SVM model using word vector features, and the multi-layer perceptron model using word vector features. The proposed deep learning methods are shown to be useful in experiments.
Deep Sentiment Classification in COVID-19 Using LSTM Recurrent Neural Network
Page: 460-468 (9)
Author: Jatin Khurana*
DOI: 10.2174/9789815305395125020043
PDF Price: $30
Abstract
Users (people/patients) concerned about health concerns have an easy outlet in online medical forums along with other public social media on the Internet. The World Health Organization declared a global public health emergency in response to the emergence of a new coronavirus (infection which causes the disease termed COVID-19) in late December 2019. In this research, we employed a natural language processing (NLP) technique based on topic modeling to automatically extract COVID19-related talks from social media and discover numerous concerns linked to COVID19 from public viewpoints. As an added bonus, we look into the possibility of employing a long short-term memory (LSTM) recurrent neural network to accomplish the same task with COVID-19 remarks. Our research highlights the value of incorporating public opinion and appropriate computational approaches into the process of learning about and making decisions on COVID-19. The trials also showed that the study model was able to reach an accuracy of 81.15 percent, which is greater than the accuracy attained by many other popular machine-learning methods for COVID-19 sentiment classification.
Machine Learning-Based Data Preprocessing as well as Visualization Techniques for Predicting Students' Tasks
Page: 469-482 (14)
Author: Pratik Mahajan*
DOI: 10.2174/9789815305395125020044
PDF Price: $30
Abstract
A student's chances of landing a job after graduation are influenced by their performance in school and their history of academic accomplishment. Students who want to succeed in the working world need to develop skills including technical proficiency, critical thinking, and effective communication. Here we are making an attempt to figure out how students' academic performance affects their future opportunities. Algorithms for data mining play a key role in analysing and forecasting the chance for students' placement according to their previous academic achievement. We polled students at a prominent technical institution for this piece. Multiple factors that influence students' probabilities are included in the dataset, and these factors are examined and shown graphically. We made an effort to assess the data and provide visualisations and insights before running or applying algorithms for machine learning to this dataset. Data analysis, understanding, and preparation are the key focuses of this research.
The Prediction of Faults Using Large Amounts of Industrial Data
Page: 483-498 (16)
Author: Jagtej Singh*
DOI: 10.2174/9789815305395125020045
PDF Price: $30
Abstract
An essential feature of an intelligent workplace is error-free manufacturing. The data-driven approaches rely on fundamental status data with minimal volume, while the conventional model-based methods rely heavily on accurate equipment models. Not only are these aspects problematic, but they also prevent them from meeting the real-time need of evaluating industrial big data in an IoT setting. In this research, we provide a fault prediction approach that uses industrial big data to unearth the connection between data (such as status and sound data) and equipment failures using machine learning techniques. Not only that but the breakdown could be investigated promptly since the equipment's status could be tracked in real-time. Our solution outperforms the current ones in terms of accuracy and real-time capabilities, according to the simulation results.
Comparison Analysis of Logical Regression and Random Forest with Word Embedding Techniques for Twitter Sentiment Analysis
Page: 499-509 (11)
Author: Dhiraj Singh*
DOI: 10.2174/9789815305395125020046
PDF Price: $30
Abstract
In order to carry out categorization and generate new categories, the huge volumes of textual materials produced nowadays must be immediately organised. The fundamental method of gaining insights from organising textual data is text classification. Then, we further classify the classes based on the discovered text types. Separated into four stages—pre-treatment, text representation, classifier execution, and classification—we use a wide range of machine learning approaches to classify texts. In this study, we utilise real-world data from Twitter to evaluate and compare several sentiment analysis approaches. We clean the data and divide it into train and text set before developing models using various vectorising approaches and compare the outcomes. Based on a comparison of the models with various vectorizations, it was found that the best performance was provided by the Logical Regression (LR) models using TF-IDF, with an f1 value of 0.81 and good accuracy and recollection values.
The Classification of News Articles Through the Use of Deep Learning and the Doc2Vec Modeling
Page: 510-522 (13)
Author: Himanshu Makhija*
DOI: 10.2174/9789815305395125020047
PDF Price: $30
Abstract
The exponential growth in internet use has also led to the proliferation of textual information in large quantities. Since handling unstructured material manually is difficult, there is a need to explore novel techniques for automated categorization of textual information. The primary goal of text categorization is to teach a model to correctly categorise an unseen text. In this research, the Doc2vec word embedding technique was used to classify stories in Turkish from the TTC-3600 database of Turkish news and BBC news stories in English. In addition to the CNN based on deep learning, traditional machine learning classification methods including Gauss Naive Bayes (GNB), Random Forest (RF), Naive Bayes (NB), and Support Vector Machine (SVM) are used. The best classification results using CNN were achieved with the proposed model, scoring 94.17% on the Turkish database and 96.41% on the English database.
Investigating the Utility of Data Mining for Automated Credit Scoring
Page: 523-540 (18)
Author: Amarpal Yadav*
DOI: 10.2174/9789815305395125020048
PDF Price: $30
Abstract
Banks and other financial organisations rely heavily on credit scoring (CS) as a method of risk management since it is both effective and necessary. It reduces financial risks and gives sound advice on loan disbursement. As a result, businesses and financial institutions are exploring innovative automated solutions to the CS dilemma in an effort to safeguard their resources and those of their clients. The use of various machine learning (ML) as well as data mining (DM) approaches has led to significant progress in CS prediction in recent years. The Deep Genetic Hierarchical Network of Learners (DGHNL) is a novel approach developed for this study. Support Vector Machines (SVMs), k-nearest Neighbours (kNNs), Probabilistic Neural Networks (PNNs), and fuzzy structures are just some of the many types of methods that may be used in the suggested method. The Statlog German (1000 occurrences) approval of credit dataset from the UCI machine learning library was used to evaluate our model. We used a DGHNL model with five unique learner types, two feature extraction methods, three kernel functions, and three methods for optimising model parameters. In addition to conventional cross-validation (CV) and train-testing (stratified 10-fold) methods, this model employs a cutting-edge biological layered training (participant selection) approach. Using data on German credit approvals from Statlog, we show that the suggested DGHNL model can obtain a prediction accuracy of 94.60% (54 errors per 1000 classifications) with its 29-layer architecture.
Investigating the Use of Data Mining for Knowledge Discovery
Page: 541-553 (13)
Author: Sover Singh Bisht*
DOI: 10.2174/9789815305395125020049
PDF Price: $30
Abstract
The practice of “lifelogging” involves documenting an increasing amount of one's subjective everyday experience with the intention of using the recordings in the future as a memory aid or the foundation for data-driven self-development. Therefore, the usefulness of the generated lifelogs depends on the lifeloggers' ability to efficiently sift through them. The logs' intrinsic multi-modality and semi-structure allow them to combine data from a variety of sources, including cameras and other wearable physical and virtual sensors. As a result, expressing the data in a graph structure allows for the effective capturing of all created interrelations. Alternative methods must be developed to capture the higher-level semantics because it is impossible to manually or mechanically annotate each entry with a significant amount of semantic context. We describe an Improved Life Graph (ILG), a first method for building a Knowledge Graph-based lifelog representation and retrieval solution, which can capture a lifelog in a graph structure and augment it with external data to help with the connection of higher-level semantic information.
Exploring the Role of Big Data in Predictive Analytics
Page: 554-569 (16)
Author: T. R. Mahesh*
DOI: 10.2174/9789815305395125020050
PDF Price: $30
Abstract
Cardiovascular illness is afflicting enormous monetary and psychological costs. The development of an ASHRO-based model for forecasting healthcare resource use and its link with clinical outcomes was driven by a desire to improve the economy and provide a high-quality evaluation of the healthcare system. Data included in this analysis were taken from a big database that included doctor visits, insurance claims across several years, and results of preventive health screenings. Hospitalized patients with heart illness (ICD-10 I00-I99) comprised the study population. Broadly defined composites compliance served as the explanatory variable, while medical as well as long-term care costs served as the objective variable. Using a combination of random forest learning (AI) and multiple regression analysis, predictive models were calibrated. These models were then used to create ASHRO scores. Two measures, the area under the curve as well as the Hosmer-Lemeshow test, were used to assess the prediction model's effectiveness. After controlling for clinical risk variables, we compared the two ASHRO 50% threshold groups' total morbidity at 48 months of follow-up using matching propensity scores. Heart disease affected 61.9% of the 48,456 patients surveyed, with an average age of 68.3 9.9 years at hospital release. For the purpose of adherence score classification, machine learning was employed to combine eight factors into a single index: generic drug rate, interconnecting outpatient visits/clinical laboratory as well as physiological tests, the proportion of days addressed, secondary mitigation, rehabilitation magnitude, direction, and a single index that adjusted for eight factors. In the end, the multiple regression study yielded a 0.313 (p 0.001) coefficient of determination. Medical as well as long-term care expenditures had a statistically significant total coefficient of determination (p 0.001) in a logistic regression study using 50% along with 25%/75% cut-off values. At the 50% level of significance (2% vs. 7%; p 0.001), the relationship between ASHRO score and mortality rate was statistically significant.
Implementing Automated Reasoning in Natural Language Processing
Page: 570-583 (14)
Author: N. Sengottaiyan* and Rohaila Naaz
DOI: 10.2174/9789815305395125020051
PDF Price: $30
Abstract
One deep learning method is the Convolutional Neural Network (CNN). Natural language processing problems like text classification are simplified using this approach. In this study, we use a deep learning strategy, namely the CNN method to deal with the issue of text classification. CNNs, which require a large deal of time as well as finances to train and use, have been greatly impeded by the rise of Big Data and the increased complexity of tasks. To get around these problems, we introduce a MapReduce-based CNN that rethinks what a CNN has learned by breaking it down into a series of smaller networks and training them in parallel. Subsets of incoming text are analysed by many autonomous networks.
Subject Index
Page: 584-589 (6)
Author: Pankaj Kumar Mishra and Satya Prakash Yadav
DOI: 10.2174/9789815305395125020052
Introduction
Demystifying Emerging Trends in Machine Learning (Volume 2) offers a deep dive into emerging and trending topics in the field of machine learning (ML). This edited volume showcases several machine learning methods for a variety of tasks. A key focus of this volume is the application of text classification for cybersecurity, E-commerce, sentiment analysis, public health and web content analysis. The 49 chapters highlight a wide variety of machine learning methods including SVNs, K-Means Clustering, CNNs, DCNNs, among others. Each chapter includes accessible information through summaries, discussions and reference lists. This comprehensive volume is essential for students, researchers, and professionals eager to understand the emerging trends reshaping machine learning today.

