The Science and Information (SAI) Organization
  • Home
  • About Us
  • Journals
  • Conferences
  • Contact Us

Publication Links

  • IJACSA
  • Author Guidelines
  • Publication Policies
  • Digital Archiving Policy
  • Promote your Publication
  • Metadata Harvesting (OAI2)

IJACSA

  • About the Journal
  • Call for Papers
  • Editorial Board
  • Author Guidelines
  • Submit your Paper
  • Current Issue
  • Archives
  • Indexing
  • Fees/ APC
  • Reviewers
  • Apply as a Reviewer

IJARAI

  • About the Journal
  • Archives
  • Indexing & Archiving

Special Issues

  • Home
  • Archives
  • Proposals
  • Guest Editors
  • SUSAI-EE 2025
  • ICONS-BA 2025
  • IoT-BLOCK 2025

Future of Information and Communication Conference (FICC)

  • Home
  • Call for Papers
  • Submit your Paper/Poster
  • Register
  • Venue
  • Contact

Computing Conference

  • Home
  • Call for Papers
  • Submit your Paper/Poster
  • Register
  • Venue
  • Contact

Intelligent Systems Conference (IntelliSys)

  • Home
  • Call for Papers
  • Submit your Paper/Poster
  • Register
  • Venue
  • Contact

Future Technologies Conference (FTC)

  • Home
  • Call for Papers
  • Submit your Paper/Poster
  • Register
  • Venue
  • Contact
  • Home
  • Call for Papers
  • Editorial Board
  • Guidelines
  • Submit
  • Current Issue
  • Archives
  • Indexing
  • Fees
  • Reviewers
  • Subscribe

IJACSA Volume 15 Issue 5

Copyright Statement: This is an open access publication licensed under a Creative Commons Attribution 4.0 International License, which permits unrestricted use, distribution, and reproduction in any medium, even commercially as long as the original work is properly cited.

View Full Issue

Paper 1: Multiview Outlier Filtered Pediatric Heart Sound Classification

Abstract: The advancements in deep learning has generated a large-scale interest in development of black-box models for various use cases in different domains such as healthcare, in both at-home and critical setting for diagnosis and monitoring of various health conditions. The use of audio signals as a view for diagnosis is nascent and the success of deep learning models in ingesting multimedia data provides an opportunity for use as a diagnostic medium. For the widespread use of these decision support systems, it is prudent to develop high performing systems which require large quantities of data for training and low-cost method of data collection making it more accessible for developing regions of the world and general population. Data collected from low-cost collection especially wireless devices are prone to outliers and anomalies. The presence of outliers skews the hypothesis space of the model and leads to model drift on deployment. In this paper, we propose a multiview pipeline through interpretable outlier filtering on the small Mendeley Children Heart Sound dataset collected using wireless low-cost digital stethoscope. Our proposed pipeline explores and provides dimensionally reduced interpertable visualizations for functional understanding of the effect of various outlier filtering methods on deep learning model hypothesis space and fusion strategies for multiple views of heart sound data namely raw time-series signal and Mel Frequency Cepstrum Coefficients achieving 98.19% state-of-the-art testing accuracy.

Author 1: Sagnik Dakshit

Keywords: Deep learning; outlier filtering; machine learning; ECG

PDF

Paper 2: Trigger Screen Restriction Framework, iOS use Case Towards Building a Gamified Physical Intervention

Abstract: The growing trend of inactive lifestyles caused by excessive use of mobile devices raises severe concerns about people’s health and well-being. This paper illustrates the technical implementation of the Trigger Screen Restriction (TSR) framework, which integrates advanced technologies, including machine learning and gamification techniques, to address the limitations of traditional gamified physical interventions. The TSR framework encourages physical activity by leveraging the fear of missing out phenomenon, strategically restricting access to social media applications based on activity goals. The framework’s components, including the Screen Time Restriction, Notification Triggers, Computer Vision Model, and Reward Engine, work together to create an engaging and personalized experience that motivates users to engage in regular physical activity. Although the TSR framework represents a potentially significant step forward in gamified physical activity interventions, it remains a theoretical model requiring further investigation and rigorous testing.

Author 1: Majed Hariri
Author 2: Richard Stone

Keywords: Gamification; physical activity; screen-time restriction; triggered screen restriction framework; TSR Framework; personalized interventions; gamified physical intervention

PDF

Paper 3: An Intelligent Method for Collecting and Analyzing Voice Reviews to Gauge Customer Satisfaction

Abstract: Customer loyalty and customer satisfaction are premier goals of modern business since these factors indicate customers’ future behaviour and ultimate impact on the revenue and value of a business. The customers’ reviews, ratings, and rankings are a primary source for gauging customer satisfaction levels. Similar efforts have been reported in the literature. However, there has been no solution that can record real-time views of customers and provide analysis of the views. In this paper, a novel approach is presented that records, stores, and analyzes the customer live reviews and uses text mining to perform various levels of analysis of the reviews. The used approach also involves steps like void-to-text conversion, pre-processing, sentiment analysis, and sentiment report generation. This paper also presents a prototype tool that is the outcome of the present research. This research not only provides novel functionalities in the domain but also outperforms similar solutions in performance.

Author 1: Nail Khabibullin

Keywords: Voice reviews; customer satisfaction; text mining; sentiment analysis

PDF

Paper 4: Intelligent Framework in a Serverless Computing for Serving using Artificial Intelligence and Machine Learning

Abstract: Serverless computing has grown in popularity as a paradigm for deploying applications in the cloud due to its ability to scale, cost-effectiveness, and simplified infrastructure management. Serverless architectures can benefit AI and Machine Learning (ML) models, which are becoming increasingly complex and resource-intensive. This study investigates the integration of AI/ML frameworks and models into serverless computing environments. It explains the steps involved, including model training, deployment, packaging, function implementation, and inference. Serverless platforms' auto-scaling capabilities allow for seamless handling of varying workloads, while built-in monitoring and logging features ensure effective management. Continuous integration and deployment pipelines simplify the deployment process. Using serverless computing for AI/ML models offers developers scalability, flexibility, and cost savings, allowing them to focus on model development rather than infrastructure issues. The proposed model leverages performance forecasting and serverless computing model deployment using virtual machines, specifically utilizing the Knative platform. Experimental validation demonstrates that the model effectively predicts performance based on specific parameters with minimal data collection. The results indicate significant improvements in scalability and cost efficiency while maintaining optimal performance. This performance model can guide application owners in selecting the best configurations for varying workloads and assist serverless providers in setting adaptive defaults for target value configurations.

Author 1: Deepak Khatri
Author 2: Sunil Kumar Khatri
Author 3: Deepti Mishra

Keywords: Machine learning; data analytics; serverless computing; performance testing

PDF

Paper 5: Find a Research Collaborator: An Ontology-Based Solution to Find the Right Resources for Research Collaboration

Abstract: Researchers in Higher Education (HE) institution-s/academia and in industry are continuously engaged in gen-erating new solutions and products for existing and emergent problems. Doing quality research and producing better scientific results depend greatly on solid research teams and scientific collaborators. Research output in HE institutions and industry can be optimized with appropriate resources in research teams and collaborations with suitable research partners. The main challenge in finding suitable resources for joint research projects and scientific collaborations pertains to the availability of data and metadata of researchers and their scientific work in tradi-tional formats, for instance, websites, portals, documents, and traditional databases. However, these traditional data sources do not support intelligent and smart ways of finding and querying the right resources for joint research and scientific collaboration. A possible solution resides in the deployment of Semantic Web (SW) techniques and technologies for representing researcher and their research contribution data in a machine-understandable format, thus ultimately proving useful for smart and intelligent query-answering purposes. In pursuit of this, we present a general Methodology for Ontology Design and Development (MODD). We also describe the use of this methodology to design and develop Higher Education Ontology (HEO). This HEO can be used to automate various activities and processes in HE. In addition, we describe the use and adoption of the HEO through a case study on the topic of “finding the right resources for joint research and scientific collaboration”. Finally, we provide an analysis and evaluation of our methodology for posing smart queries and evaluating the results based on machine reasoning.

Author 1: Nada Abdullah Alrehaili
Author 2: Muhammad Ahtisham Aslam
Author 3: Amani Falah Alharbi
Author 4: Rehab Bahaaddin Ashari

Keywords: Higher Education Ontology (HEO); Linked Open Data (LOD); Machine Reasoning; Semantic Web (SW); SPARQL Queries

PDF

Paper 6: Advancing Hospital Cybersecurity Through IoT-Enabled Neural Network for Human Behavior Analysis and Anomaly Detection

Abstract: The integration of Internet of Things (IoT) technologies in hospital environments has introduced transformative changes in patient care and operational efficiency. However, this increased connectivity also presents significant cybersecurity challenges, particularly concerning the protection of patient data and healthcare operations. This research explores the application of advanced machine learning models, specifically LSTM-CNN hybrid architectures, for anomaly detection and behavior analysis in hospital IoT ecosystems. Employing a mixed-methods approach, the study utilizes LSTM -CNN models, coupled with the Mobile Health Human Behavior Analysis dataset, to analyze human behavior in a cybersecurity context in the hospital. The model architecture, tailored for the dynamic nature of hospital IoT activities, features a layered. The training accuracy attains an impressive 99.53%, underscoring the model's proficiency in learning from the training data. On the testing set, the model exhibits robust generalization with an accuracy of 91.42%. This paper represents a significant advancement in the convergence of AI and healthcare cybersecurity. The model's efficacy and promising outcomes underscore its potential deployment in real-world hospital scenarios.

Author 1: Faisal ALmojel
Author 2: Shailendra Mishra

Keywords: IoT security; cyber security; network security; machine learning; LSTM

PDF

Paper 7: Tile Defect Recognition Network Based on Amplified Attention Mechanism and Feature Fusion

Abstract: For the current situation of low AP of tile defect detection with incomplete detection of defect types, this paper proposes YOLO-SA, a detection neural network based on the enhanced attention mechanism and feature fusion. We propose an enhanced attention mechanism named amplified attention mechanism to reduce the information attenuation of the defect information in the neural network and improve the AP of the neural network. Then, we use the EIoU loss function, the four-layer feature fusion, and let the backbone network directly involved in the detection and other methods to construct an excellent tile defect detection and recognition model Yolo-SA. In the experiments, this neural network achieves better experimental results with an improvement of 8.15 percentage points over Yolov5s and 8.93 percentage points over Yolov8n. The model proposed in this paper has high application value in the direction of tile defect recognition.

Author 1: JiaMing Zhang
Author 2: ZanXia Qiang
Author 3: YuGang Li

Keywords: Amplified attention mechanism; defect recognition; small target recognition; Yolo; feature fusion

PDF

Paper 8: Tendon-Driven Robotic Arm Control Method Based on Radial Basis Function Adaptive Tracking Algorithm

Abstract: With the rapid development of intelligent technology, robotic arms are widely used in different fields. The study combines the tendon drive theory and radial basis function neural network to construct the robotic arm model, and then combines the back-stepping method and non-singular fast terminal sliding mode to improve the controller and system optimization of the tendon drive robotic arm model. Simulation tests on commercial mathematical software platforms yielded that joint 2 achieves stable overlap of position trajectory and velocity trajectory after 0.2s and 0.5s with errors of 1° and 1°/s, respectively. Radial basis function neural network approximation of robotic arm error converged to the true value at 14s. The optimized joint achieved the accuracy of trajectory tracking after 0.2s. Also the control torque of joint 2 changes at 1.5s, 4.5s and 8s and its change is small. The tendon tension curve was smoother and more stable in the range of -0.05N~0.0.5N to show that the robotic arm model has superiority after the optimization of the controller, and the interference observer had accurate estimation of the tracking trajectory of the tendon-driven robotic arm. Therefore, the radial basis function-based adaptive tracking algorithm had higher accuracy for the tendon-driven robotic arm model and provided technical reference for the control system of the intelligent robotic arm.

Author 1: Xiaoke Fang

Keywords: Tendon drive; adaptive neural network; dynamic relationship; sliding membrane control; trajectory tracking

PDF

Paper 9: Construction of Cloud Computing Task Scheduling Model Based on Simulated Annealing Hybrid Algorithm

Abstract: With the development of cloud computing technology, effective task scheduling can help people improve work efficiency. Therefore, this study presented a hybrid algorithm on the grounds of simulated annealing and taboo search to optimize task scheduling in cloud computing. This study presented a hybrid algorithm for optimizing the cloud computing task scheduling model. The model used simulated annealing algorithm and taboo search algorithm to convert the objective function into an energy function, allowing atoms to quickly arrange in terms of a certain rule for obtaining the optimal solution. The study analyzed the model through simulation experiments, and the experiment showed that the optimal value of the hybrid algorithm in high-dimensional unimodal testing was 7.15E-247, far superior to the whale optimization algorithm's 3.99E-28 and the grey wolf optimization algorithm's 1.10E-28. The completion time of the hybrid algorithm decreased with the growth of virtual machines, and the shortest time was 8.6 seconds. However, the load balancing degree of the hybrid algorithm increased with the growth of virtual machines. The final results indicated that the proposed hybrid algorithm exhibits high efficiency and superior performance in cloud computing task scheduling, especially when dealing with large-scale and complex optimization problems.

Author 1: Kejin Lv
Author 2: Tianxu Huang

Keywords: Simulated annealing algorithm; taboo search optimization algorithm; cloud computing; task scheduling; completion time; load balancing degree

PDF

Paper 10: Ensemble Empirical Mode Decomposition Based on Sparse Bayesian Learning with Mixed Kernel for Landslide Displacement Prediction

Abstract: Inspired by the principles of decomposition and ensemble, we introduce an Ensemble Empirical Mode Decomposition (EEMD) method that incorporates Sparse Bayesian Learning (SBL) with Mixed Kernel, referred to as EEMD-SBLMK, specifically tailored for landslide displacement prediction. EEMD and Mutual Information (MI) techniques were jointly employed to identify potential input variables for our forecast model. Additionally, each selected component was trained using distinct kernel functions. By minimizing the number of Relevance Vector Machine (RVM) rules computed, we achieved an optimal balance between kernel functions and selected parameters. The EEMD-SBLMK approach generated final results by summing the prediction values of each subsequence along with the residual function associated with the corresponding kernel function. To validate the performance of our EEMD-SBLMK model, we conducted a real-world case study on the Liangshuijing (LSJ) landslide in China. Furthermore, in comparison to RVM-Cubic and RVM-Bubble, EEMD-SBLMK emerged as the most effective method, delivering superior results in the same measurement metrics.

Author 1: Ping Jiang
Author 2: Jiejie Chen

Keywords: Bubble; cublic; ensemble empirical mode decomposition; landslide; Sparse Bayesian Learning

PDF

Paper 11: Adaptive Scheduling of Robots in the Mixed Flow Workshop of Industrial Internet of Things

Abstract: With the deep integration of industrial Internet of Things technology and artificial intelligence technology, the material robot has been widely used in the Internet of Things workshop. In view of many complex factors such as real-time dynamic change and uncertain condition in workshop, this paper proposes to realize workshop adaptive scheduling decision with component layer construction and SPMCTS search method with real-time state as the root node. This method transforms the robot scheduling problem into a Markov decision process and describes a detailed representation of workshop states, actions, rewards, and strategies. In the real-time scheduling process, the search method is based on the artifact component layer construction, and only considers the state relationship between two adjacent groups, so as to simplify the calculation difficulty. In the subtree search, SPMCTS is applied to search the real-time state as the root node, and the extension method and shear method are applied to conduct strategy exploration and information accumulation, so that the deeper the real-time state node in the subtree, the more the optimal strategy can be obtained quickly and accurately. Finally, the effectiveness and superiority of the proposed method are verified by real case simulation analysis.

Author 1: Dejun Miao
Author 2: Rongyan Xu
Author 3: Yizong Dai
Author 4: Jiusong Chen

Keywords: Industrial Internet of Things; mixed flow workshop; robot; Markov decision-making process; SPMCTS

PDF

Paper 12: Optimization of Student Behavior Detection Algorithm Based on Improved SSD Algorithm

Abstract: Despite advancements in educational technology, traditional action recognition algorithms have struggled to effectively monitor student behavior in dynamic classroom settings. To address this gap, the Single Shot Detector (SSD) algorithm was optimized for educational environments. This study aimed to assess whether integrating the Mobilenet architecture with the SSD algorithm could improve the accuracy and speed of detecting student behavior in classrooms, and how these enhancements would impact the practical implementation of behavior-monitoring technologies in education. An improved SSD algorithm was developed using Mobilenet, known for its efficient data processing capabilities. A dataset of 2,500 images depicting various student behaviors was collected and enhanced through preprocessing methods to train the model. The optimized SSD model outperformed traditional algorithms in accuracy and speed, thanks to the integration of Mobilenet. Evaluation metrics such as precision, recall, and frames per second (fps) confirmed the superior performance of the Mobilenet-enhanced SSD algorithm in real-time environmental analysis. This advancement represents a significant improvement in surveillance technologies for educational settings, enabling more precise and timely assessments of student behavior. Despite the promising outcomes, the study faced limitations due to the uniformity of the dataset, which mainly consisted of controlled environment images. To improve the generalizability of the findings, it is suggested that future research should broaden the dataset to encompass a wider range of educational settings and student demographics. Additionally, it is encouraged to explore alternative advanced machine learning frameworks and conduct longitudinal studies to evaluate the influence of real-time behavior monitoring on educational outcomes.

Author 1: Yongqing CAO
Author 2: Dan LIU

Keywords: Improved single shot detector (SSD) model; mobilenet network; class behavior recognition; artificial intelligence

PDF

Paper 13: Detecting User Credibility on Twitter using a Hybrid Machine Learning Model of Features’ Selection and Weighting

Abstract: With the pervasive and rapidly growing presence of the internet and social media, creating untrustworthy accounts has become effortless, allowing fake news to be spread for personal or private interests. As a result, it is crucial in this era to investigate the credibility of users on social networking platforms such as Twitter. In this research, we aim to integrate existing solutions from previous research to create a hybrid model. Our approach is based on selecting and weighting features using supervised machine learning methods such as ExtraTressClarifier, correlation-based algorithm methods, and SelectKBest to extract new ranked and weighted features in the dataset and then use them to train our model to discover their impact on the accuracy of user credibility detection issues. The research objective is to combine feature selection and weighting methods with Supervised Machine Learning to evaluate their impact on the accuracy of user credibility detection on Twitter. In addition, we measure the effectiveness of different feature categories on this detection. Experiments are conducted on one of the online available datasets. We seek to employ extracted features from a user's profile and statistical and emotional information. Then, the experimental results are compared to discover the effectiveness of the proposed solution. This study focuses on revealing the credibility of Twitter (or X-platform as recently renamed) accounts, which may result in the need for some adjustments to the generalization of Twitter’s outputs to other social media accounts such as LinkedIn, Facebook, and others.

Author 1: Nahid R. Abid-Althaqafi
Author 2: Hessah A. Alsalamah

Keywords: User credibility; supervised machine learning; feature selection; feature weighting; social network; twitter

PDF

Paper 14: Automated Motor Imagery Detection Through EEG Analysis and Deep Learning Models for Brain-Computer Interface Applications

Abstract: The classification of motor imagery holds significant importance within brain-computer interface (BCI) research as it allows for the identification of a person's intention, such as controlling a prosthesis. Motor imagery involves the brain's dynamic activities, commonly captured using electroencephalography (EEG) to record nonstationary time series with low signal-to-noise ratios. While various methods exist for extracting features from EEG signals, the application of deep learning techniques to enhance the representation of EEG features for improved motor imagery classification performance has been relatively unexplored. This research introduces a new deep learning approach based on two-dimensional CNNs with different architectures. Specifically, time-frequency domain representations of EEGs obtained by the wavelet transform method with different mother wavelets (Mexicanhat, Cmor, and Cgaus). The BCI competition IV-2a dataset held in 2008 was utilized for testing the proposed deep learning approaches. Several experiments were conducted and the results showed that the proposed method achieved better performance than some state-of-the-art methods. The findings of this study showed that the architecture of CNN and specifically the number of convolution layers in this deep learning network has a significant effect on the classification performance of motor imagery brain data. In addition, the mother wavelet in the wavelet transform is very important in the classification performance of motor imagery EEG data.

Author 1: Yang Li
Author 2: Bocheng Liu
Author 3: Yujia Tian

Keywords: Brain-computer interface (BCI); Electroencephalogram (EEG); motor imagery; deep learning; classification

PDF

Paper 15: Exploring Differential Entropy and Multifractal Cumulants for EEG-based Mental Workload Recognition

Abstract: In the current research, two nonlinear features were utilized for the design of EEG-based mental workload recognition: one feature based on differential entropy and the other feature based on multifractal cumulants. Clean EEGs recorded from 36 healthy volunteers in both resting and task states were subjected to feature extraction via differential entropy and multifractal cumulants. Then, these nonlinear features were utilized as input for a fuzzy KNN classifier. Experimental results showed that the multifractal cumulants feature vector achieved an AUC of 0.951, which is larger than the differential entropy feature vector (AUC = 0.935). However, the combination of both feature sets resulted in added value in identifying these two mental workloads (AUC = 0.993). Furthermore, the multifractal cumulants feature vector (best classification accuracy = 94.76%) obtained better classification results than the differential entropy feature vector (best classification accuracy = 92.61%). However, the combination of these two feature vectors achieved the best classification results: accuracy of 96.52%, sensitivity of 97.68%, specificity of 95.58%, and F1-score of 96.61%. This shows that these two feature vectors are complementary in identifying different mental workloads.

Author 1: Yan Lu

Keywords: Mental workload; EEG; nonlinear analysis; multifractal; differential entropy; fuzzy KNN; classification

PDF

Paper 16: Predicting Math Performance in High School Students using Machine Learning Techniques

Abstract: In the field of education, understanding and predicting student performance plays a crucial role in improving the quality of system management decisions. In this study, the power of various machine learning techniques to learn the complicated task of predicting students’ performance in math courses using demographic data of 395 students was investigated. Predicting students' performance through demographic information makes it possible to predict their performance before the start of the course. Filtered and wrapper feature selection methods were used to find 10 important features in predicting students' final math grades. Then, all the features of the data set as well as the 10 selected features of each of the feature selection methods were used as input for the regression analysis with the Adaboost model. Finally, the prediction performance of each of these feature sets in predicting students' math grades was evaluated using criteria such as Pearson's correlation coefficient and mean squared error. The best result was obtained from feature selection by the LASSO method. After the LASSO method for feature selection, the Extra Tree and Gradient Boosting Machine methods respectively had the best prediction of the final math grade. The present study showed that the LASSO feature selection technique integrated with regression analysis with the Adaboost model is a suitable data mining framework for predicting students' mathematical performance.

Author 1: Yuan hui

Keywords: Student performance; math grade prediction; feature selection; regression analysis; machine learning; data mining

PDF

Paper 17: IoT Device Identity Authentication Method Based on rPPG and CNN Facial Recognition

Abstract: This study aims to address the insufficient model recognition accuracy and limitations of authentication techniques in current IoT authentication methods. The research presents a more accurate face video image authentication technique by using a new authentication method that combines convolutional neural networks (CNN) and remote Photoplethysmography (rPPG) volumetric tracing. This method comprehensively analyzes facial video images to achieve effective authentication of user identity. The results showed that the new method had higher recognition accuracy when the light was weak. The new method performed better in ablation experiments. The error rate was 1.12% lower than the separate CNN model and 1.73% lower than the rPPG model. The half-error rate was lower than the traditional face authentication recognition model, and the method had better performance effect. Meanwhile, the images with high similarity showed better recognition stability. It can be seen that the new method is able to solve problems such as the recognition accuracy in identity authentication, but the recognition effect under extreme conditions requires further research. The research provides a new technical solution for the authentication of Internet of Things devices, which helps to improve the security and accuracy of the authentication system. By combining the CNN model and rPPG, the research not only improves the recognition accuracy in complex environments, but also enhances the system's adaptability to environmental changes. The new method provides a new solution for the advancement of Internet of Things authentication technology.

Author 1: Liwan Wu
Author 2: Chong Yang

Keywords: Internet of Things; identity authentication; facial recognition; remote photoplethysmography; error rate

PDF

Paper 18: Logistics Path Planning Method using NSGA-II Algorithm and BP Neural Network in the Era of Logistics 4.0

Abstract: The distribution of fresh food is affected by its perishable characteristics, and compared with ordinary logistics distribution, the distribution path needs to be very reasonably planned. However, the complexity of the actual road network and the time variation of traffic conditions are not considered in the existing food logistics planning methods. In order to solve this problem, this study takes road section travel prediction as the starting point, and uses the non-dominant ranking genetic algorithm II and the backpropagation network to construct a new logistics path planning model. Firstly, the road condition information detected by the retainer detection and the floating vehicle technology is integrated, and the travel time prediction is input into the backpropagation network model. In order to make the prediction model perform better, it is improved using a whale optimization algorithm. Then, according to the prediction results, the non-dominant ranking genetic algorithm II is used for distribution route planning. Through experimental analysis, the average distribution cost of method designed by this study was 9476 yuan, and the average carbon emission was 2871kg. Compared with the other three algorithms, the distribution cost was more than 15% lower, and the carbon emission was more than 12.5% lower. The planning method designed by the institute can achieve more reasonable, low-cost, and environmentally friendly logistics and distribution, and bring more satisfactory services to the lives of urban residents.

Author 1: Liuqing Li

Keywords: Whale optimization algorithm; non-dominant ordering genetic algorithm; backpropagation network; logistics and distribution; path planning

PDF

Paper 19: Cloud-Enabled Real-Time Monitoring and Alert System for Primary Network Resource Scheduling and Large-Scale Users

Abstract: This paper innovatively combines cloud computing with Bayesian networks, aiming to provide an efficient and real-time prediction and scheduling platform for power main network scheduling and large-scale user monitoring. The core of the research lies in the development of a set of novel intelligent scheduling algorithms, which integrates multi-objective optimization theory and deep reinforcement learning technology to achieve dynamic and optimal allocation of power grid resources in the cloud environment. By constructing a comprehensive evaluation system, this study verifies the advancement of the proposed model in multiple dimensions: not only does it make breakthroughs in the in-depth parsing and accurate prediction of electric power data, but it also significantly improves the prediction accuracy of the main grid load changes, tariff dynamic adjustments, grid security posture, and power consumption patterns of large users. The empirical study shows that compared with the existing methods, the model proposed in this study effectively reduces energy consumption and operation costs while improving prediction accuracy and dispatching efficiency, demonstrating its significant innovative value and practical significance in the field of intelligent grid management. The innovation of this paper lies in the development of a composite prediction model that integrates the powerful classification and prediction capabilities of Bayesian networks and the efficient learning mechanism of deep reinforcement learning in complex decision-making scenarios.

Author 1: Bin Zhang
Author 2: Hongchun Shu
Author 3: Dajun Si
Author 4: Jinding He
Author 5: Wenlin Yan

Keywords: Cloud computing; main network scheduling; large users; real-time monitoring; monitoring and prediction; systems research

PDF

Paper 20: A Comparative Work to Highlight the Superiority of Mouth Brooding Fish (MBF) over the Various ML Techniques in Password Security Classification

Abstract: Within the domain of password security classification, the pursuit of practical and dependable methodologies has prompted the examination of both biological and technological paradigms. The present study investigates the efficacy of Mouth Brooding Fish (MBF) as an innovative method in contrast to conventional Machine Learning (ML) approaches for classifying password security. The research approach entails a rigorous examination of the comparative analysis of MBF and ML algorithms, evaluating their effectiveness in password classification using many criteria, including accuracy, robustness, flexibility, and durability against adversarial assaults. The findings suggest that ML approaches have shown significant effectiveness in classifying passwords. However, using methodologies inspired by the minimum Bayes risk framework demonstrates a higher degree of resistance against typical cyber dangers. The intrinsic biological mechanisms of MBF, encompassing adaptive behaviors and inherent protection, play a role in enhancing the resilience and adaptability of the password security categorization system. The results offer significant insights that can inform the evolution of password security systems, integrating biological principles with technical progress to enhance safeguarding measures in digital environments. To emphasize the advantages of the suggested approach, several ML approaches are investigated, such as Support Vector Machines (SVM), AdaBoost, Multilayer Perceptron (MLP), Gaussian Kernel (GK), and Random Forest (RF). The F-score, accuracy, sensitivity, and specificity metrics for MBF exhibit noteworthy performance compared to the other selected models, with values of 100%.

Author 1: Yan Shi
Author 2: Yue Wang

Keywords: Mouth Brooding Fish (MBF); password security; Sber dataset; SVM; Random Forest; AdaBoost

PDF

Paper 21: Comparative Study: Mouth Brooding Fish (MBF) as a Novel Approach for Android Malware Detection

Abstract: Android Malware Detection has become increasingly prevalent, with the highest market share among all other mobile operating systems due to its open-source nature and user-friendliness. This has resulted in an uncontrolled proliferation of malicious applications targeting the Android platform. Emerging trends of Android malware are employing highly sophisticated detection and analysis evasion techniques, rendering traditional signature-based detection methods less effective in identifying modern and unknown malware. Alternative approaches, such as Machine Learning methods, have emerged as leading solutions for timely zero-day anomaly detection. Ensemble learning, a common meta-approach in machine learning, seeks to improve predictive performance by amalgamating predictions from multiple models. This paper introduces an enhanced strategy, Mouth Brooding Fish (MBF), based on ensemble learning for Android Malware Detection (AMD). The findings are further compared with the outputs of various algorithms including Support Vector Machine (SVM), AdaBoost, Multilayer Perceptron (MLP), Gaussian Kernel (GK), and Random Forest (RF). Compared to the other selected models, MBF exhibits remarkable performance with an F-score of 98.57%, precision of 99.65%, sensitivity of 97.51%, and specificity of 97.51%. Thus, the significant novelty of this work lies in the accuracy and authenticity of the selected algorithms, demonstrating their superior performance overall.

Author 1: Kangle Zhou
Author 2: Panpan Wang
Author 3: Baiqing He

Keywords: Android malware detection; ensemble learning; SVM; MLP; RF

PDF

Paper 22: Examining the Various Neural Network Algorithms Considering the Superiority of Mouth Brooding Fish in Data Classification

Abstract: Data classification, a crucial practice in information management, involves categorizing data based on its sensitivity to determine appropriate access levels and protection measures. This paper explores the utilization of novel algorithms, including mouth-brooding fish (MBF), alongside machine learning techniques, for the analysis of medical health data. The SVM exhibits suboptimal performance in the task of data categorization. Therefore, Adaboost may be considered a viable substitute for MBF due to its superior performance in terms of F-score, accuracy, specificity, and sensitivity. The accuracy of MBF, which stands at about 95%, surpasses that of Adaboost by a significant margin of 77%. The F-score, accuracy, and specificity values obtained for MBF are exceptional when compared to the other chosen models, with values of 97.17%, 93.6%, and 96.5%, respectively. The proposed algorithm exhibits promising advancements in health data categorization, offering a potential breakthrough in data classification methodologies. Leveraging this innovative approach could facilitate more accurate and efficient management of sensitive medical data, thereby enhancing healthcare systems' capabilities for data protection and analysis. The main novelty of this study lies in the introduction and evaluation of the MBF algorithm for data classification within the medical domain. Unlike traditional algorithms, MBF draws inspiration from the collective behavior of mouth-brooding fish, offering a unique optimization strategy that enhances both exploration and exploitation of the solution space. This novel approach presents a promising avenue for advancing healthcare analytics and decision-making processes.

Author 1: Lang Liu
Author 2: Yong Zhu

Keywords: Medical data analysis; clinical decision support; dataset classification; Mouth Brooding Fish; Support Vector Machine (SVM)

PDF

Paper 23: A Method for Assessing Financial Market Price Behavior: An Analysis of the Shanghai Stock Exchange Index

Abstract: A stock market is a venue where the shares of publicly traded companies are available for purchase and sale by individuals. The financial markets exert a substantial influence on various domains, including technology, employment, and business. Given the substantial rewards and risks associated with stock trading, investors are exceedingly concerned with the precision of future stock value forecasts. They modify their investment strategies in an effort to achieve even greater returns. Accurate stock price forecasting can be challenging in the securities industry due to the complex nature of the problem and the requirement for a comprehensive understanding of various interconnected factors. The stock market is influenced by a variety of factors, including politics, society, and economics. A multitude of interrelated factors contribute to these behaviors, and stock price fluctuations are capricious. In order to tackle a range of these difficulties, the present investigation proposes an innovative framework that integrates a Grasshopper optimization method with the gated recurrent unit model, a machine-learning approach. The research used data from the Shang Hai Stock Exchange Index for the period of 2015–2023. The proposed hybrid model was also tested on the 2013–2022 S&P 500 and Nikkei 225. The proposed model demonstrated optimal performance, exhibiting a minimal error rate and exceptional effectiveness. The study's findings demonstrate that the proposed model is more suitable for the volatile stock market and surpasses other existing strategies in terms of predictive accuracy.

Author 1: Zhi Huang
Author 2: Jiansheng Li

Keywords: Financial market; shanghai stock exchange price; gated recurrent unit; grasshopper optimization algorithm

PDF

Paper 24: Stock Market Volatility Estimation: A Case Study of the Hang Seng Index

Abstract: Among the influential elements in the national economy is the stock market. The stock market is a multifaceted system that combines economics, investor psychology, and other market mechanics. The objective of the financial market investment is to maximize profits; but, due to the market's complexity and the multitude of factors that might impact it, it is challenging to predict its future behavior. The challenging process of stock price prediction requires the analysis of a wide range of social, political, and economic factors. These variables include market trends, financial statements, earnings reports, and other data. The goal of this project is to develop an accurate hybrid stock price forecasting model using Random Forest which is combined with the optimization. Random Forest is one type of machine learning that is often used in time series analysis. This study provides stock price forecasting using the Hang Seng index market, which consists of the largest and most liquid corporations that are publicly traded on the Hong Kong Stock Exchange, data from 2015 to 2023. The Dow Jones and KOSPI were evaluated as two additional indices. This study demonstrates some optimization approaches including genetic algorithm, grey wolf optimization, and biogeography-based optimization, which drew inspiration from the phenomenon of species migrating between islands in search of a suitable habitat. Biogeography-based optimization has shown the best result among these optimizations. The proposed hybrid model obtained the values 0.992, 0.997, and 0.9937 for the coefficient of determination for HSI, Dow Jones, and KOSPI markets, respectively. These results indicate the ability of the model in order to predict the stock market with a high degree of accuracy.

Author 1: Shengwen Wu
Author 2: Qiqi Lin
Author 3: Xuefeng liu

Keywords: Hang Seng index; financial market; stock price prediction; Random Forest; biological bases optimization

PDF

Paper 25: Presenting a New Approach for Clustering Optimization in Wireless Sensor Networks using Fuzzy Cuckoo Search Algorithm

Abstract: Because of the developments in this technology, wireless sensor networks are now among the most commonly used in the domains of agriculture, harsh environments, medical, and the military. Among the many problems with these networks is their limited lifespan. Much work has been done in the fields of sensor communication, routing, and data gathering to reduce energy usage and increase network life. Routing protocols and clustering algorithms are two techniques for reducing energy use. Selecting the cluster head is the most important stage in any clustering technique. The objectives of this article are to decrease total energy consumption, increase packet delivery rates, and lengthen the network's lifetime. In order to do this, the LEACH protocol uses cuckoo search instead of probability distribution during the cluster head selection step and fuzzy logic during the routing phase. A MATLAB environment was utilized to evaluate the proposed method with the LEACH algorithm under identical conditions. The results of the comparison show that the recommended approach does a better job of prolonging the network's lifetime than the LEACH protocol.

Author 1: Bing ZHOU
Author 2: Youyou LI

Keywords: Wireless sensor network; fuzzy cuckoo search algorithm; clustering; fuzzy model

PDF

Paper 26: Enhancing Fraud Detection in Credit Card Transactions using Optimized Federated Learning Model

Abstract: In recent years, credit card transaction fraud has inflicted significant losses on both consumers and financial institutions. To address this critical issue, we propose an optimized framework for fraud detection. This study deals with non-identically independent distributions (IIDs) involving different numbers of clients. The proposed framework empowers banks to construct robust fraud detection models using their internal training data. Specifically, by optimizing the initial global model before to the federated learning phase, the suggested optimization technique accelerates convergence speed by reducing communication costs when moving forward with federal training. The optimization techniques using the three most recent metaheuristic Optimizers, namely: An improved gorilla troops optimizer (AGTO), Coati Optimization Algorithm (CoatiOA), Coati Optimization Algorithm (COA). Furthermore, credit card data is highly skewed, which makes it challenging to predict fraudulent transactions. The resampling strategy is used as a preprocessing step to improve the outcomes of unbalanced or skewed data. The performance of these algorithms is documented and compared. Computation time, accuracy, precision, recall, F-measure, loss, and computation time are used to assess the algorithms' performance. The experimental results show that AGTO and (CoatiOA) exhibit higher accuracy, precision, recall, and F1 scores compared to the baseline FL Model. Additionally, they achieve lower loss values.

Author 1: Mustafa Abdul Salam
Author 2: Doaa L. El-Bably
Author 3: Khaled M. Fouad
Author 4: M. Salah Eldin Elsayed

Keywords: Credit card fraud detection (CCFD); federated learning; optimization algorithms; identically independent distributions (IIDs); metaheuristic optimization techniques

PDF

Paper 27: Embedding Emotions in the Metaverse: The Emotive Keywords for Augmented Reality Mobile Library Application

Abstract: The emergence of the metaverse, marked by the seamless integration of augmented reality (AR) applications across various sectors is driving a profound transformation in the digital landscape. As we delve into the digital realm of the metaverse, just like other applications, it unfolds as an equally captivating canvas for emotional exploration, where a comprehensive understanding of human emotion for better user experience (UX) is vital. Although the efforts to investigate emotions within the metaverse are in progress, however there is a notable absence of extensive research that examines the user’s emotional experiences which incorporates a tailored set of keywords specifically for designing user interface (UI) products within this context, resulting in a substantial void in this particular domain. Therefore, the objective of this research is to synthesise and validate an extensive array of emotive keywords explicitly tailored for AR-based Mobile Library Application (MLA) design. This endeavor involves an exhaustive review of literature and a rigorous validation process, encompassing input from both linguistic and technical experts in the field. The result is an explicit collection of sixty emotive keywords that will significantly contribute to the metaverse realm by adding a layer of emotional depth to enrich the AR-based MLA experience. These findings offer valuable guidance for practitioners and researchers, advancing the landscape of MLA interface design and ultimately boosting UX in the educational sector.

Author 1: Nik Azlina Nik Ahmad
Author 2: Munaisyah Abdullah
Author 3: Ahmad Iqbal Hakim Suhaimi
Author 4: Anitawati Mohd Lokman

Keywords: Affective engineering; emotional design; human factor; Kansei engineering; metaverse library; mobile augmented reality; user experience

PDF

Paper 28: The Impact of Dual Objective Optimization Model Combining Non-Dominated Genetic Algorithm on Rural Industrial Ecological Economy

Abstract: Due to the development of industrial economy, it has caused serious damage to the ecological environment. Based on the industrial structure and production scale, rural industrial economic parks are planned to analyze the quantity and weight of pollutants emitted from the original industries. The results showed that the quantity and weight of hydrogen sulfide in the coking industry were 10kg/t and 94, respectively. The weight of smoke and carbon monoxide in the steelmaking industry was relatively high, with 54 and 34, respectively. Non-dominated sorting genetic algorithm and multi-objective programming model are used to optimize the comprehensive benefits and industrial structure of rural industrial ecological economy. According to the experimental results, when the scale of the coking industry was 135600 tons, the steelmaking industry was 314900 tons, the ironmaking industry was 148100 tons, and the underground coal gasification industry was 424.76 million Nm3. The comprehensive economic benefits of the industry reached the optimal level of 0.6415. The environmental and comprehensive benefits generated by the increased power generation industry were 64.98 and 40.87, respectively. Therefore, it indicates that the dual objective programming model combining non-dominated sorting genetic algorithm can improve the rural industrial ecological economy.

Author 1: Ying Wang

Keywords: Industrial chain production mode; ecological economy; environmental benefits; non-dominated sorting genetic algorithm; dual objective programming model

PDF

Paper 29: Performance Enhancement of Wi-Fi Fingerprinting-based Indoor Positioning using Truncated Singular Value Decomposition and LSTM Model

Abstract: Wi-Fi based indoor positioning has been considered as the most promising approach for civil location-based service due to the widespread availability Wi-Fi systems in many buildings. One of the most favorable approaches is to employ received signal strength indicator (RSSI) of Wi-Fi access points as the signals for estimating the mobile object locations. However, developing a solution to obtain high positioning accuracy while reducing system complexity using traditional methods as well as deep learning based methods is still a very challenging task. This paper presents a proposal to combine the Truncated Singular Value Decomposition (SVD) technique with a Long Short -Term Memory (LSTM) model to enhance the performance of indoor positioning system. Experimental results on a public dataset demonstrate that the proposed approach outperforms other state-of-the-art solutions by means of positioning accuracy as well as computational cost.

Author 1: Duc Khoi Nguyen
Author 2: Thi Hang Duong
Author 3: Le Cuong Nguyen
Author 4: Manh Kha Hoang

Keywords: Indoor positioning; Wi-Fi fingerprinting; Truncated Singular Value Decomposition; LSTM

PDF

Paper 30: Design and Implementation of an Information Management System for College Students in Higher Education Institutions Based on Cloud Computing

Abstract: A cloud computing-based system has been developed to enhance the efficiency and practicality of the information management system for college students in higher vocational colleges. This system incorporates a well-defined architecture that leverages cloud computing technology. The management layer's logic module ensures the security of vocational college students' information by deploying virtual gateways at strategic points within the system, thereby controlling access, sharing, and exchange of information. The resource module in the application layer optimizes server cluster load balancing by minimizing task completion time and improving load balancing effectiveness. Additionally, the M-Cloud storage mode is employed to store and back up application layer cloud information, along with the distributed Bigtable information base. The user access layer provides users with convenient services through the corresponding cloud service access interface in the application layer. Furthermore, the employment information of college students and enterprise position information are clustered using the K-means algorithm based on data mining, and personalized employment recommendations are made using similarity calculations. Experimental results demonstrate that the system boasts a user-friendly interface design, efficient operation, and comprehensive management functions. The system's server cluster exhibits strong load-balancing capabilities, effectively mitigating network congestion and minimizing the risks of network storms and paralysis.

Author 1: Mo Bin

Keywords: Cloud computing; student information; management system; load balancing; virtual gateway; personalized recommendation

PDF

Paper 31: A Deep Learning-based Method for Determining Semantic Similarity of English Translation Keywords

Abstract: In the English translation task, the semantics of context play an important role in correctly understanding the subtle differences between keywords. The bidirectional LSTM includes a positive LSTM and a reverse LSTM. When processing sequence data, you can consider the information of the preceding and following text at the same time. Therefore, to capture the subtle semantic differences between English translation keywords and accurately evaluate their similarity, a new semantic similarity determination method for English translation keywords is studied with the bidirectional LSTM neural network in deep learning as the main algorithm. This method introduces an English translation keyword extraction algorithm based on word co-occurrence and uses the co-occurrence relationship between words to identify and extract keywords in English translation. The extracted keywords are input into the bidirectional LSTM neural network keyword semantic similarity judgment model based on deep learning, and the weight of the bidirectional LSTM neural network is set by using the sparrow search algorithm to optimize. After the bidirectional LSTM neural network is trained, the information on keyword word vectors is captured, and the similarity between keyword word vectors is evaluated. The experimental results show that the sentence similarity calculated by the proposed method for English translation is very close to the result of professional manual scoring. The Spearman rank correlation coefficient of the semantic similarity determination result is 1, and the determination result is accurate.

Author 1: Wu Zhili
Author 2: Zhang Qian

Keywords: Deep learning; English translation; keyword; semantic similarity; co-occurrence of words; bidirectional LSTM neural network

PDF

Paper 32: An Improved VMD and Wavelet Hybrid Denoising Model for Wearable SSVEP-BCI

Abstract: The brain-computer interface (BCI) based on steady-state visual evoked potentials (SSVEP) has attracted considerable attention due to its non-invasiveness, low user training requirements, and efficient information transfer rate. To optimize the accuracy of SSVEP detection, we propose an innovative hybrid EEG denoising model combining variational mode decomposition (VMD) with wavelet packet transform(WPT). This model ingeniously integrates VMD decomposition and WPT denoising techniques, employing detrended fluctuation analysis (DFA) thresholding to deeply filter the noisy data collected from wearable devices. The filtered components are then reconstructed alongside the unprocessed components. Finally, three classification algorithms are used to validate the proposed method on a wearable SSVEP-BCI dataset. Our proposed algorithm achieves accuracies of 71.27% and 86.35% on dry and wet electrodes, respectively. Comparing the use of VMD combined with adaptive wavelet denoising and direct denoising with VMD, the classification accuracy of our method improved by 3.68% and 0.26% on dry electrodes, respectively, and by 3.28% and 0.66% on wet electrodes, respectively. The proposed approach demonstrates excellent performance and holds promising potential for application and generalization in the field of wearable EEG denoising.

Author 1: Yongquan Xia
Author 2: Keyun Li
Author 3: Duan Li
Author 4: Jiaofen Nan
Author 5: Ronglei Lu

Keywords: Brain-computer interface; steady-state visual evoked potential; style; variational mode decomposition; wavelet packet transform

PDF

Paper 33: Analyzing Quantity-based Strategies for Supply Chain Sustainability and Resilience in Uncertain Environment

Abstract: In today's interconnected world, where supply chains are the backbone of commerce, ensuring their resilience and sustainability is paramount. This study investigates how quantity-based strategies in supply chain networks are influenced by sustainability and resilience considerations. A conceptual framework is devised, focusing on a two-echelon supply chain network comprising a central supplier and multiple stores. A stochastic mathematical model is constructed to tackle demand uncertainty while incorporating parameters related to sustainability and resilience. Competitive negotiations between suppliers and stores aim at maximizing expected profits. Two store configurations are examined: non-cooperative and cooperative. Supplier resilience is reinforced through strategies like security stocks and diversified sourcing, while sustainability efforts are considered by the supplier and stores. Results show that demand following a uniform distribution benefits stores and suppliers, and cooperative behavior among stores leads to higher profitability. Sustainability initiatives impact expected profits, with security stocks particularly advantageous for supplier profitability. The utilization of foreign products has a detrimental effect on expected profits, emphasizing the significance of government regulation via customs fees. The study underscores the importance of integrating sustainability and resilience in supply chain networks. It concludes with reflections on model limitations and proposes avenues for future research in this domain.

Author 1: Dounia SAIDI
Author 2: Aziz AIT BASSOU
Author 3: Mustapha HLYAL
Author 4: Jamila EL ALAMI

Keywords: Supply chain management; competition; sustainability; resilience; demand uncertainty

PDF

Paper 34: Hardhat-YOLO: A YOLOv5-based Lightweight Hardhat-Wearing Detection Algorithm in Substation Sites

Abstract: Accidents at substation sites have occurred frequently in recent years due to workers violating power safety regulations by not wearing hardhats. Therefore, it is necessary to provide real-time warnings when detecting workers without hardhats. Nevertheless, the deployment of deep learning-based algorithms necessitates the utilization of a multitude of parameters and computations, which consequently engenders an augmented expenditure on hardware. Therefore, using a lightweight backbone can address this issue well. This paper explored methods, such as deep learning, power Internet of Things (PIoT), and edge computing and proposed a lightweight and effective method called hardhat-YOLO for hardhat-wearing detection. First, the MobileNetv3-small backbone replaced the backbone of You Only Look Once (YOLO) v5s to reduce parameters and increase detection speed. In addition, the Convolutional Block Attention Module (CBAM) was effectively integrated into the network to improve detection precision. Finally, the hardhat-YOLO model, trained with a customized dataset, was transmitted to edge computing terminals in substations through PIoT for hardhat-wearing detection. Compared to the YOLOv5s model, the Parameters and Giga Floating Point Operations (GFLOPs) of the proposed model decreased by about 35.5% and 54.4%, respectively, and Frame per Second (FPS) increased by 17.3% approximately. The experimental results indicated that the hardhat-YOLO model achieved a Mean Average Precision of 83.3% at 50% intersection over union (mAP50), correctly and effectively conducting hardhat-wearing detection tasks.

Author 1: Wanbo Luo
Author 2: Ahmad Ihsan Mohd Yassin
Author 3: Khairul Khaizi Mohd Shariff
Author 4: Rajeswari Raju

Keywords: Hardhat-wearing detection; You Only Look Once (YOLO); MobileNet; Substation; power Internet of Things (PIoT)

PDF

Paper 35: Model for Responsive Agriculture Hub via e-Commerce to Sustain Food Security

Abstract: Ensuring food security in the face of evolving environmental, economic, and societal challenges requires innovative solutions that leverage emerging technologies. This paper proposes a model for a responsive agriculture hub facilitated through e-commerce platforms to address the dynamic demands of food production, distribution, and consumption. The model integrates data-driven decision-making, supply chain optimization, and digital marketplaces to enhance the efficiency and resilience of agricultural systems. By harnessing real-time data analytics, predictive algorithms, and smart logistics, the proposed hub enables agile responses to fluctuating market conditions, climatic variability, and resource constraints. Through case studies and simulation analyses, we demonstrate the effectiveness of the model in enhancing the accessibility, affordability, and sustainability of food systems. Furthermore, we discuss the implications of this approach for stakeholders across the agricultural value chain, including farmers, distributors, retailers, and consumers. The findings underscore the potential of leveraging e-commerce platforms as catalysts for transformative change in agriculture, contributing to the overarching goal of achieving food security in an increasingly uncertain world.

Author 1: Wan Nurhayati Wan Ab. Rahman
Author 2: Wan Nurfarah Wan Zulkifli
Author 3: Nur Nabilah Zainuri
Author 4: Hanis Amira Khairol Anwar

Keywords: Digital agriculture hub model; digital value chain; responsive agriculture hub; food security; multi-sided e-commerce

PDF

Paper 36: Digital Public System of Urban Art: Navigating Human-Computer Interaction in Artistic Design for Innovative Urban Expressions

Abstract: The convergence of digital technology and urban art has given rise to novel urban art digitalization systems. This paper investigates the relationship between Human-Computer Interaction (HCI) and creative design, particularly in the age of 3D printing, Virtual Reality (VR), and digital art. We highlight the transformative potential of these technologies using examples that include interactive public installations and VR art exhibitions, thereby providing empirical evidence to ground our discussion of the evolving paradigms of technology and public art mutual constitution. We also contribute prescriptive guidance for bringing digital art into cities. We hope to offer a full guide to understanding how digital innovation could catalyze the growth and evolution of the wealth of cultural assets that characterize cities.

Author 1: Yuan Yao
Author 2: Ying Liu

Keywords: Urban art; digitalization; human-computer interaction; creative expression; public art; technological innovation

PDF

Paper 37: Fusion Lightweight Steel Surface Defect Detection Algorithm Based on Improved Deep Learning

Abstract: In industrial production, timely and accurate detection and identification of surface defects in steel materials were crucial for ensuring product quality, enhancing production efficiency, and reducing production costs. This study addressed the problem of surface defect detection in steel materials by proposing an algorithm based on an improved version of YOLOv5. The algorithm achieved lightweight and high efficiency by incorporating the MobileNet series network. Experimental results demonstrated that the improved algorithm significantly reduced inference time and model file size while maintaining performance. Specifically, the YOLOv5-MobileNet-Small model exhibited slightly lower performance but excelled in inference time and model file size. On the other hand, the YOLOv5-MobileNet-Large model achieved a slight performance improvement while significantly reducing inference time and model file size. These results indicated that the improved algorithm could achieve lightweighting while maintaining performance, showing promising applications in steel surface defect detection tasks. It provided an efficient and feasible solution for this important domain, offering new insights and methods for similar surface defect detection problems and contributing to research and applications in related fields.

Author 1: Fei Ren
Author 2: Jiajie Fei
Author 3: HongSheng Li
Author 4: Bonifacio T. Doma Jr

Keywords: Deep learning; improved YOLOv5; YOLOv5-mobilenet; surface defects

PDF

Paper 38: AdvAttackVis: An Adversarial Attack Visualization System for Deep Neural Networks

Abstract: Deep learning has been widely used in various scenarios such as image classification, natural language processing, and speech recognition. However, deep neural networks are vulnerable to adversarial attacks, resulting in incorrect predictions. Adversarial attacks involve generating adversarial examples and attacking a target model. The generation mechanism of adversarial examples and the prediction principle of the target model for adversarial examples are complicated, which makes it difficult for deep learning users to understand adversarial attacks. In this paper, we present an adversarial attack visualization system called AdvAttackVis to assist users in learning, understanding, and exploring adversarial attacks. Based on the designed interactive visualization interface, the system enables users to train and analyze adversarial attack models, understand the principles of adversarial attacks, analyze the results of attacks on the target model, and explore the prediction mechanism of the target model for adversarial examples. Through real case studies on adversarial attacks, we demonstrate the usability and effectiveness of the proposed visualization system.

Author 1: DING Wei-jie
Author 2: Shen Xuchen
Author 3: Yuan Ying
Author 4: MAO Ting-yun
Author 5: SUN Guo-dao
Author 6: CHEN Li-li
Author 7: CHEN bing-ting

Keywords: Deep learning; deep neural networks; adversarial attacks; adversarial examples; interactive visualization

PDF

Paper 39: Exploring the Impact of PCA Variants on Intrusion Detection System Performance

Abstract: Intrusion detection systems (IDS) play a critical role in safeguarding network security by identifying malicious activities within network traffic. However, the effectiveness of an IDS hinges on its ability to extract relevant features from the vast amount of data it collects. This study investigates the impact of different feature extraction methods on the performance of IDS. We compare the performance of various feature extraction techniques on two widely used intrusion detection datasets: KDD Cup 99 and NSL-KDD. By evaluating these techniques on both datasets, we aim to gain insights into the generalizability and robustness of each method across different dataset characteristics. The study compares the performance of these methods using standard metrics like detection rate, F-measure and FPR for intrusion detection.

Author 1: CHENTOUFI Oumaima
Author 2: CHOUKHAIRI Mouad
Author 3: CHOUGDALI Khalid
Author 4: ALLOUG Ilyas

Keywords: Intrusion detection; dimensionality reduction; feature extraction; KDDCup’99; NSL-KDD

PDF

Paper 40: Enhancing Whale Optimization Algorithm with Differential Evolution and Lévy Flight for Robot Path Planning

Abstract: Path planning is a prominent and essential part of mobile robot navigation in robotics. It allows robots to determine the optimal path from a given beginning point to a desired end goal. Additionally, it enables robots to navigate around obstacles, recognize secure pathways, and select the optimal route to follow, considering multiple aspects. The Whale Optimization Algorithm (WOA) is a frequently adopted approach to planning mobile robot paths. However, conventional WOA suffers from drawbacks such as a sluggish convergence rate, inefficiency, and local optimization traps. This study presents a novel methodology integrating WOA with Lévy flight and Differential Evolution (DE) to plan robot paths. As WOA evolves, the Levy flight promotes worldwide search capabilities. On the other hand, DE enhances WOA's ability to perform local searches and exploitation while also maintaining a variety of solutions to avoid getting stuck in local optima. The simulation results demonstrate that the proposed approach offers greater planning efficiency and enhanced route quality.

Author 1: Rongrong TANG
Author 2: Xuebang TANG
Author 3: Hongwang ZHAO

Keywords: Path planning; mobile robot; differential evolution; Whale Optimization Algorithm; lévy flight

PDF

Paper 41: A Capacitance-base System Design for Measurement of Crude Oil Moisture

Abstract: Challenges including difficulty in cleaning and low measurement accuracy widely exist in traditional methods for measuring moisture in crude oil. In order to solve these problems, a capacitance measurement device that combines PCAP01 and STM32 has been designed. PCAP01 is employed as the processing core of the sensor, which significantly enhances the measurement accuracy of capacitance-based methods. As to STM32, it plays a critical role in data acquisition, signal processing, and data transmission. Besides, the capacitance-based device contains two symmetric half-cylindrical electrode plates that are closely attached to the outer wall of the cylindrical glass vessel, where the crude oil sample to be tested is contained. This design prevents the direct contact between the liquid sample and the electrode plates, thus eliminating issues related to cleaning difficulties. Time-frequency domain expansion is presented to realize the fit between moisture and the capacitance. Experimental results indicate that the designed system delivers a high accuracy across the entire 0-100% range.

Author 1: ZhixueShi
Author 2: Xudong Zhao

Keywords: Capacitance measurement; crude oil moisture; PCAP01; STM32; time-frequency domain

PDF

Paper 42: SchemaLogix: Advancing Interoperability with Machine Learning in Schema Matching

Abstract: Schema matching, a fundamental process in data integration, traditionally employs pairwise comparisons to discern semantic correspondences among elements in disparate schemas. However, recent developments underscore the necessity of concurrent matching of interconnected schemas, termed schema alignment, to reconcile heterogeneous elements. This paper presents SchemaLogix, an innovative machine learning-based approach for schema matching. SchemaLogix addresses challenges such as data scarcity and domain-specific constraints through an inventive bootstrapping method, autonomously generating extensive datasets. Furthermore, SchemaLogix capitalizes on inherent alignment context constraints to optimize learning and improve precision across varied schema structures. Additionally, SchemaLogix incorporates user contributions to validate chosen correspondences, refining outputs based on valuable feedback. Empirical evaluations establish SchemaLogix's superiority over traditional methods, achieving an exceptional maximum S1 score of 0.90. These results offer practical insights for real-world applications, substantially advancing data integration and interoperability endeavors.

Author 1: Mohamed Raoui
Author 2: Mohammed Ennaouri
Author 3: Moulay Hafid El Yazidi
Author 4: Ahmed Zellou

Keywords: Interoperability; data integration; schema matching; machine learning

PDF

Paper 43: Image Segmentation in Complex Backgrounds using an Improved Generative Adversarial Network

Abstract: As technology advances, solving image segmentation challenges in complex backgrounds has become a key issue across various fields. Traditional image segmentation methods underperform in addressing these challenges, and existing generative adversarial networks (GANs) also face several problems when applied in complex environments, such as low generation quality and unstable model training. To address these issues, this study introduces an improved GAN approach for image segmentation in complex backgrounds. This method encompasses preprocessing of complex background image datasets, feature reduction encoding based on cerebellar neural networks, image data augmentation in complex backgrounds, and the application of an improved GAN. In this paper, new generator and discriminator network structures are designed and image data enhancement is implemented through self-play learning. Experimental results demonstrate significant improvements in image segmentation tasks in various complex backgrounds, enhancing the accuracy and robustness of segmentation. This research offers new insights and methodologies for image processing in complex backgrounds, holding substantial theoretical and practical significance.

Author 1: Mei Wang
Author 2: Yiru Zhang

Keywords: Generative Adversarial Networks (GANs); complex backgrounds; image segmentation; data augmentation; feature dimensionality reduction encoding

PDF

Paper 44: A Novel Quantum Orthogonal Frequency-Division Multiplexing Transmission Scheme

Abstract: Recently, extensive research attention has been dedicated to enabling Orthogonal Frequency-Division Multiplexing (OFDM) waveforms to be compatible with a modern communication system. Encoding data as OFDM wavelengths still has a lot of problems, like the peak-to-average power ratio (PAPR) and the cyclic prefix (CP), which are important factors that affect how efficiently the spectrum is used. To meet the quality-of-service requirements imposed by communication system applications, this paper proposes to replace the classical encoding and decoding schemes, classical channel, discrete Fourier transform (DFT), and inverse discrete Fourier transform (IDFT) with their classical counterparts. This new quantum OFDM transmission scheme allows for the preparation of a quantum OFDM symbol without the need to incorporate a CP. To validate the accuracy of the suggested quantum OFDM transmission scheme, we compared it with the most widely recognised reference quantum transmission scheme. We have demonstrated that increasing the channel resistivity results in a higher probability of correctly measuring the quantum state in the quantum OFDM transmission scheme compared to the reference quantum transmission scheme. The results are verified by IBM's Qiskit.

Author 1: Mohammed R. Almasaoodi
Author 2: Abdulbasit M. A. Sabaawi
Author 3: Sara El Gaily
Author 4: Sándor Imre

Keywords: Discrete Fourier transform; quantum Fourier transform; orthogonal frequency-division multiplexing; quantum channel

PDF

Paper 45: Deep Learning-based Classification of MRI Images for Early Detection and Staging of Alzheimer's Disease

Abstract: Alzheimer's disease (AD) poses a significant challenge to modern healthcare, as effective treatment remains elusive. Drugs may slow down the progress of the disease, but there is currently no cure for it. Early AD identification is crucial for providing the required medications before brain damage occurs. In this course of research, we studied various deep learning techniques to address the challenge of early AD detection by utilizing structural MRI (sMRI) images as biomarkers. Deep learning techniques are pivotal in accurately analyzing vast amounts of MRI data to identify Alzheimer's and anticipate its progression. A balanced MRI image dataset of 12,936 images was used in this study to extract sufficient features for accurately distinguishing Alzheimer’s disease stages, due to the similarities in the characteristics of its early stages, necessitating more images than previous studies. The GoogLeNet model was utilized in our investigation to derive features from each MRI scan image. These features were then inputted into a feed-forward neural network (FFNN) for AD stage prediction. The FFNN model, utilizing GoogLeNet features, underwent rigorous training over multiple epochs using a small batch size to ensure robust performance on unseen data and achieved 98.37% accuracy, 98.39% sensitivity, 98.50% precision, and 99.45% specificity. Most remarkably, our results show that the model detected AD with an amazing average accuracy rate of 99.01%.

Author 1: Parvatham Niranjan Kumar
Author 2: Lakshmana Phaneendra Maguluri

Keywords: Alzheimer’s disease (AD); Convolution Neural Network (CNN); Deep Learning (DL); Transfer Learning (TL); imaging pre-processing

PDF

Paper 46: Image Processing-based Performance Evaluation of KNN and SVM Classifiers for Lung Cancer Diagnosis

Abstract: It is important to note that the cure rates in cases of advanced stages of lung cancer are remarkably low, which stresses out the importance for early detection as means to increase survival chances. A strong area of focus when it comes to increased research in the lung cancer diagnosis is the search for ways through which this disease can be identified at its early stages. The methodology described below is proposed as a means to facilitate early detection of lung cancer There are two phases in this approach. The study deals with effectiveness of three types of classifiers K-Nearest Neighbors (KNN), Random Forest and Support Vector Machine (SVM) to identify cases related to lung cancer via relevant medical data assessment. In this application, the eval axis performs profiling or measures the accuracy of applying these classifiers and discriminating between cancerous instances versus non-cancerous ones within the dataset. To rate the adequacy of classifiers in distinguishing classes, performance metrics like accuracy, precision, recall and F1- score are used. Furthermore, the research compares KNN, Random Forest and SVM, explaining their specific advantages as well as disadvantages logically referring to how they can or cannot be applied while detecting lung cancer. This investigation shows helpful results in suggesting the possibility that machine learning techniques could assist to identify lung cancer as exact and timely as possible, providing more successful diagnostic procedures and patient outcomes. The experimental findings show that SVM gives the best result at 95.06%, KNN comes second with a percentage of 86.89.

Author 1: Kavitha B C
Author 2: Naveen K B

Keywords: K-Nearest Neighbors; lung cancer detection; machine learning; medical data; performance metrics; support vector machine

PDF

Paper 47: Generative AI-Powered Predictive Analytics Model: Leveraging Synthetic Datasets to Determine ERP Adoption Success Through Critical Success Factors

Abstract: Data scarcity is a significant problem in Enterprise Resource Planning (ERP) adoption prediction, limiting the accuracy and reliability of traditional predictive models. This study addresses this issue by integrating Generative Artificial Intelligence (AI) technologies, specifically Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs), to generate synthetic data that supplements sparse real-world data. A systematic literature review identified critical gaps in existing ERP adoption models, underscoring the need for innovative approaches. The generated synthetic data, validated through comprehensive statistical analyses including mean, variance, skewness, kurtosis, and the Kolmogorov-Smirnov test, demonstrated high accuracy and reliability, aligning closely with real-world data. A hybrid predictive model was developed, combining Generative AI with Pearson Correlation Coefficient (PCC) and Random Forest techniques. This model was rigorously tested and compared against traditional models such as SVM, Neural Networks, Linear Regression, and Decision Trees. The hybrid model achieved superior performance, with an accuracy of 90%, precision of 88%, recall of 89%, and Area Under the Receiver Operating Characteristic Curve (AUC-ROC) score of 0.91, significantly outperforming traditional models in predicting ERP adoption outcomes. The research also established continuous monitoring and adaptation mechanisms to ensure the model's long-term effectiveness. The findings provide practical insights for organizations, offering a robust tool for forecasting ERP adoption success and facilitating more informed decision-making and resource allocation. This study not only advances theoretical understanding by addressing data scarcity through synthetic data generation but also provides a practical framework for enhancing ERP adoption strategies.

Author 1: Koh Chee Hong
Author 2: Abdul Samad Bin Shibghatullah
Author 3: Thong Chee Ling
Author 4: Samer Muthana Sarsam

Keywords: ERP adoption; predictive analytics; generative AI; synthetic data; GANs; VAEs; Pearson's correlation coefficient; random forest

PDF

Paper 48: An Investigation of Scalability in EHRs using Healthcare 4.0 and Blockchain

Abstract: In the past decade, Electronic Health Records (EHRs) based on clouds have become popular in empowering remote patient monitoring. The rise of Health 4.0, which includes using system elements and cloud services to access health records remotely, has gained highest attention of the experts. Healthcare 4.0 requires the consistent collection, combination, transmission, exchange, and storage of medical information related to the patients. Because patient information is a private data, it might be challenging to keep hackers out of the reach. As a result, secure cloud storage, access, and exchange of patient medical information is critical in ensuring that the information is not exposed in any unauthorized manner. Security mechanisms that employ Blockchain technology have become popular in recent years since they can provide robust data sharing amongst large number of users and provide storage protection with low computing costs. Researchers have now shifted their focus to using Blockchain to protect healthcare information administration. This work presents an architecture to investigate the scalability of the Healthcare 4.0 systems that use Blockchain. The investigations are carried out under different test scenarios and are evaluated under numerous circumstances, including varying user and data volumes, while also considering the presence of cyber threats. The results demonstrate interesting findings related to the efficiency and effectiveness of deploying Healthcare 4.0 and Blockchain in EHRs.

Author 1: Ahmad Fayyaz Madni
Author 2: Munam Ali Shah
Author 3: Muhammad Al-Naeem

Keywords: EHRs; secure cloud; Healthcare 4.0; Blockchain; scalability; cyber threats; medical information; security

PDF

Paper 49: Traffic Flow Prediction at Intersections: Enhancing with a Hybrid LSTM-PSO Approach

Abstract: The growing challenge of increasing traffic volumes presents a real challenge for road safety, emergency response and overall transport efficiency. Intelligent transportation systems play a fundamental role in solving these challenges, through accurate traffic prediction. In this study, we propose a hybrid model that combines the Long-Term Memory Algorithm (LSTM) and Particle Swarm Optimization (PSO) to predict traffic flow more accurately at intersections. Our approach takes advantage of the strength of PSO, a robust optimization technique inspired by swarm intelligence, to optimize the hyperparameters of the LSTM algorithm. Through in-depth benchmarking, we evaluate the performance of our hybrid LSTM-PSO model against other existing models. By evaluating measures such as root mean square error and mean absolute error, we demonstrate the superior efficiency of the proposed hybrid model. Our results highlight the effectiveness of our approach in outperforming alternative models, offering a promising solution for intelligent transportation systems to accurately predict traffic flow at intersections and improve overall traffic management efficiency.

Author 1: Chaimaa CHAOURA
Author 2: Hajar LAZAR
Author 3: Zahi JARIR

Keywords: Deep learning; intersection congestion; intelligent transport systems; traffic flow prediction

PDF

Paper 50: Remote Palliative Care: A Systematic Review of Effectiveness, Accessibility, and Patient Satisfaction

Abstract: Remote palliative care has emerged as a viable option to address the complex needs of patients facing life-limiting illnesses, particularly in the context of evolving healthcare landscapes and technological advancements. This systematic review aims to comprehensively examine the effectiveness, accessibility, and patient satisfaction of remote palliative care interventions. Through a meticulous analysis of empirical studies, clinical trials, and qualitative research, this review synthesizes evidence about the impact of remote palliative care on clinical outcomes, patient access to services, and overall satisfaction levels. Our findings highlight the benefits of remote palliative care, including improved symptom management, enhanced patient autonomy, and greater convenience in accessing care, particularly for individuals in rural or underserved areas. Moreover, we identify key facilitators and barriers influencing the implementation and uptake of remote palliative care services, such as technological proficiency, infrastructure limitations, and concerns regarding the quality of interpersonal communication. By critically evaluating the existing literature, this review underscores the significance of remote palliative care as a patient-centred approach to delivering compassionate end-of-life care. Furthermore, it underscores the need for ongoing research efforts and policy initiatives to optimize the effectiveness and accessibility of remote palliative care services to ensure equitable and high-quality care for all patients facing serious illnesses.

Author 1: Rihab El Sabrouty
Author 2: Abdelmajid Elouadi
Author 3: Mai Abdou Salifou Karimoune

Keywords: Palliative care; eHealth; patient; artificial intelligence; well-being

PDF

Paper 51: Enhancing SDN Anomaly Detection: A Hybrid Deep Learning Model with SCA-TSO Optimization

Abstract: The paper explores the evolving landscape of network security, in Software Defined Networking (SDN) highlighting the challenges faced by security measures as networks transition to software-based control. SDN revolutionizes Internet technology by simplifying network management and boosting capabilities through the OpenFlow protocol. It also brings forth security vulnerabilities. To address this we present a hybrid Intrusion Detection System (IDS) tailored for SDN environments leveraging a state of the art dataset optimized for SDN security analysis along with machine learning and deep learning approaches. This comprehensive research incorporates data preprocessing, feature engineering and advanced model development techniques to combat the intricacies of cyber threats in SDN settings. Our approach merges feature from the sine cosine algorithm (SCA) and tuna swarm optimization (TSO) to optimize the fusion of Long Short Term Memory Networks (LSTM) and Convolutional Neural Networks (CNN). By capturing both spatial aspects of network traffic dynamics our model excels at detecting and categorizing cyber threats, including zero-day attacks. Thorough evaluation includes analysis using confusion matrices ROC curves and classification reports to assess the model’s ability to differentiate between attack types and normal network behavior. Our research indicates that improving network security using software defined methods can be achieved by implementing learning and machine learning strategies paving the way, for more reliable and effective network administration solutions.

Author 1: Ahmed Mohanad Jaber ALHILO
Author 2: Hakan Koyuncu

Keywords: SDN; Intrusion Detection System; deep learning; CNN; LSTM; SCA; TSO

PDF

Paper 52: Comprehensive and Simulated Modeling of a Centralized Transport Robot Control System

Abstract: This work proposes a new simulation model for a centralized transport robot control system that was created with the AnyLogic environment and a special blend of agent-based and discrete-event approaches. The model attempts to do a comprehensive analysis of the centralized request distribution algorithm among robots, gauging the effectiveness of the transport system based on service arrival times. For in-depth testing, a transport robot model was developed using Arduino microcontrollers and NRF24L01 transceivers for communication. Item movement test sequences were created to be uniform in both full-scale and simulation testing. Good, though not perfect, agreement was found between the simulation and experimental results, underscoring the difficulty of obtaining high accuracy in real-time coordinate identification in the absence of sensors. This shortcoming notwithstanding, the novel simulation model provides an invaluable instrument for determining the viability and efficiency of transportation systems as well as analyzing decentralized control mechanisms prior to actual deployment. The novelty of this paper in that it builds a thorough simulation model for a centralized transport robot control system using an AnyLogic environment and a unique blend of discrete-event and agent-based approaches. This comprehensive technique is a novel contribution to the discipline since it enables a thorough evaluation of a centralized request distribution system.

Author 1: Murad Bashabsheh

Keywords: Artificial intelligence; centralized control system; transport robots; automatic system; agent-based modeling; AnyLogic; Arduino microcontroller

PDF

Paper 53: Estimating Stock Market Prices with Histogram-based Gradient Boosting Regressor: A Case Study on Alphabet Inc

Abstract: One of the most important and common activities mentioned while discussing the financial markets is stock market trading. An investor is constantly searching for methods to estimate future trends to minimize losses and maximize profits due to the unavoidable volatility in stock prices. It is undeniable, nonetheless, that there is currently no mechanism for accurately estimating future market patterns despite numerous approaches being investigated to enhance model performance as much as feasible. Findings indicate notable improvements in accuracy compared to traditional Histogram-based gradient-boosting models. Experiments conducted on historical stock price datasets verify the efficacy of the proposed method. The combined strength of HGBoost and optimization techniques, including Particle Swarm Optimization, Slime Mold Algorithm, and Grey Wolf Optimization, not only increases prediction accuracy but also fortifies the model's ability to adjust to changing market conditions. The results for HGBoost, PSO- HGBoost, SMA- HGBoost, and GWO- HGBoost were 0.964, 0.973, 0.981, and 0.988, in that order. Compared to HGBoost, the result of GWO- HGBoost shows how combining with the optimizer can enhance the output of the given model.

Author 1: Shigen Li

Keywords: Alphabet Inc.; market movement; stock; financial markets; Histogram-based gradient boosting

PDF

Paper 54: A Study on Wireless Sensor Node Localization and Target Tracking Based on Improved Locust Algorithm

Abstract: To improve the positioning accuracy of wireless sensor nodes and ensure the target tracking effect, a wireless sensor node positioning and target tracking method based on an improved locust algorithm is proposed. The DV Hop algorithm is used to calculate the minimum hops and average hops distance between the unknown node and each anchor node to obtain the location of the unknown node, realize the rough positioning of wireless sensor nodes, and analyze the positioning error to determine the positioning accuracy target function; The improved locust algorithm is used to solve the positioning accuracy objective function to obtain the sensor node positioning results with the minimum error; The target tracking model and the target is calculated. According to the target observation information obtained by all sensor nodes, the target state in the wireless sensor network model is tracked using the probability hypothesis density filtering algorithm. The test results show that the algorithm has better performance, the spatial evaluation index results are all lower than 0.020, and the individual distribution in the solution set is better; The location of each unknown node in different node distribution states can be obtained; The positioning error under the surface and plane is less than 0.012; The maximum error of target tracking is 0.142m; It can track single target and multiple targets.

Author 1: Tan SONGHE
Author 2: Qin Qi

Keywords: Improved locust algorithm; wireless sensors; node localization; target tracking; target state; unknown node position

PDF

Paper 55: Toward Optimal Service Composition in the Internet of Things via Cloud-Fog Integration and Improved Artificial Bee Colony Algorithm

Abstract: In the quest to delve deeper into the burgeoning realm of the service-oriented Internet of Things (IoT), the pressing challenge of smoothly integrating functionalities within smart objects emerges prominently. IoT devices, notorious for their resource constraints, often lean heavily on cloud infrastructures to function effectively. However, the emergence of fog computing offers a promising alternative, allowing the processing of IoT applications closer to the sensors and thereby slashing delays. This research develops a novel method for IoT service composition that leverages both fog and cloud computing, utilizing an enhanced version of the Artificial Bee Colony (ABC) algorithm to refine its convergence rate. The approach introduces a Dynamic Reduction (DR) mechanism designed to perturb dimensions innovatively. Traditionally, the ABC algorithm generates new solutions that closely mimic their parent solutions, which unfortunately slows down convergence. By initiating the process with significant dimension disparities among solutions and gradually reducing these disparities over successive iterations, this method strikes an optimal balance between exploration and exploitation through dynamic adjustment of dimension perturbation counts. Comparative analyses against contemporary methodologies reveal significant improvements: a 17% decrease in average energy consumption, a 10% boost in availability, an 8% enhancement in reliability, and a remarkable 23% reduction in average cost. Combining the strengths of fog and cloud computing with the refined ABC algorithm through the Dynamic Reduction mechanism significantly advances the efficiency and effectiveness of IoT service compositions.

Author 1: Guixia Xiao

Keywords: Internet of Things (IoT); fog computing; service composition; Artificial Bee Colony (ABC) Algorithm; Dynamic Reduction Mechanism

PDF

Paper 56: Emotion-based Autism Spectrum Disorder Detection by Leveraging Transfer Learning and Machine Learning Algorithms

Abstract: Autism Spectrum Disorder (ASD) presents as a neurodevelopmental condition impacting social interaction, communication, and behavior, underscoring the imperative of early detection and intervention to enhance outcomes. This paper introduces a novel approach to ASD detection utilizing facial features extracted from the Autistic Children Facial Dataset. Leveraging transfer learning models, including VGG16, ResNet, and Inception, high-level features are extracted from facial images. Additionally, fine-grained details are captured through the utilization of handcrafted image features such as Histogram of Oriented Gradients, Local Binary Patterns, Scale-Invariant Feature Transform, PHASH descriptors. Integration of these features yields three distinct feature vectors, combining image features with VGG16, ResNet, and Inception features. Subsequently, multiple machine learning classifiers, including Random Forest, KNN, Decision Tree, SVM, and Logistic Regression, are employed for ASD classification. Through rigorous experimentation and evaluation, the performance of these classifiers across three datasets is compared to identify the optimal approach for ASD detection. By evaluating multiple classifiers and feature combinations, this work offers insights into the most effective approaches for ASD detection.

Author 1: I. Srilalita Sarwani
Author 2: D. Lalitha Bhaskari
Author 3: Sangeeta Bhamidipati

Keywords: Autism Spectrum Disorder; transfer learning; image features; VGG; ResNet; Inception

PDF

Paper 57: Offensive Language Detection on Social Media using Machine Learning

Abstract: This research paper addresses the critical issue of cyberbullying detection within the realm of social networks, employing a comprehensive examination of various machine learning and deep learning techniques. The study investigates the performance of these methodologies through rigorous evaluation using standard metrics, including Accuracy, Precision, Recall, F-measure, and AUC-ROC. The findings highlight the notable efficacy of deep learning models, particularly the Bidirectional Long Short-Term Memory (BiLSTM) architecture, in consistently outperforming alternative methods across diverse classification tasks. Confusion matrices and graphical representations further elucidate model performance, emphasizing the BiLSTM-based model's remarkable capacity to discern and classify cyberbullying instances accurately. These results underscore the significance of advanced neural network structures in capturing the complexities of online hate speech and offensive content. This research contributes valuable insights toward fostering safer and more inclusive online communities by facilitating early identification and mitigation of cyberbullying. Future investigations may explore hybrid approaches, additional feature integration, or real-time detection systems to further refine and advance the state-of-the-art in addressing this critical societal concern.

Author 1: Rustam Abdrakhmanov
Author 2: Serik Muktarovich Kenesbayev
Author 3: Kamalbek Berkimbayev
Author 4: Gumyrbek Toikenov
Author 5: Elmira Abdrashova
Author 6: Oichagul Alchinbayeva
Author 7: Aizhan Ydyrys

Keywords: Machine learning; deep learning; hate speech; CNN; RNN; LSTM

PDF

Paper 58: A Deep Residual Network Designed for Detecting Cracks in Buildings of Historical Significance

Abstract: This research paper investigates the application of deep learning techniques, specifically convolutional neural networks (CNNs), for crack detection in historical buildings. The study addresses the pressing need for non-invasive and efficient methods of assessing structural integrity in heritage conservation. Leveraging a dataset comprising images of historical building surfaces, the proposed CNN model demonstrates high accuracy and precision in identifying surface cracks. Through the integration of convolutional and fully connected layers, the model effectively distinguishes between positive and negative instances of cracks, facilitating automated detection processes. Visual representations of crack finding cases in ancient buildings validate the model's efficacy in real-world applications, offering tangible evidence of its capability to detect structural anomalies. While the study highlights the potential of deep learning algorithms in heritage preservation efforts, it also acknowledges challenges such as model generalization, computational complexity, and interpretability. Future research endeavors should focus on addressing these challenges and exploring new avenues for innovation to enhance the reliability and accessibility of crack detection technologies in cultural heritage conservation. Ultimately, this research contributes to the development of sustainable solutions for safeguarding architectural heritage, ensuring its preservation for future generations.

Author 1: Zlikha Makhanova
Author 2: Gulbakhram Beissenova
Author 3: Almira Madiyarova
Author 4: Marzhan Chazhabayeva
Author 5: Gulsara Mambetaliyeva
Author 6: Marzhan Suimenova
Author 7: Guldana Shaimerdenova
Author 8: Elmira Mussirepova
Author 9: Aidos Baiburin

Keywords: Crack detection; historical buildings; deep learning; convolutional neural networks; heritage conservation; image analysis; machine learning; non-destructive testing; preservation

PDF

Paper 59: Mobile Application with Augmented Reality Applying the MESOVA Methodology to Improve the Learning of Primary School Students in an Educational Center

Abstract: "The application was developed using the MESOVA methodology, employing technologies such as Unity, Vuforia, and Visual Studio with the purpose of enhancing the educational experience for elementary school students. This innovative tool integrates augmented reality with the pedagogical principles of MESOVA, standing out notably from other research. Focusing on topics such as scientific knowledge and design and construction skills, the application not only provides information but also includes games that encourage interaction with the universe and planets, offering a participative and meaningful educational experience. The pretest results revealed an average scientific knowledge of 9.75%, significantly increasing to 15.55% in the posttest. Similarly, design and construction skills, initially evaluated at 8.24%, experienced a remarkable increase to 14.99% in the posttest. The adaptability of the application to the specific needs of elementary school students creates a stimulating and personalized learning environment. The combination of MESOVA and augmented reality enriches the educational experience, promoting understanding, collaboration, and critical thinking among students. In conclusion, the initiative goes beyond providing basic information; it becomes a transformative educational resource that equips students with fundamental cognitive and social skills as they explore the universe through augmented reality. Ultimately, it highlights the potential of technology and pedagogy to create a dynamic and enriching educational environment for elementary school students."

Author 1: Anthony Wilder Arias Vilchez
Author 2: Tomas Silvestre Marcelo Lloclla Soto
Author 3: Giancarlo Sanchez Atuncar

Keywords: Augmented reality; mobile application; MESOVA methodology; Kolmogorov-Smirnov; Wilcoxon; education; Vuforia

PDF

Paper 60: An Improved MobileNet Model Integrated Spatial and Channel Attention Mechanisms for Tea Disease

Abstract: Aiming at addressing the challenges of large model parameters, high computational cost, and low accuracy of the traditional tea disease identification model, an improved MobileNet model integrated spatial and channel attention mechanisms (MobileNet-SCA) was proposed for tea disease identification. Firstly, the tea disease identification dataset was augmented through random clipping, rotation transformation, and perspective transformation to simulate diverse image acquisition perspectives and mitigate overfitting effects. Secondly, based on the convolutional neural network (CNN) framework, the Channel Attention (CA) mechanism and Spatial Attention (SA) mechanism were introduced to carry out global average pooling and group normalization operations on input feature maps respectively, and adjust the channel weights using the learned parameters. Then the h-swish activation function was utilized to scale, and the two kinds of attention mechanisms were spliced and mixed to improve the channel and spatial information. In addition, the MobileNetV3 network's structure underwent optimization by adjusting the number of input channels, the size of the convolution kernel, and the number of channels in the residual block. The experimental results showed that the identification accuracy of MobileNet-SCA for tea diseases was 5.39% higher than the original model. This method can balance the identification accuracy and identification time well, and it meets the requirements for accurate and rapid identification of tea diseases.

Author 1: Li Zhang
Author 2: Jiacheng Sun
Author 3: Minghui Yang

Keywords: Tea disease; MobileNetV3; attention mechanism; convolutional neural network component

PDF

Paper 61: Tourist Attraction Recommendation Model Based on RFPAP-NNPAP Algorithm

Abstract: Driven by globalization and digitization, the tourism industry is facing new challenges and opportunities brought about by big data and artificial intelligence. The recommendation of tourist attractions, as an important part of the industry, has a direct influence on the tourist experience. However, with the diversification and personalization of tourism demand, traditional recommendation methods have shown shortcomings: weak processing ability for complex nonlinear data, affecting recommendation accuracy and personalization, and insufficient efficiency and stability when processing large-scale data. Faced with this challenge, this study proposed a hybrid tourist attraction recommendation model with random forest, artificial neural network, and frequent pattern growth. This model utilized the powerful classification and regression capabilities of random forests, as well as the complex nonlinear mapping ability of artificial neural networks, to predict tourist attraction preferences. And on this basis, the frequent pattern growth algorithm was introduced to mine the associated attractions of tourist preferences, thereby achieving accurate recommendation of tourist attractions. In experimental verification, the proposed model demonstrated superior performance. It not only surpassed traditional tourist attraction recommendation methods in accuracy and personalization, but also exhibited efficient and stable characteristics when processing large-scale data. After about 16 iterations, the MAPE value of the mixed model decreased to 0.44%. After about 39 iterations, the MAPE value of the mixed model decreased to 0.40%. The average accuracy, recall rate and F-value of the proposed model are 92.26%, 82.11% and 84.43%, respectively, which are superior to the comparison algorithm. Its error correction accuracy fluctuates around 90%. This study provides a new solution to the problem of personalized recommendation of tourist attractions, providing theoretical guidance for the tourism applications of random forests and artificial neural networks, and improving the tourist experience, promoting the development of the tourism industry.

Author 1: Jun Li

Keywords: Tourist attractions; recommendation model; RF; ANN; FP-Growth

PDF

Paper 62: Ontology Driven for Mapping a Relational Database to a Knowledge-based System

Abstract: The mapping of a relational database system to a knowledge-based system is a key stage in developing an online analytical processing (OLAP) system. OLAP is a cornerstone in discovering hidden knowledge in any business. Hence, the existence of an OLAP system is one of the modern success factors in a business environment. Mapping has proven benefits for knowledge-based systems in terms of enabling the discovery of hidden relationships among objects and the inference of new information. However, there remains room for improvement in respect of the quality of the mapping output. Therefore, in this paper, a rule-based method for mapping a relational database to a knowledge-based system is introduced. First, the proposed mapping process, which involves converting the tables and relationships of a relational database into facts and rules for a knowledge-based system, is illustrated through the use of a detailed case study. Then the correctness of the proposed method is proved by testing the tautology results against equivalent SQL queries. In addition, the completeness of the proposed method is proved by demonstrating that the used predicates are sufficient to allow a complete modeling of the required system. Furthermore, the experimental results show that the performance of the knowledge-based system that was developed using the proposed method is much better than that of an equivalent relational database.

Author 1: Abdelrahman Osman Elfaki
Author 2: Yousef H. Alfaifi

Keywords: Mapping knowledge; ontology-based; relational database; online analytical processing

PDF

Paper 63: A Novel Controlling System for Smart Farming-based Internet of Things (IoT)

Abstract: The integration of IoT systems in agriculture has become a very important need amid the high population and increasingly limited farmland, which demands researchers to be more innovative in addressing these issues. Using IoT systems for automatic irrigation, fertilization, and cooling based on sensor values through internet networks. Poor internet connection leads to the failure of automation and sustainability in online conditions, which can be very dangerous for plants. This paper presents a new IoT-based control system divided into two parts: an automation system and an IoT system, which can maintain sustainability in online conditions to ensure that plants in the planting area are always controlled. In addition, the sensors used have undergone calibration processes to determine the increase in precision of the sensor values produced. The research results show that the system can maintain sustainability under online conditions. Mobile apps are available for control when the system is online, but if it goes offline and is unable to reconnect, the Arduino Mega will fully manage control using soil moisture sensor values for irrigation processes if the values fall below a certain threshold. This demonstrates the sustainability of the system in online conditions, allowing continuous control and reducing the risk of plant death in the planting area. The calibration result shows an increase in precision for the air temperature and humidity (DHT 11 sensor) by 7.14 and 6.15, respectively. Additionally, the precision improvement for the soil pH sensor is 1.81, while for the soil moisture sensor and the water flow sensor, it is 0.13 and 0.008, respectively.

Author 1: Dodi Yudo Setyawan
Author 2: Warsito
Author 3: Roniyus Marjunus
Author 4: Sumaryo

Keywords: IoT; agriculture; automation; sustainability

PDF

Paper 64: Method of Budding Detection with YOLO-based Approach for Determination of the Best Time to Plucking Tealeaves

Abstract: Method of budding detection with YOLO (You Only Look Once) for determination of the best time to plucking tealeaves is proposed. In order to get the best quality and quantity of tealeaves, it is very important to determine the best time to plucking date. It is most likely that the number of days elapsed after the budding of the tealeaves are the most effective for determine the best plucking day. Therefore, method for detect the budding is getting much important. In this paper, YOLO-based object detection is proposed. Hyperparameter of the YOLO has to be optimized. Also, a comparative study is conducted for the resolution of the cameras used for acquisition of tealeaves from a point of view for learning performance of YOLO. Through experiments, it is found that the proposed method for detection of budding is effective in terms of learning performance for getting the best quality and quantity of tealeaves harvested.

Author 1: Kohei Arai
Author 2: Yoho Kawaguchi

Keywords: Budding; YOLO; plucking; tealeaves; quality; quantity; hyperparameter; learning performance

PDF

Paper 65: Road Accident Detection using SVM and Learning: A Comparative Study

Abstract: Everyday, a great deal of children and young adults (aged five to 29) lives are lost in road accidents. The most frequent causes are a driver’s behavior, the streets infrastructure is of lower quality and the delayed response of emergency services especially in rural areas. There is a need for automatics road accident systems detection that can assist in recognizing road accidents and determining their positions. This work reviews existing machine learning approaches for road accidents detection. We propose three distinct classifiers: Convolutional Neural Network CNN, Recurrent Convolution Neural Network R-CNN and Support Vector Machine SVM, using a CCTV footage dataset. These models are evaluated based on ROC curve, F1 measure, precision, accuracy and recall, and the achieved accuracies were 92%, 82%, and 93%, respectively. In addition, we suggest using an ensemble learning strategy to maximize the strengths of individual classifiers, raising detection accuracy to 94%.

Author 1: Fatima Qanouni
Author 2: Hakim El Massari
Author 3: Noreddine Gherabi
Author 4: Maria El Badaoui

Keywords: Road accidents; road traffic management; machine learning; SVM; deep learning; ensemble learning

PDF

Paper 66: Recognition of Hate Speech using Advanced Learning Model-based Multi-Layered Approach (MLA)

Abstract: Hate speech becomes more complicated for the users of social media. Some users on online social networking sites (OSNS) create a lot of nonsense by uploading hate speech. OSNS applications developing many models to prevent this hate speech in terms of text and videos. However, these messages still need to be fixed for OSNS users. Sophisticated techniques must automatically identify and detect hate speech material to solve this problem. This paper proposes an advanced learning model-based Multi-Layered Approach (MLA) for hate speech recognition. The proposed model analyses textual data and finds hate speech patterns using multiple deep learning (DL) architectures. The algorithm can generalize well across settings and languages because it was trained on text datasets that include various hate speech types. The final step is an integrated model called Text Convolutional Neural Networks (TCNN), which combines hate text pattern detection with T-Convolutionals. Essential components of the model include the pre-trained model for DistilBERT, integrated pre-processing techniques like Text Cleaning, Lemmatization, and Stemming, and feature extraction techniques like GloVe and Bi-grams (2-grams) to capture contextual information and nuances within language. The model integrates continuous learning techniques to handle the dynamic nature of hate speech. It enables the model to update its comprehension of new language patterns and evolving forms of objectionable content. The evaluation of the proposed model involves benchmarking against existing hate speech detection methods, demonstrating superior precision, recall, and overall accuracy. Finally, the proposed MLA offers a practical and adaptable solution for recognizing hate speech, contributing to creating safer online environments.

Author 1: Puspendu Biswas
Author 2: Donavalli Haritha

Keywords: Multi-Layered Approach (MLA); Deep Learning (DL); DistilBERT; GloVe; Bi-grams (2-grams)

PDF

Paper 67: Method Resource Sharing in On-Premises Environment Based on Cross-Origin Resource Sharing and its Application for Safety-First Constructions

Abstract: The method of resource sharing in an on-premises environment based on Cross-Origin Resource Sharing (CORS) is proposed for security reasons. However, using CORS entails several risks: Cross-Site Request Forgery (CSRF), difficulties in secure configuration, handling credentials, controlling complex requests, and restrictions associated with using wildcards. (1) To mitigate these risks, the following countermeasures are proposed: (2) Use CSRF tokens and the “SameSite” attribute. (3) Minimize preflight requests by allowing only specific origins. (4) Use the “withCredentials” flag or set the “Access-Control-Allow-Credentials” header on the server. (5) Handle custom headers by adding the required headers to CORS settings. (6) Specify a specific origin in the “Access-Control-Allow-Origin” header instead of using wildcards. Additionally, applying CORS for safety-first constructions, which helps raise awareness of dangerous actions in construction fields, is also being explored.

Author 1: Kohei Arai
Author 2: Kodai Norikoshi
Author 3: Mariko Oda

Keywords: Cross-Origin Resource Sharing: CORS; CSRF (Cross-Site Request Forgery); SameSite; withCredentials flag; Access-Control-Allow-Credentials header; safety first constructions

PDF

Paper 68: A Raise of Security Concern in IoT Devices: Measuring IoT Security Through Penetration Testing Framework

Abstract: Despite the widespread adoption of IoT devices across different industries to enhance human activities, there is a pressing need to address the vulnerabilities associated with these devices, as they can potentially give rise to a plethora of cyber threats. Cyberattacks targeting IoT devices are predominantly attributed to inadequate patching and security updates. Furthermore, the current atmosphere pertaining to IoT penetration tests primarily focuses on specific devices and sectors while leaving certain fields behind, such as household devices. This study delves into recent penetration testing on IoT devices. Further, it discusses and critically analyzes the significance and issues in conducting IoT penetration tests. The findings of this study reveal a substantial demand for automated IoT penetration testing to serve diverse industries because conducting such testing has the capacity to diminish the consequences of cyber-attacks across numerous industries that utilize IoT devices for various purposes. This study is intended to be a ready reference for the research community to construct effective and innovative solutions in IoT penetration testing, which covers various fields.

Author 1: Abdul Ghafar Jaafar
Author 2: Saiful Adli Ismail
Author 3: Abdul Habir
Author 4: Khairul Akram Zainol Ariffin
Author 5: Othman Mohd Yusop

Keywords: IoT Security; IoT penetration testing; security assessment; automated penetration testing; penetration testing framework

PDF

Paper 69: Decision Making Systems for Pneumonia Detection using Deep Learning on X-Ray Images

Abstract: This research paper investigates the application of Convolutional Neural Networks (CNNs) for the classification of pneumonia using chest X-ray images. Through rigorous experimentation and data analysis, the study demonstrates the model's impressive learning capabilities, achieving a notable accuracy of 96% in pneumonia classification. The consistent decrease in training and validation losses across 25 learning epochs underscores the model's adaptability and proficiency. However, the research also highlights the challenge of dataset imbalance and the need for improved model interpretability. These findings emphasize the potential of deep learning models in enhancing pneumonia diagnosis but also underscore the importance of addressing existing limitations. The study calls for future research to explore techniques for addressing dataset imbalances, enhance model interpretability, and extend the scope to address nuanced diagnostic challenges within the field of pneumonia classification. Ultimately, this research contributes to the advancement of medical image analysis and the potential for deep learning models to aid in early and accurate pneumonia diagnosis, thereby improving patient care and clinical outcomes.

Author 1: Zhadra Kozhamkulova
Author 2: Elmira Nurlybaeva
Author 3: Madina Suleimenova
Author 4: Dinargul Mukhammejanova
Author 5: Marina Vorogushina
Author 6: Zhanar Bidakhmet
Author 7: Mukhit Maikotov

Keywords: CNN; machine learning; pneumonia; X-ray; image analysis; classification

PDF

Paper 70: Data Security Optimization at Cloud Storage using Confidentiality-based Data Classification

Abstract: Data is the most assets for any organization, stored either in individual systems, server, or cloud platform. Cloud, one of the trending storage systems being adapted now a day is the state-of-the-art of the advanced technology. The major concern with this technological growth is privacy and security of data. Hoisting of data in this platform must be with privacy and security. Hence, there is an urge for service that provides security associated with data to the stake holders. Though the existing security for the data is provided at different levels incurred high cost in terms of processing time. This research aims at providing novel classification-based security algorithm (CBSA) composed with confidential-based classification and encryption with low cost. The confidential-based classification classifies the data into three different levels based on its degree of confidentiality; confidential-based encryption applies a suitable and proportional security mechanism dynamically to each of the levels of data. Thus, the data security process will become optimal and cost effective. The proposed algorithm has outperformed the existing algorithms in terms of processing time and entropy. The processing time and entropy of proposed algorithm has improved by 10%.

Author 1: Dorababu Sudarsa
Author 2: A. Nagaraja Rao
Author 3: A. P. Sivakumar

Keywords: Cloud storage; data privacy; data security; data classification; degree of confidentiality

PDF

Paper 71: Optimal Trajectory Planning for Robotic Arm Based on Improved Dynamic Multi-Population Particle Swarm Optimization Algorithm

Abstract: In response to the problem of easy falling into local optima and low execution efficiency of the basic particle swarm optimization algorithm for 6-degree-of-freedom robots under kinematic constraints, a trajectory planning method based on an improved dynamic multi-population particle swarm optimization algorithm is proposed. According to the average fitness value, the population is divided into three subpopulations. The subpopulation with fitness values higher than the average is classified as the inferior group, while the subpopulation with fitness values lower than the average is classified as the superior group. An equal number of populations are selected from both to form a mixed group. The inferior group is updated using Gaussian mutation and mixed particles, while the superior group is updated using Levy flight and greedy strategies. The mixed group is updated using improved learning factors and inertia weights. Simulation results demonstrate that the improved dynamic multi-population particle swarm optimization algorithm enhances work efficiency and convergence speed, validating the feasibility and effectiveness of the algorithm.

Author 1: Rong Wu
Author 2: Yong Yang
Author 3: Xiaotong Yao
Author 4: Nannan Lu

Keywords: Particle swarm optimization; Gaussian mutation; mixed particles; levy flight; greedy strategy

PDF

Paper 72: Migration Learning and Multi-View Training for Low-Resource Machine Translation

Abstract: This paper discusses the main challenges and solution strategies of low-resource machine translation, and proposes a novel translation method combining migration learning and multi-view training. In a low-resource environment, neural machine translation models are prone to problems such as insufficient generalization performance, inaccurate translation of long sentences, difficulty in processing unregistered words, and inaccurate translation of domain-specific terms due to their heavy reliance on massively parallel corpora. Migration learning gradually adapts to the translation tasks of low-resource languages in the process of fine-tuning by borrowing the general translation knowledge of high-resource languages and utilizing pre-training models such as BERT, XLM-R, and so on. Multi-perspective training, on the other hand, emphasizes the integration of source and target language features from multiple levels, such as word level, syntax and semantics, in order to enhance the model's comprehension and translation ability under limited data conditions. In the experiments, the study designed an experimental scheme containing pre-training model selection, multi-perspective feature construction, and migration learning and multi-perspective fusion, and compared the performance with randomly initialized Transformer model, pre-training-only model, and traditional statistical machine translation model. The experiments demonstrate that the model with multi-view training strategy significantly outperforms the baseline model in evaluation metrics such as BLEU, TER, and ChrF, and exhibits stronger robustness and accuracy in processing complex language structures and domain-specific terminology.

Author 1: Jing Yan
Author 2: Tao Lin
Author 3: Shuai Zhao

Keywords: Low-resource machine translation; migration learning; multi-view training; continual pretraining; multidimensional linguistic feature integration

PDF

Paper 73: Visual Communication Design Based on Sparsity-Enhanced Image Processing Models

Abstract: In the field of visual communication, image clarity and accuracy are the key to convey effective information. A new sparsity-enhanced image processing model is introduced to address the limitations of traditional image processing models in terms of image resolution and fidelity. This model combines a deep neural networks learning framework with a sparse convolutional neural networks enhancement module to complete image reinforcement processing, thereby achieving more accurate image reconstruction techniques. Dictionary learning is used to train models so that the sparse representation of low resolution and high-resolution images has the same dictionary coefficients. By comparing with the existing techniques Enhanced Super-Resolution Generative Adversarial Network, Wide Activation for Efficient and Accurate Image Super-Resolution, and Bicubic Interpolation, and the new model achieves an average peak signal-to-noise ratio of 32.9334 dB, which significantly outperforms the comparison group, respectively, with improvements of 1.9252 dB, 6.6509 dB, and 9.7297 dB, respectively. In addition, the new model demonstrates advantages in structural similarity and learning to perceive image block similarity, implying that it not only enhances the objective quality of the image, but also improves the subjective visual effect of the image. The improved resolution and fidelity of the output image confirms the model's superior performance in processing details and textures. This advancement not only improves the accuracy and efficiency of image processing techniques, but also provides strong technical support for the creation and dissemination of high-quality visual content, which is particularly suitable for application scenarios requiring high-precision visual displays, such as satellite image analysis, remote sensing detection and medical imaging.

Author 1: Zheng Wang
Author 2: Dongsik Hong

Keywords: Deep neural networks; convolutional neural networks; sparsity; dictionary learning; image reinforcement processing

PDF

Paper 74: Enhancing Age Estimation from Handwriting: A Deep Learning Approach with Attention Mechanisms

Abstract: Currently, age estimation is a hot research topic in the field of forensic biology. Age estimation methods based on facial or brain features are easily affected by external factors. In contrast, handwriting analysis is a more reliable method for age estimation. This paper aims to improve the accuracy and efficiency of age prediction using handwriting analysis by proposing a novel method that integrates a coordinate attention mechanism in a deep residual network (CA-ResNet). This method can more accurately capture important features in the input handwritten images while reducing the number of model parameters, thereby improving the accuracy (Acc) and efficiency of the model for age estimation. The proposed method is evaluated on standard handwriting datasets and the created dataset, and it is compared with the current state-of-the-art methods. The results show that the method consistently outperforms others, achieving an accuracy of 79.60% on the IAM handwriting dataset, with a 6.31% improvement over other methods.

Author 1: Li Zhao
Author 2: Xiaoping Wu
Author 3: Xiaoming Chen

Keywords: Age estimation; coordinate attention mechanism; handwriting analysis; accuracy

PDF

Paper 75: Generation of Topical Educational Content by Estimation of the Number of Patents in the Digital Field

Abstract: Analysis of trends in the development of emerging technologies based on patents is a well-recognized approach. An increase in the number of patents precedes the extensive spread of technological solutions and their incorporation in production and professional activities, making it possible to perform predictive analysis. The field of digital technology, which is changing most dynamically among production areas, was chosen as the object of study. The study develops an approach to the analysis of emerging technologies that are related to a given domain. Methods for obtaining quantitative parameters have been developed based on time series representing the number of patents per year. The concept of a parameter plane has been introduced. It includes the parameters of stable growth/decline and annual quantity of patents. A special feature of the approach is the calculation of parameters for the last observed segment of the stable dynamic behavior of the time series based on the developed algorithm. The work takes the Digital Marketing domain as an example and presents analysis of 296 keywords related to this concept. Based on time series constructed from the patent database for 2000-2021, the most promising technologies were identified. The application of the results for the generation of topical educational content in the Digital Marketing field is considered.

Author 1: Evgeny Nikulchev
Author 2: Dmitry Ilin

Keywords: Time series; patent analysis; parameter plane; predictive analysis

PDF

Paper 76: Enhancing Smart Contract Security Through Multi-Agent Deep Reinforcement Learning Fuzzing: A Survey of Approaches and Techniques

Abstract: Multi-Agent Systems (MAS) and Deep Reinforcement Learning (DRL) have emerged as powerful tools for enhancing security measures, particularly in the context of smart contract security in blockchain technology. This literature review explores the integration of Multi-Agent DRL fuzzing techniques to bolster the security of smart contracts. The study delves into the formalization of emergence in MAS, the comprehensive survey of multi-agent reinforcement learning, and progress on the state explosion problem in model checking. By addressing challenges such as state space explosion, real-time detection, and adaptability across blockchain platforms, researchers aim to advance the field of smart contract security. The review emphasizes the significance of Multi-Agent DRL fuzzing in improving security testing processes and calls for future research and collaboration to enhance the resilience and integrity of decentralized applications. Through advancements in algorithmic efficiency, the incorporation of Explainable AI, cross-domain applications of MAS, and cooperation with blockchain development teams, the future of smart contract security holds promise for robust and secure blockchain ecosystems.

Author 1: Muhammad Farman Andrijasa
Author 2: Saiful Adli Ismail
Author 3: Norulhusna Ahmad
Author 4: Othman Mohd Yusop

Keywords: Smart contract security; multi-agent systems; deep reinforcement learning; fuzzing techniques; blockchain technology

PDF

Paper 77: Exhaustive Insights Towards Social-Media Driven Disaster Management Approaches

Abstract: The manuscript presents discussion about the disaster management approaches using social media. It is noted that rising popularity of social media has been witnessed to significantly contribute towards information propagation and community participation to deal with the event of disaster. Different from conventional disaster management policies, the scope of inclusion of social media-based approaches are quite novel and yet promising. However, the problem is towards unclear information about the effectivity of such schemes. Hence, this manuscript contributes towards bridging this information gap by carrying out an exhaustive and systematic review of existing methodology frequently adopted towards disaster management using social media viz. early warning methods, information dissemination methods, crisis mapping method, and predictive approach, where Artificial Intelligence was noted to be quite dominant scheme. The contributory findings of this review study contribute towards clear visualization of updated research trends, critical learning outcomes associated with identified research gap with illustrated discussion of the reviewed articles. A clear and informative study findings contributes towards future researchers. The result of review has also answered the formed research question to give potential insight towards existing system. The result of the review finds that existing approaches has both beneficial aspect and limitation associated with complex learning approaches, higher infrastructural cost, model complexities, security threats, higher resource dependencies.

Author 1: Nethravathy Krishnappa
Author 2: D Saraswathi
Author 3: Chandrasekar Chelliah

Keywords: Artificial intelligence; disaster management; information propagation; social media; community

PDF

Paper 78: Network Security Evaluation Based on Improved Genetic Algorithm and Weighted Error Backpropagation Algorithm

Abstract: As the speed advancement of network technology and the popularization of applications, network security problems are becoming more and more prominent, all kinds of network attacks and security threats are increasing, and the demand for network security evaluation is becoming more and more urgent. To address the issues of long time-consuming and low accuracy in the traditional network security evaluation model, the study proposes a network security evaluation model based on improved genetic algorithm and weighted error BP algorithm. The study first combines the weighted error BP algorithm with the improved genetic algorithm for data analysis and research, and then integrates the two to construct a network security evaluation model. The results show that in the detection of network security vulnerabilities, the evaluation model of the data processing vulnerability detection accuracy, risk detection rate of 93.28%, 91.88%, respectively. The function training error of the model is 8.93% respectively, while the decoding accuracy and stability are 90.43% and 92.07% respectively, which are better than the comparison method. This indicates that the method has high accuracy and robustness in network security evaluation, and can provide network administrators and users with a more scientific and reliable basis for decision-making.

Author 1: Jinlong Pang
Author 2: Chongwei Liu

Keywords: Genetic algorithm; return propagation algorithm; cybersecurity evaluation; weighting; network vulnerability

PDF

Paper 79: Application Analysis of Network Security Situational Awareness Model for Asset Information Protection

Abstract: The popularity of the Internet makes the network develop rapidly. However, the network security threat is more complex and hidden. The traditional network security alarm system has the problems of low accuracy and low efficiency when dealing with huge redundant data. Therefore, the research comprehensively considers the network security problems, proposes a network security situational awareness model for asset information protection combined with knowledge graph, establishes an asset-based network security knowledge graph, utilizes attribute graphs to complete the network attack scenario discovery and network situational understanding, and verifies the effectiveness and superiority of the model. The experimental results show that the research-proposed model detects an average of 9706 attacks out of 10000 attacks. For 100 high-risk level attacks, the number of detections is higher than 98. The average correctness, recall, and false alarm rates of the research proposed model are 99.48%, 99.04%, and 0.86%, respectively. In addition, when the model is running, its maximum memory usage is only 22.67%, and the time to complete the attack detection at the same time is 258.4s, both of which are much lower than the comparison algorithms. Finally, the research-proposed model is able to effectively reflect the impact of attack events on the posture of asset nodes. The proposed cybersecurity situational awareness model is of great theoretical and practical significance for improving organizational cybersecurity, innovating cybersecurity solutions, and maintaining the security of asset information in the digital era.

Author 1: Yuemei Ren
Author 2: Xianju Feng

Keywords: Asset information protection; cyber security; situational awareness; knowledge graph; attack scenarios

PDF

Paper 80: Big Data Multi-Strategy Predator Algorithm for Passenger Flow Prediction

Abstract: Faced with the rapidly recovering tourism market, accurate prediction of passenger flow can help local authorities achieve more effective resource regulation. Therefore, based on big data technology, a multi-strategy predator algorithm is proposed, which uses the Marine Predator Algorithm, combined with regularized extreme learning machines and Collaborative Filtering Algorithms, to achieve accurate passenger flow prediction. The experiment findings denote that the performance parameters of the algorithm are excellent, with extremely strong convergence performance, and only 30 iterations are needed to reach the optimal solution. The fitting degree of this algorithm is 97.8%, which is 6.27% -19.31% higher than that of long and short-term memory networks, random forest algorithms, and support vector machine regression. In actual passenger flow prediction, the error rate of this algorithm is only 2.29%, which is 3.47% -6.50% higher than the three comparison algorithms. This study provides a new and efficient prediction method for passenger flow prediction. Its excellent predictive performance can not only help relevant departments predict and manage passenger traffic more accurately, but also provide reference for traffic prediction in other fields. Overall, this study has important reference value and practical significance for the research and practice of passenger flow prediction.

Author 1: Peng Guo

Keywords: Passemger Flow Prediction; regularized extreme learning machine; Collaborative Filtering Algorithm; Marine Predator Algorithm

PDF

Paper 81: Computer Simulation Study of Stiffness Variation of Stewart Platform under Different Loads

Abstract: The ability of Stewart platform to resist deformation is an important target for designing and optimizing the platform, and studying the variation rule of stiffness of Stewart platform under different loads can help us to understand the dynamic characteristics of the platform, guide the design and control of the platform, and improve the performance and stability of the platform. The purpose of this paper is to change the law of stiffness variation and influence factors of Stewart platform under different loads, aiming to study the change of stiffness of Stewart platform under different loads as well as the influence factors, and the influence of stiffness change on the performance and stability of the platform. Firstly, using MATLAB software, the kinematic and mechanical model of Stewart platform was established, the analytical expression of the stiffness matrix of the platform was deduced, and the stiffness characteristics and stiffness singularity of the platform were analyzed. Then, using ADAMS software, the dynamic simulation model of the Stewart platform was established, and the stiffness of the platform was simulated and analyzed. The results show that the stiffness of the Stewart platform will appear singularity or sudden change under some special positions or loads, which should be avoided as much as possible so as not to affect the performance and stability of the platform. There is a certain correlation between the dynamic and static stiffness, but it is also affected by the nonlinearity of the structure, damping, coupling and other factors.

Author 1: Zhiqiang Zhao
Author 2: Yuetao Liu
Author 3: Changsong Yu
Author 4: Peicen Jiang

Keywords: Stewart; different loads; stiffness variation; computer simulation

PDF

Paper 82: Transforming Pixels: Crafting a 3D Integer Discrete Cosine Transform for Advanced Image Compression

Abstract: We propose an innovative technique for image compression based on the 3-dimensional Integer Discrete Cosine Transform (3D-Integer DCT), which will serve as an alternative to the existing DCT-based compression technique. If an image is encoded as cubes [row × column × temporal length] instead of blocks [row × column], higher compression can be achieved. Here, the number of blocks is represented as the temporal length. To construct cubes, we use highly correlated blocks, and the correlation level is determined using the mean absolute difference (MAD). The suggested 3D-Integer DCT-based coder can achieve a higher compression ratio while maintaining the required image quality. It also needs fewer coefficients to encode an image than the usual Joint Photographic Expert Group (JPEG) coder. Adopting integer DCT further reduces the computational complexity of the proposed algorithm, given the abundance of methods available in the literature to determine equivalent integers for DCT. We choose an optimum integer group that minimizes mean squared error (MSE) and improves coding efficiency for computing 3D-Integer DCT. We also conducted a detailed analysis to examine the impact of implementing integer DCT in image compression. When we look at peak signal-to-noise noise ratio (PSNR), bits per pixel, and structural similarity index (SSIM), we see that the proposed algorithm does a better job than the standard real-value DCT-based compression algorithm like JPEG.

Author 1: R. Rajprabu
Author 2: T. Prathiba
Author 3: Deepa Priya V.
Author 4: Arthy Rajkumar
Author 5: Rajkannan. C
Author 6: P. Ramalakshmi

Keywords: Discrete cosine transform; 3D integer DCT; Image compression; JPEG algorithm

PDF

Paper 83: Real-Time Road Lane-Lines Detection using Mask-RCNN Approach

Abstract: This paper presents a novel approach to real-time road lane-line detection using the Mask R-CNN framework, with the aim of enhancing the safety and efficiency of autonomous driving systems. Through extensive experimentation and analysis, the proposed system demonstrates robust performance in accurately detecting and segmenting lane boundaries under diverse driving conditions. Leveraging deep learning techniques, the system exhibits a high level of accuracy in handling complex scenarios, including variations in lighting conditions and occlusions. Real-time processing capabilities enable instantaneous feedback, contributing to improved driving safety and efficiency. However, challenges such as model generalizability, interpretability, computational efficiency, and resilience to adverse weather conditions remain to be addressed. Future research directions include optimizing the system's performance across different geographic regions and road types and enhancing its adaptability to adverse weather conditions. The findings presented in this paper contribute to the ongoing efforts to advance autonomous driving technology, with implications for improving road safety and transportation efficiency in real-world settings. The proposed system holds promise for practical deployment in autonomous vehicles, paving the way for safer and more efficient transportation systems in the future.

Author 1: Gulbakhram Beissenova
Author 2: Dinara Ussipbekova
Author 3: Firuza Sultanova
Author 4: Karasheva Nurzhamal
Author 5: Gulmira Baenova
Author 6: Marzhan Suimenova
Author 7: Kamar Rzayeva
Author 8: Zhanar Azhibekova
Author 9: Aizhan Ydyrys

Keywords: Lane lines; detection; classification; segmentation; Mask-RCNN; deep learning

PDF

Paper 84: Hybrid Convolutional Recurrent Neural Network for Cyberbullying Detection on Textual Data

Abstract: With the burgeoning use of social media platforms, online harassment and cyberbullying have become significant concerns. Traditional mechanisms often falter, necessitating advanced methodologies for efficient detection. This study presents an innovative approach to identifying cyberbullying incidents on social media sites, employing a hybrid neural network architecture that amalgamates Long Short-Term Memory (LSTM) and Convolutional Neural Network (CNN). By harnessing the sequential processing capabilities of LSTM to analyze the temporal progression of textual data, and the spatial discernment of CNN to pinpoint bullying keywords and patterns, the model demonstrates substantial improvement in detection accuracy compared to extant methods. A diverse dataset, encompassing multiple social media platforms and linguistic styles, was utilized to train and test the model, ensuring robustness. Results evince that the LSTM-CNN amalgamation can adeptly handle varied sentence structures and contextual nuances, outstripping traditional machine learning classifiers in both specificity and sensitivity. This research underscores the potential of hybrid neural networks in addressing contemporary digital challenges, urging further exploration into blended architectures for nuanced problem-solving in cyber realms.

Author 1: Altynzer Baiganova
Author 2: Saniya Toxanova
Author 3: Meruert Yerekesheva
Author 4: Nurshat Nauryzova
Author 5: Zhanar Zhumagalieva
Author 6: Aigerim Tulendi

Keywords: CNN; RNN; LSTM; urban sounds; impulsive sounds

PDF

Paper 85: Studying the Behavior of a Modified Deep Learning Model for Disease Detection Through X-ray Chest Images

Abstract: In modern medical diagnostics, Deep Learning models are commonly used for illness diagnosis, especially over X-ray chest images. Deep Learning approaches provide unmatched promise for early identification, prognosis, and treatment evaluation across a range of illnesses, by combining sophisticated algorithms with large datasets. It is crucial to research these models to lead to improved ones to progress toward disease identification's precision, effectiveness, and scalability. This paper presents the study of a CNN+VGG19 Deep Learning architecture (subsets of machine learning), both before and after its modification. The same dataset is used over the existing and modified models to compare metrics under the same conditions. They are compared using metrics like loss, accuracy, precision, sensitivity, and AUC. These metrics display lower values in the updated model than in the original one. The numbers demonstrate the occurrence of the overfitting phenomenon, which is most likely the result of the model's increased complexity for a small dataset. The noise in the images included in the dataset may also be the cause. As a result, it can be stated that regularization techniques should be applied; otherwise, layers of extraction and classification should not be added to the model to prevent overfitting.

Author 1: Elma Zanaj
Author 2: Lorena Balliu
Author 3: Gledis Basha
Author 4: Elunada Gjata
Author 5: Elinda Kajo Meçe

Keywords: Machine learning; big data; X-ray chest image; CNN; VGG19

PDF

Paper 86: Establishment of Economic Analysis Model Based on Artificial Intelligence Technology

Abstract: With the continuous evolution of artificial intelligence technology, its integration into economic analysis models is becoming increasingly prevalent. This paper employs the Lasso Back Propagation neural network method to conduct financial analysis and prediction for major global economies, focusing on total Gross Domestic Product, combined Gross Domestic Product growth rate, and Consumer Price Index. The real Gross Domestic Product of the top 30 countries in the global ranking is meticulously analyzed and categorized into various economic types. This categorization, coupled with the utilization of neural network multi-hidden layer variable analysis, facilitates the analysis and prediction of national economic trends. The findings reveal that overall economic growth among the top 30 countries is sluggish, albeit showing a growth trajectory. However, the driving force for economic growth remains notably inadequate. Moreover, employing a single time series model effectively predicts Gross Domestic Product and Consumer Price Index growth rates, alongside other macroeconomic indicators. Notably, the absence of autocorrelation in the fitting residual series underscores the applicability of the time series method for combined forecasting, affirming the robustness of the predictive framework.

Author 1: Jiqing Shi

Keywords: Artificial intelligence; lasso regression; BP neural network; economic analysis model; major global economies; multi-hidden layer variable analysis; economic trends

PDF

Paper 87: Deep Learning Approach to Classify Brain Tumors from Magnetic Resonance Imaging Images

Abstract: Brain tumor is one of the primary causes of mortality all over the globe, and it poses as one of the most complicated tasks in contemporary medicine when it comes to its proper diagnosis and classification into its many different types. Both benign and malignant tumors affect the lives of their respective patients as they may lead to mortality, or in the least many related difficulties and sicknesses. Typically, MRI (Magnetic Resonance Imaging) is used as a diagnostic technique where experts manually analyze the images to detect tumors. On the other hand, advanced technologies such as deep learning can step into the light and aid in the diagnosis and classification procedures in a much more time-efficient and precise manner. MRI images are an effective input that can be used in deep learning technologies such as CNN in order to accurately detect brain tumors. In this study, VGG-16, ResNet50, and Xception were trained on a Kaggle dataset consisting of brain tumor MRI images. The performance of the models was evaluated where it was found that brain tumors can be efficiently detected from MRI images with high accuracy and precision using VGG-16, ResNet50, and Xception. The highest performing model was the proposed XCeption model with perfect scores.

Author 1: Asma Ahmed A. Mohammed

Keywords: Deep learning; brain tumor; MRI images; Convolutional Neural Networks (CNN); Xception; VGG-16; ResNet50

PDF

Paper 88: Multi-Objective Optimization of Oilfield Development Planning Based on Shuffled Frog Leaping Algorithm

Abstract: Oilfield development planning is a complex task that involves multiple optimization objectives and constraints. Therefore, a study proposes an improved shuffled frog leaping algorithm to achieve multi-objective optimization tasks. In multi-objective problems, the fitness value of the algorithm is not adaptive to the memetic evolution, resulting in local search failures. Research is conducted on improving the shuffled frog leaping algorithm through non-dominated sorting genetic algorithm-II, memetic evolution, and traversal methods, and then verifying the effectiveness of the algorithm. The outcomes denoted that when the population was 30 and the grouping was 5, the algorithm proposed in the study had the fastest search speed and better optimization effect. The improved shuffled frog leaping algorithm had advantages in both construction period and cost compared to the shuffled frog leaping algorithm, with a construction period difference of 19 days and a cost difference of $13871. In comparative experiments with other algorithms, the average optimal solution and running time of the proposed algorithm were 0.324 and 7.2 seconds, respectively, which can quickly find the optimal solution in a short time. The algorithm proposed in the study can effectively optimize the complex objectives and constraints in oilfield development planning problems.

Author 1: Jun Wei

Keywords: Shuffled frog leaping algorithm; oilfield development; multi-objective; optimization; improve

PDF

Paper 89: Investigating an Ensemble Classifier Based on Multi-Objective Genetic Algorithm for Machine Learning Applications

Abstract: Ensemble learning in machine learning applications is crucial because it leverages the collective wisdom of multiple models to enhance predictive performance and generalization. Ensemble learning is a method to provide a better approximation of an optimal classifier. A number of basic classifiers are used in ensemble learning. In order to improve performance, it is important for the basic classifiers to possess adequate efficacy and exhibit distinct classification errors. Additionally, an appropriate technique should be employed to amalgamate the outcomes of these classifiers. Numerous methods for ensemble classification have been introduced, including voting, bagging and reinforcement methods. In this particular study, an ensemble classifier that relies on the weighted mean of the basic classifiers' outputs was proposed. To estimate the combination weights, a multi-objective genetic algorithm, considering factors such as classification error, diversity, sparsity, and density criteria, was utilized. Through implementations on UCI datasets, the proposed approach demonstrates a significant enhancement in classification accuracy compared to other conventional ensemble classifiers. In summary, the obtained results showed that genetic-based ensemble classifiers provide advantages such as enhanced capability to handle complex datasets, improved robustness and generalization, and flexible adaptability. These advantages make them a valuable tool in various domains, contributing to more accurate and reliable predictions. Future studies should test and validate this method on more and larger datasets to determine its actual performance.

Author 1: Zhiyuan LIU

Keywords: Machine learning; genetic algorithm; ensemble classification; classification error

PDF

Paper 90: Automatic Detection of Ascaris Lumbricoides in Microscopic Images using Convolutional Neural Networks (CNN)

Abstract: Parasites are disease-causing agents both in Peru and worldwide. In many contexts, diagnosis is done manually by observing microscopic images, where it's necessary to identify parasite eggs. However, this process is notably slow, and sometimes image clarity may be insufficient, making rapid and accurate identification challenging. This can be due to various factors, such as image quality or the presence of noise. This paper focused on a Convolutional Neural Network (CNN) model. Through this approach, the training, testing, and validation stages of our CNN model to detect and identify Ascaris lumbricoides parasite eggs. The results show that the proposed CNN model, combined with image preprocessing, yielded highly favorable results in parasite egg identification. Additionally, very satisfactory values were achieved in model testing and validation, indicating its effectiveness and precision in diagnosing parasite presence. This research represents a significant advancement in the field of parasitological diagnosis, offering an efficient and accurate solution for parasite detection through microscopic image analysis. It is hoped that these results contribute to improving diagnosis and treatment methods for parasitic diseases.

Author 1: Giovanni Gelber Martinez Pastor
Author 2: Cesar Roberto Ancco Ruelas
Author 3: Eveling Castro-Gutierrez
Author 4: Victor Luis Vásquez Huerta

Keywords: Ascaris lumbricoides; Convolutional Neural Networks; OpenCV; microscopic images; moment-based detection

PDF

Paper 91: Automatic Personality Recognition in Videos using Dynamic Networks and Rank Loss

Abstract: There are a few difficulties with current automatic personality recognition technologies. Two of these are discussed in this article. They use of very brief video segments or individual frames to come to conclusion with personality factors rather than long-term behavior; and absence of techniques to record individuals' facial movements for personality recognition. To address these concerns, this work first offers a unique Rank Loss for self-regulated learning of facial movements that uses the innate time related development of facial movements in lieu of personality traits. Our method begins by training a basic U-net type system that can predict broad facial movements from a collection of unlabeled face recordings. The robust model is frozen subsequently, and a series of intermediary filters is added to the architecture. The self-regulated education is then restarted, but only with films tailored to the individual. As a result, the weights of the learnt filters are individual-specific, making it a useful tool for simulating individual facial dynamics. The weights of the learnt filters are then concatenated as an individual-specific representation, to forecast personality factors without the assistance of other components of the network. The proposed strategy is tested on ChaLearn personality dataset. We infer that the tasks performed by the individual in the video matter, merging or combined application of tasks achieves the high-rise precision. Also, multi-scale characteristics are better penetrating than single-scale dynamics, along with achieving impressive outcomes as process innovation in prediction of the personality factors scores through videos.

Author 1: Nethravathi Periyapatna Sathyanarayana
Author 2: Karuna Pandith
Author 3: Manjula Sanjay Koti
Author 4: Rajermani Thinakaran

Keywords: Automatic personality recognition; facial movements; individual-specific representation; personality factors; convolutional neural networks

PDF

Paper 92: Contrastive Learning and Multi-Choice Negative Sampling Recommendation

Abstract: Most existing recommendation models that directly model user interests on user-item interaction data usually ignore the natural noise present in the interaction data, leading to bias in the model's learning of user preferences during data propagation and aggregation. In addition, the currently adopted negative sampling strategy does not consider the relationship between the prediction scores of positive samples and the degree of difficulty of negative samples, and is unable to adaptively select a suitable negative sample for each positive sample, leading to a decrease in the model recommendation performance. In order to solve the above problems, this paper proposes a Contrastive Learning and Multi-choice Negative Sampling Recommendation. Firstly, an improved topology-aware pruning strategy is used to process the user-item bipartite graph, which uses the topology information of the graph to remove noise and improve the accuracy of model prediction. In addition, a new multivariate selective negative sampling module is designed, which ensures that each positive sample selects a negative sample of appropriate hardness through two sampling principles, improving the model embedding space representation capability, which in turn leads to improved model recommendation accuracy. Experimental results on the Urban-Book and Yelp2018 datasets show that the proposed algorithm significantly improves all the metrics compared to the state-of-the-art model, which proves the effectiveness and sophistication of the algorithm in different scenarios.

Author 1: Yun Xue
Author 2: Xiaodong Cai
Author 3: Sheng Fang
Author 4: Li Zhou

Keywords: Recommendation algorithms; comparative learning; negative sampling; pruning strategies

PDF

Paper 93: Development of Deep Learning Models for Traffic Sign Recognition in Autonomous Vehicles

Abstract: This research paper investigates the development of deep learning models for traffic sign recognition in autonomous vehicles. Leveraging convolutional neural networks (CNNs), the study explores various architectural configurations and evaluation methodologies to assess the efficacy of CNNs in accurately identifying and classifying traffic signs. Through a systematic evaluation process utilizing metrics such as accuracy, precision, recall, and F-score, the research demonstrates the robustness and generalization capability of the developed models across diverse environmental conditions. Furthermore, the utilization of visualization techniques, including the Matplotlib library, enhances the interpretability of model training dynamics and optimization progress. The findings highlight the significance of CNN architecture in facilitating hierarchical feature extraction and spatial dependency learning, thereby enabling reliable and efficient traffic sign recognition. The successful recognition of traffic signs under varying lighting conditions underscores the resilience of the developed models to environmental perturbations. Overall, this research contributes to advancing the capabilities of autonomous vehicle systems and lays the groundwork for the implementation of intelligent traffic sign recognition systems aimed at enhancing road safety and navigational efficiency.

Author 1: Zhadra Kozhamkulova
Author 2: Zhanar Bidakhmet
Author 3: Marina Vorogushina
Author 4: Zhuldyz Tashenova
Author 5: Bella Tussupova
Author 6: Elmira Nurlybaeva
Author 7: Dastan Kambarov

Keywords: Traffic sign recognition; machine learning; deep learning; computer vision; image classification

PDF

Paper 94: Enhanced U-Net Architecture for Lung Segmentation on Computed Tomography and X-Ray Images

Abstract: In the expanding field of medical imaging, precise segmentation of anatomical structures is critical for accurate diagnosis and therapeutic interventions. This research paper introduces an innovative approach, building upon the established U-Net architecture, to enhance lung segmentation techniques applied to Computed Tomography (CT) images. Traditional methods of lung segmentation in CT scans often confront challenges such as heterogeneous tissue densities, variability in human anatomy, and pathological alterations, necessitating an approach that embodies greater robustness and precision. Our study presents a modified U-Net model, characterized by an integration of advanced convolutional layers and innovative skip connections, improving the reception field and facilitating the retention of high-frequency details essential for capturing the lung's intricate structures. The enhanced U-Net architecture demonstrates substantial improvements in dealing with the subtleties of lung parenchyma, effectively distinguishing between precarious nuances of tissues, and pathologies. Rigorous quantitative evaluations showcase a significant increase in the Dice coefficient and a decrease in the Hausdorff distance, indicating a more refined segmentation output compared to predecessor models. Additionally, the proposed model manifests exceptional versatility and computational efficiency, making it conducive for real-time clinical applications. This research underlines the transformative potential of employing advanced deep learning architectures for biomedical imaging, paving the way for early intervention, accurate diagnosis, and personalized treatment paradigms in pulmonary disorders. The findings have profound implications, propelling forward the nexus of artificial intelligence and healthcare towards unprecedented horizons.

Author 1: Gulnara Saimassay
Author 2: Mels Begenov
Author 3: Ualikhan Sadyk
Author 4: Rashid Baimukashev
Author 5: Askhat Maratov
Author 6: Batyrkhan Omarov

Keywords: Lung disease; deep learning; U-Net; computed tomography; segmentation; diagnosis

PDF

Paper 95: EfficientSkinCaSV2B3: An Efficient Framework Towards Improving Skin Classification and Segmentation

Abstract: Ozone layer depletion has gained attention as a serious environmental issue. Because of its effects on human health especially skin cancer. Besides, Ultraviolet (UV) radiation is known to be a major risk factor for skin cancer. For instance, it can damage the DNA in skin cells leading to mutations that may eventually result in cancerous growth. Basal cell carcinoma, squamous cell carcinoma, and melanoma are the three primary forms of skin cancer linked to UV exposure. Additionally, it triggers associated illnesses including nevus, seborrheic keratosis, actinic keratosis, dermatofibroma, and vascular lesions. Many medical and computer studies were published as a result to address these disorders. Especially, using an aspect of deep learning that is transfer learning and fine-tuning for the classification of skin images. In this research, the EffecientSkinCaSV2B3 framework was proposed and applied to classify and segment the skin cancer dataset, which were collected and validated by The International Skin Imaging Collaboration (ISIC). In addition, Gradient-weighted Class Activation Mapping (Grad-CAM) is used in skin cancer classification to visually explain images, aiding in understanding model decisions and highlighting important areas. Based on color and texture, k-means clustering was used for the segmentation between portions that were healthy and those that were unhealthy. The study reached a surprising accuracy of 84.91% in nine classes of classifying skin cancer. In other experiments, the customized EfficientNetV2B3 model achieved 94.00% in classifying malign and benign. Moreover, scenarios pointed out that in classifying six classes (i.e., between benign skin diseases) and three classes (i.e., between malign skin diseases) the model earned a high accuracy of 89.56% and 96.74%, respectively.

Author 1: Quy Lu Thanh
Author 2: Triet Minh Nguyen

Keywords: Skin cancer; Convolutional Neural Network (CNN); transfer learning; fine tuning; classification; segmentation; EffecientNetB3V2

PDF

Paper 96: Cross-Modal Fine-Grained Interaction Fusion in Fake News Detection

Abstract: The popularity of social media has significantly increased the speed and scope of news dissemination, making the emergence and spread of fake news easier. Current fake news detection methods often ignore the correlation between text and images, leading to insufficient modal interaction and fusion. To address these issues, a cross-modal fine-grained interaction and fusion model for fake news detection is proposed. Specifically, this study addresses the correlation problem between text and image modalities by designing an interaction similarity domain. It extracts features of text word weight distribution using an attention mechanism network, guides the features of different regions of the image, and calculates the local similarity between the two. This approach analyzes positive and negative correlations between modalities at a fine-grained level, thereby strengthening the intermodal connection. Additionally, to tackle the problem of insufficient fusion of semantic feature vectors between text and images, this paper designs a fusion network that employs improved encoding and decoding using a Transformer for inter-modal information fusion, achieving the final multimodal feature representation. Experimental results show that our proposed method achieves excellent performance on WeiboA and Twitter, with accuracies of 88.2% and 89%, respectively, outperforming the benchmark model in several evaluation metrics.

Author 1: Zhanbin Che
Author 2: GuangBo Cui

Keywords: Fake news detection; attention mechanism; multimodal feature fusion; local similarity

PDF

Paper 97: A Stepwise Discriminant Analysis and FBCSP Feature Selection Strategy for EEG MI Recognition

Abstract: Accurate decoding of brain intentions is a pivotal technology within Brain-Computer Interface (BCI) systems that rely on Motor Imagery (MI). The effective extraction of information features plays a critical role in the precise decoding of these brain intentions. However, there exists significant individual and environmental variability in signals, and the sensitivity of EEG signals from different subjects also varies, imposing higher demands on both feature exploration and accurate decoding. To address these challenges, we employ adaptive sliding time windows and a stepwise discriminant analysis strategy to selectively extract features obtained through the Filter Bank Common Spatial Pattern (FBCSP). This entails the identification of an optimal feature combination tailored to specific patients, thereby mitigating individual differences and environmental variations. Initially, adaptive sliding time windows are applied to segment electroencephalogram (EEG) data for different subjects, followed by FBCSP for feature extraction. Subsequently, a stepwise discriminant analysis (SDA) incorporating prior knowledge is employed for optimal feature selection, effectively and adaptively identifying the best feature combination for specific subjects. The proposed method is evaluated using two publicly available datasets, the EEG recognition accuracy for Dataset A is 98.47%, and for Dataset B, it is 95.2%. In comparison to current publicly reported research results (utilizing Power Spectral Density (PSD) + Support Vector Machine (SVM) methods) for Dataset A, the proposed method improves MI recognition accuracy by 25.37%. For Dataset B, compared to current publicly reported results (FBCNet method), the proposed method improves MI recognition accuracy by 26.4%. The experimental results underscore the method's broad applicability, scalability, and substantial value for promotion and application.

Author 1: YingHui Meng
Author 2: YaRu Su
Author 3: Duan Li
Author 4: JiaoFen Nan
Author 5: YongQuan Xia

Keywords: Stepwise discriminant analysis; electroenceph alogram; motor imagery; sliding time window; filter bank common spatial pattern

PDF

Paper 98: Enhancing Sentiment Analysis on Social Media Data with Advanced Deep Learning Techniques

Abstract: This paper introduces a comprehensive methodology for conducting sentiment analysis on social media using advanced deep learning techniques to address the unique challenges of this domain. As digital platforms play an increasingly pivotal role in shaping public discourse, the demand for real-time sentiment analysis has expanded across various sectors, including policymaking, brand monitoring, and personalized services. Our study details a robust framework that encompasses every phase of the deep learning process, from data collection and preprocessing to feature extraction and model optimization. We implement sophisticated data preprocessing techniques to improve data quality and adopt innovative feature extraction methods such as TF-IDF, Word2Vec, and GloVe. Our approach integrates several advanced deep learning configurations, including variants of BiLSTMs, and employs tools like Scikit-learn and Gensim for efficient hyperparameter tuning and model optimization. Through meticulous optimization with GridSearchCV, we enhance the robustness and generalizability of our models. We conduct extensive experimental analysis to evaluate these models against multiple configurations using standard metrics to identify the most effective techniques. Additionally, we benchmark our methods against prior studies, and our findings demonstrate that our proposed approaches outperform comparative techniques. These results provide valuable insights for implementing deep learning in sentiment analysis and contribute to setting benchmarks in the field, thus advancing both the theoretical and practical applications of sentiment analysis in real-world scenarios.

Author 1: Huu-Hoa Nguyen

Keywords: Sentiment analysis; deep learning; hyperparameter; feature extraction; social media; digital platform; gridsearchcv; BiLSTM; TF-IDF; word2vec; glove; Scikit-learn and Gensim

PDF

Paper 99: An Integrated Generalized Linear Regression with Two Step-AS Algorithm for COVID-19 Detection

Abstract: This research introduces a computer-aided intelligence model designed to automatically identify positive instances of COVID-19 for routine medical applications. The model, built on the Generalized Linear architecture, employs the TwoStep-AS cluster method with diverse screen relatives, Weight sharing and stripping characteristics automatically identify distinctive features in chest X-ray images. Unlike the conventional transformational learning approach, our model underwent training both before and after clustering. The dataset was subjected to a compilation process that involved subdividing samples and categories into multiple sub-samples and subgroups. New cluster labels were then assigned to each cluster, treating each subject cluster as a distinct category. Discriminant features extracted from this process were used to train the Generalized Linear model, which was subsequently applied to classify instances. The TwoStep-AS clustering method underwent modification using pre-compiling the data earlier then employing the Generalized Linear model to identify COVID samples from X-ray chest results. Tests were conducted by the COVID-radiology data guaranteed the correctness of the results. The suggested model demonstrated an impressive accuracy of 90.6%, establishing it as a highly efficient, cost-effective, and rapid intelligence tool for the detection of Coronavirus infections.

Author 1: Ahmed Hamza Osman
Author 2: Hani Moetque Aljahdali
Author 3: Sultan Menwer Altarrazi
Author 4: Altyeb Taha

Keywords: Generalized Linear model; COVID-19; TwoStep-AS; clustering; X-ray images

PDF

Paper 100: Local Path Planning of Mobile Robots Based on the Improved SAC Algorithm

Abstract: This paper proposes a new EP-PER-SAC algorithm to solve the problems of slow training speed and low learning efficiency of the SAC (Soft Actor Critic) algorithm in the local path planning of mobile robots by introducing the Priority Experience Replay (PER) strategy and Experience Pool (EP) adjustment technique. This algorithm replaces equal probability random sampling with sampling based on the priority experience to increase the frequency of extracting important samples, thereby improves the stability and convergence speed of model training. On this basis, it requires to continuously monitor the learning progress and exploration rate changes of the robot to dynamically adjust the experience pool, so the robot can adapt effectively to the environment changes and the storage requirements and learning efficiency of the algorithm are balanced. Then, the algorithm's reward and punishment function is improved to reduce the blindness of algorithm training. Finally, experiments are conducted under different obstacle environments to verify the feasibility of the algorithm based on ROS (Robot Operating System) simulation platform and real environment. The results show that the improved EP-PER-SAC algorithm has a shorter path length and faster model convergence speed than the original SAC algorithm and PER-SAC algorithm.

Author 1: Ruihong Zhou
Author 2: Caihong Li
Author 3: Guosheng Zhang
Author 4: Yaoyu Zhang
Author 5: Jiajun Liu

Keywords: Mobile robots; local path planning; reinforcement learning; SAC algorithm; priority experience replay; experience pool adjustment; Robot Operating System (ROS)

PDF

Paper 101: Exploring Music Style Transfer and Innovative Composition using Deep Learning Algorithms

Abstract: Automatic music generation represents a challenging task within the field of artificial intelligence, aiming to harness machine learning techniques to compose music that is appreciable by humans. In this context, we introduce a text-based music data representation method that bridges the gap for the application of large text-generation models in music creation. Addressing the characteristics of music such as smaller note dimensionality and longer length, we employed a deep generative adversarial network model based on music measures (MT-CHSE-GAN). This model integrates paragraph text generation methods, improves the quality and efficiency of music melody generation through measure-wise processing and channel attention mechanisms. The MT-CHSE-GAN model provides a novel framework for music data processing and generation, offering an effective solution to the problem of long-sequence music generation. To comprehensively evaluate the quality of the generated music, we used accuracy, loss rate, and music theory knowledge as evaluation metrics and compared our model with other music generation models. Experimental results demonstrate our method's significant advantages in music generation quality. Despite progress in the field of automatic music generation, its application still faces challenges, particularly in terms of quantitative evaluation metrics and the breadth of model applications. Future research will continue to explore expanding the model's application scope, enriching evaluation methods, and further improving the quality and expressiveness of the generated music. This study not only advances the development of music generation technology but also provides valuable experience and insights for research in related fields.

Author 1: Sujie He

Keywords: Deep learning; style transfer; innovative composition; Generative Adversarial Networks

PDF

Paper 102: Predictive Modeling of Yoga's Impact on Individuals with Venous Chronic Cerebrospinal System

Abstract: People leading a modern lifestyle often experience varicose veins, commonly attributed to factors associated with work and diet, such as prolonged periods of standing or excess weight. These disorders include elevated blood pressure in the lower extremities, especially the legs. An often-researched metric associated with these illnesses is the Vascular Clinical Severity Score (VCSS), which is connected to discomfort and skin discolorations. However, yoga appears to be a viable way to prevent and manage these problems, significantly lessening the negative consequences of varicose veins. The investigation of yoga's effect on VCSS in this study uses a novel strategy combining machine learning with the Extra Tree Classification (ETC), which is improved by the Cheetah Optimizer (CO) and Black Widow Optimizer (BWO). In this study, the ETC model was combined with previously mentioned optimizers, and two models were amalgamated, referred to as ETBW and ETCO. Through the evaluation of the performance of these models, it was discerned that the accuracy measure for prediction was associated with the ETCO model in the context of VCSS. By revealing subtle correlations between yoga treatments and VCSS results, this multidisciplinary approach seeks to provide a thorough knowledge of preventative and control processes. This research advances the understanding of vascular health by correlating yoga interventions with VCSS outcomes using machine learning and optimization algorithms. By enhancing predictive accuracy, it promotes multidisciplinary collaboration, personalized medicine, and innovation in healthcare, promising improved patient care and outcomes in varicose vein management.

Author 1: Sanjun Qiu

Keywords: Yoga; varicose veins; Extra Tree Classification; cheetah optimization algorithm; Black Widow Optimizer

PDF

Paper 103: Modified Artificial Bee Colony Algorithm for Load Balancing in Cloud Computing Environments

Abstract: Task scheduling in cloud computing is a complex optimization problem influenced by the ever-changing user requirements and the different architectures of cloud systems. Efficiently distributing workloads across Virtual Machines (VMs) is critical to mitigate the negative consequences of inadequate and excessive workloads, such as higher power consumption and possible machine malfunctions. This paper presents a novel method for dynamic load balancing using a Modified Artificial Bee Colony (MABC) algorithm. The ABC algorithm has exceptional competence in solving complex nonlinear optimization problems based on bee colonies' foraging behavior. Nevertheless, the traditional version of the ABC algorithm cannot effectively use resources, resulting in a rapid decline in population diversity and an ineffective spread of knowledge about the best solution between generations. To address these limitations, this study integrates a genetic model into the algorithm, enhancing population diversity through crossover and mutation operators. The developed algorithm is compared with the prevailing algorithms to confirm its effectiveness. The results of the proposed MABC algorithm for the load balancing method are compared with the current ones, and it is observed that this algorithm is more beneficial in terms of cost and energy as well as resource utilization.

Author 1: Qian LI
Author 2: Xue WANG

Keywords: Resource utilization; cloud computing; task scheduling; Artificial Bee Colony; genetic algorithm

PDF

Paper 104: Cloud Workload Prediction Based on Bayesian-Optimized Autoformer

Abstract: Accurate workload forecasting plays a pivotal role in the management of cloud computing resources, enabling significant enhancement in the performance of the cloud platform and effective prevention of resource wastage. However, the complexity, variability, and strong time dependencies of cloud workloads make prediction difficult. To address the challenge of enhancing accuracy in contemporary cloud workload prediction, this paper employs empirical and quantitative research methods, introducing a cloud workload prediction method based on Bayesian-optimized Autoformer, termed BO-Autoformer. Initially, the cloud workload data were divided according to the time-sliding window to construct a continuous feature sequence, which was used as the input of the model to construct the Autoformer prediction model. Subsequently, to further enhance the model's performance, the Bayesian optimization method was employed to identify the optimal combination of hyperparameters, resulting in the development of the Bayesian optimization-based Autoformer cloud workload prediction model. Finally, experiments were conducted on a real Google dataset to evaluate the model's effectiveness. The findings reveal that, compared to alternative models, the proposed prediction model demonstrates superior performance on the cloud workload dataset, and can effectively improve the prediction accuracy of the cloud workload.

Author 1: Biying Zhang
Author 2: Yuling Huang
Author 3: Zuoqiang Du
Author 4: Zhimin Qiu

Keywords: Cloud computing; deep learning; workload prediction; Autoformer; Bayesian optimization

PDF

Paper 105: A Systematic Review on Multi-Factor Authentication Framework

Abstract: In the new era of technology, where information can be accessed and gained at the push of a button, security concerns are raised about protecting the system and data privacy and confidentiality. Traditional ways of user authentication are vulnerable to multiple attacks across all platforms. Various studies propose the use of more than one authentication process to enhance the security level of a system, either hosted on-premise or on the cloud. However, there is limited study on guidelines and appropriate authentication frameworks that suit the needs of an organization. A systematic literature review of a Multi-Factor Authentication framework was conducted through five primary databases: Scopus, IEEE, Science Direct, Springer Link, and Web of Science. The review examined the proposed solution and the underlying methods in a Multi-Factor Authentication framework. Numerous authentication methods were combined to address specific system and data security challenges. The most common authentication method is biometric authentication, which addresses the uniqueness of the user's biological identity. The majority of the proposed solutions were proof of concept and require a pilot test or experiment in the future.

Author 1: Muhammad Syahreen
Author 2: Noor Hafizah
Author 3: Nurazean Maarop
Author 4: Mayasarah Maslinan

Keywords: Data privacy; information; multi-factor authentication; security challenges

PDF

Paper 106: Improved SegNet with Hybrid Classifier for Lung Cancer Segmentation and Classification

Abstract: Prompt diagnosis is crucial globally to save lives, underscoring the urgent need in light of lung cancer's status as a leading cause of death. While CT scans serve as a primary imaging tool for LC detection, manual analysis is laborious and prone to inaccuracies. Recognizing these challenges, computational techniques, particularly ML and DL algorithms, are being increasingly explored as efficient alternatives to enhance the precise identification of cancerous and non-cancerous regions within CT scans, aiming to expedite diagnosis and mitigate errors. The proposed model employs Preprocessing to standardize image features, followed by segmentation using an Improved SegNet framework to delineate cancerous regions. Features like LGXP and MBP are then extracted, facilitating classification with a hybrid classifier which combines LSTM and LinkNet models. Implemented in Python, the model's performance is evaluated against conventional methods, showcasing superior accuracy, sensitivity, and precision. This framework promises to revolutionize LC diagnosis, enabling early intervention and improved patient outcomes.

Author 1: Rathod Dharmesh Ishwerlal
Author 2: Reshu Agarwal
Author 3: K.S. Sujatha

Keywords: Improved SegNet; LGXP; MBN; LSTM; LinkNet; lung cancer

PDF

Paper 107: New 3D Shape Descriptor Extraction using CatBoost Classifier for Accurate 3D Model Retrieval

Abstract: Given the wide application of 3D model analysis, covering domains such as medicine, engineering, and virtual reality, the demand for innovative content-based 3D shape retrieval systems capable of handling complex 3D data efficiently have significantly increased. This paper proposes a new 3D shape retrieval method that uses the CatBoost classifier, a machine learning algorithm, to capture a unique descriptor for each 3D mesh. The main idea of our method is to get a specific and a unique signature or descriptor for each 3D model by training the CatBoost classifier with features obtained directly from the 3D models. This idea not only accelerates the training process, but also ensures the consistency and relevance of the data fed to the classifier during the training process. Once fully trained, the classifier generates a descriptor that is used during the indexing and retrieval process. The efficiency of our method is demonstrated by conducting extensive experiments on the Princeton shape benchmark database. The results demonstrate high retrieval accuracy in comparison to various existing methods in the literature. Our method's ability to outperform these methods shows its potential as highly useful tool in the field of content-based 3D shape retrieval.

Author 1: Mohcine BOUKSIM
Author 2: Fatima RAFII ZAKANI
Author 3: Khadija ARHID
Author 4: Azzeddine DAHBI
Author 5: Taoufiq GADI
Author 6: Mohamed ABOULFATAH

Keywords: 3D object retrieval; 3D shape retrieval; 3D shape matching; indexing; descriptor; CatBoost

PDF

Paper 108: YOLO-T: Multi-Target Detection Algorithm for Transmission Lines

Abstract: During UAV inspections of transmission lines, inspectors often encounter long distance and obstructed targets. However, existing detection algorithms tend to be less accurate when trying to detect these targets. Existing algorithms perform inadequately in handling long-distance and occluded targets, lacking effective detection capabilities for small objects and complex backgrounds. Therefore, we propose an improved YOLOv8-based YOLO-T algorithm for detecting multiple targets on transmission lines, optimized using transfer learning. Firstly, the model is lightweight while ensuring detection accuracy by replacing the original convolution block in the C2f module of the neck network with Ghost convolution. Secondly, to improve the target detection ability of the model, the C2f module in the backbone network is replaced with the Contextual Transformer module. Then, the feature extraction of the model is improved by integrating the Attention module and the residual edge on the SPPF (Spatial Pyramid Pooling-Fast). Finally, we introduce a new shallow feature layer to enable multi-scale feature fusion, optimizing the model detection accuracy for small and obscured objects. Parameters and GFLOPs are conserved by using the Add operation instead of the Concat operation. The experiment reveals that the enhanced algorithm achieves a mean detection accuracy of 97.19% on the transmission line dataset, which is 2.03% higher than the baseline YOLOv8 algorithm. It can also effectively detect small and occluded targets at long distances with a high FPS (98.91 frames/s).

Author 1: Shengwen Li
Author 2: Huabing Ouyang
Author 3: Tian Chen
Author 4: Xiaokang Lu
Author 5: Zhendong Zhao

Keywords: Transmission line inspection; contextual transformer; attention mechanism; ghost convolution

PDF

Paper 109: Identifying Competition Characteristics of Athletes Through Video Analysis

Abstract: The vast repositories of training and competition video data serve as indispensable resources for athlete training and competitor analysis, providing a solid foundation for strategic competition analysis and tactics formulation. However, the effectiveness of these analyses hinges on the abundance and precision of data, often requiring costly professional systems for existing video analysis techniques. Meanwhile, readily accessible non-professional data frequently lacks standardization, compelling manual analysis and experiential judgments, thus limiting the widespread adoption of video analysis technologies. To address these challenges, we have devised an intelligent video analysis technology and a methodology for identifying athletes' competition characteristics. Initially, we employed target detection models, such as You Only Look Once (YOLO), renowned for their ease of deployment and low environmental dependency, to perform fundamental detection tasks. This was further complemented by the intelligent selection of standardized scenes through customizable scene rules, leading to the formation of a standardized scene dataset. On this robust foundation, we achieved classification and identification of competition participants as well as sideline recognition, ultimately compiling a comprehensive competitive dataset. Subsequently, we constructed an athlete posture estimation method utilizing OpenPose, aimed at minimizing interference caused by obstructions and enhancing the accuracy of feature extraction. In experimental validation, we gathered a diverse collection of table tennis competition video data from the internet, serving as a validation dataset. The results were impressive, with a detection success rate for standardized scenes exceeding 94% and an identification success rate for competitors surpassing 98%. The accuracy of posture reconstruction for obstructed individuals exceeded 60%, and the effectiveness of identifying athletes' main features exceeded 90%, convincingly demonstrating the effectiveness of the proposed video analysis method.

Author 1: Yuzhong Liu
Author 2: Tianfan Zhang
Author 3: Zhe Li
Author 4: Mengshuang Ma

Keywords: Video analysis technology; scene recognition method; athlete identification; posture reconstruction; table tennis competition; feature extraction

PDF

Paper 110: Differential Diagnosis of Attention-Deficit/Hyperactivity Disorder and Bipolar Disorder using Steady-State Visual Evoked Potentials

Abstract: Bipolar disorder and Attention-deficit/Hyperactivity disorder (ADHD) are two prevalent disorders whose symptoms are similar. In order to reduce the misdiagnosis between bipolar disorder and ADHD, a machine learning-based system using electroencephalography (EEG) and steady state potentials (i.e., steady-state visual evoked potential [SSVEP]) was evaluated to classify ADHD, bipolar disorder and normal conditions. Indeed, this research was conducted for the first time with the aim of designing a machine learning system for EEG detection of ADHD, bipolar disorder, and normal conditions using SSVEPs. For this purpose, both linear and nonlinear dynamics of extracted SSVEPs were analyzed. Indeed, after data preprocessing, spectral analysis and recurrence quantification analysis (RQA) were applied to SSVEPs. Then, feature selection was utilized through the DISR. Finally, we utilized various machine learning techniques to classify the linear and nonlinear features extracted from SSVEPs into three classes of ADHD, bipolar disorder and normal: k-nearest neighbors (KNN), support vector machine (SVM), linear discriminant analysis (LDA) and Naïve Bayes. Experimental results showed that SVM classifier with linear kernel yielded an accuracy of 78.57% for ADHD, bipolar disorder and normal classification through the leave-one-subject-out (LOSO) cross-validation. Although this research is the first to evaluate the utilization of signal processing and machine learning approaches in SSVEP classification of these disorders, it has limitations that future studies should investigate to enhance the efficacy of proposed system.

Author 1: Xiaoxia Li

Keywords: Attention-deficit/Hyperactivity disorder (ADHD); bipolar disorder; electroencephalography (EEG); steady-state visual evoked potential (SSVEP); machine learning; classification

PDF

Paper 111: Exploring Cutting-Edge Developments in Deep Learning for Biomedical Signal Processing

Abstract: Biomedical condition monitoring devices are progressing quickly by incorporating cost-effective and non-invasive sensors to track vital signs, record medical circumstances, and deliver meaningful responses. These sophisticated innovations rely on breakthrough technology to provide intelligent platforms for health monitoring, quick illness recognition, and precise treatment. Biomedical signal processing determines patterns of signals and serves as the backbone for reliable applications, medical diagnostics, and research. Deep Learning (DL) methods have brought significant innovation in biomedical signal processing, leading to the transformation of the health sector and medical diagnostics. This article covers an entire range of technical innovations evolved for DL-based biomedical signal processing where different modalities have been considered, including Electrocardiography (ECG), Electromyography (EMG), and Electroencephalography (EEG). A vast amount of biomedical data in various forms is available, and DL concepts are required to extract and model this data in order to identify hidden complex patterns that can be utilized to improve the diagnosis, prognosis, and personalized treatment of diseases in an individual. The nature of this developing topic certainly gives rise to a number of challenges. First, the application of sensitive and noisy time series data requires truly robust models. Second, many inferences made at the bedside must have interpretability by design. Third, the field will require that processing be performed in real-time if used for therapeutic interventions. We systematically evaluate these challenges and highlight areas where continued research is needed. The general expansion of DL technologies into the biomedical domain gives rise to novel concerns about accountability and transparency of algorithmic decision-making, a subject which we briefly touch upon as well.

Author 1: Yukun Zhu
Author 2: Haiyan Zhang
Author 3: Bing Liu
Author 4: Junyan Dou

Keywords: Biomedical signal processing; health monitoring; deep learning; electrocardiography; electromyography; electroencephalography

PDF

Paper 112: The Performance of a Temporal Multi-Modal Sentiment Analysis Model Based on Multitask Learning in Social Networks

Abstract: When conducting sentiment analysis on social networks, facing the challenge of temporal and multi-modal data, it is necessary to enable the model to deeply mine and combine information from various modalities. Therefore, this study constructs an emotion analysis model based on multitask learning. This model utilizes a comprehensive framework of convolutional networks, bidirectional gated recurrent units, and multi head self attention mechanisms to represent single modal temporal features in an innovative way, and adopts a cross modal feature fusion strategy. The experiment showed that the model accomplished 0.83 average precision and a 0.83 F1-value, respectively. In contrast with multi-scale attention (0.69, 0.70), aspect-based sentiment analysis (0.78, 0.74), and long short-term memory network (0.71, 0.78) models, this model demonstrated higher robustness and classification accuracy. Especially in terms of parallel computing efficiency, the acceleration ratio of the model reached 1.61, which is the highest among all compared models, highlighting the potential for time savings in large data volumes. This study has shown good performance in sentiment analysis in social networks, providing a novel perspective for solving complex sentiment classification problems.

Author 1: Lin He
Author 2: Haili Lu

Keywords: Multi task learning; multi-modal; emotional analysis; attention mechanism; feature fusion

PDF

Paper 113: Weighted Recursive Graph Color Coding for Enhanced Load Identification

Abstract: In the pursuit of high-precision load identification, traditional methodologies grapple with significant drawbacks, including low recognition rates, intricate signature construction, and narrow applicability. This study introduces a novel approach employing weighted recursive graph (WRG) color coding to surmount these challenges. Power consumption data, procured from advanced load monitoring devices, undergo extraction of single-cycle currents, which are then subjected to dimensional reduction via Piece-wise Aggregate Approximation (PAA). In a transformative step, these currents are encoded into load signatures through the recursive graph time series methodology, culminating in the generation of WRG images. An AlexNet neural network model is engaged to distil and assimilate the distinctive features of the WRG images. The simulation results indicate that the identification rate can exceed 97%. Additionally, an experimental platform was set up to verify the method proposed in this paper, and the results show that the actual identification rate can reach over 96%. Both the simulation results and experiments fully demonstrate that the proposed identification method has a high accuracy. This method not only sets a new standard in non-intrusive load identification but also enhances the generalization of load signature applicability across diverse scenarios.

Author 1: Li Zhang
Author 2: Hengtao Ai
Author 3: Yuhang Liu
Author 4: Shiqing Li
Author 5: Tao Zhang

Keywords: Non-Intrusive Load Monitoring (NILM); Weighted Recurrence Graph (WRG); color coding; AlexNet neural network; load signature

PDF

Paper 114: Diagnosis of NEC using a Multi-Feature Fusion Machine Learning Algorithm

Abstract: Necrotizing enterocolitis (NEC) is a severe gastrointestinal emergency in neonates, marked by its complex etiology, ambiguous clinical manifestations, and significant morbidity and mortality, profoundly affecting long-term pediatric health outcomes. The prevailing diagnostic approaches for NEC, including traditional manual auscultation of bowel sounds, suffer from limited sensitivity and specificity, leading to potential misdiagnoses and delayed treatment. In this paper, we introduce a groundbreaking NEC diagnostic framework employing machine learning algorithms that utilize multi-feature fusion of bowel sounds, significantly improving the diagnostic accuracy. Bowel sounds from NEC patients and healthy newborns are meticulously captured using a specialized acquisition system, designed to overcome the inherent challenges associated with the low amplitude, substantial background noise, and high variability of neonatal bowel sounds. To enhance the diagnostic framework, we extract mel-frequency cepstral coefficient (MFCC), short-time energy (STE), and zero-crossing rate (ZCR) to capture comprehensive frequency and time domain features, ensuring a robust representation of bowel sound characteristics. These features are then integrated using a multi-feature fusion technique to form a singular feature vector, providing a rich, integrated dataset for the machine learning algorithm. Employing the support vector machine (SVM), the algorithm achieved an accuracy (ACC) of 88.00%, sensitivity (SEN) of 100.00%, and an area under the receiver operating characteristic (ROC) curve (AUC) of 97.62%, achieving high accuracy in diagnosing NEC. This innovative approach not only improves the accuracy and objectivity of NEC diagnosis but also shows promise in revolutionizing neonatal care through facilitating early and precise diagnosis. It significantly enhances clinical outcomes for affected neonates.

Author 1: Jiahe Li
Author 2: Yue Han
Author 3: Yunzhou Li
Author 4: Jin Zhang
Author 5: Ling He
Author 6: Tao Xiong
Author 7: Qian Gao

Keywords: Diagnosis of necrotizing enterocolitis (NEC); bowel sound; feature fusion; machine learning

PDF

Paper 115: Towards Optimal Image Processing-based Internet of Things Monitoring Approaches for Sustainable Cities

Abstract: Population growth and urbanization demand innovative strategies for sustainable city management. This paper focuses on the integration of the Internet of Things (IoT) and image processing technologies for environmental monitoring in sustainable urban development. The IoT forms an integral part of the Information and Communication Technology (ICT) infrastructure in smart sustainable cities. It offers a new model for urban design, due to the ability to offer environmentally sustainable alternatives. Furthermore, image processing is a method employed in computer vision that provides reliable approaches for extracting significant data from images. The convergence of these technologies has the capacity to enhance the effectiveness and durability of our urban surroundings. This paper discusses the current state-of-the-art in both IoT and image processing, highlighting their individual applications, architectures, and challenges. This paper explores the integration of the aforementioned technologies in a harmonized monitoring system to promote synergies and complementarities. Several case studies demonstrate the successful adoption of the harmonized approach in urban contexts, focusing on the environmental monitoring, energy management, transportation, and social well-being. The combination of IoT with image processing raises concerns regarding privacy, standardization, and scalability. The study has provided a direction for future research and suggested that more participant and multiple-strategy approaches could be beneficial to address some existing limitations and move toward a more sustainable urban context. It should therefore be viewed as a compass or a roadmap for future research in the areas of IoT and image processing-based monitoring towards todays and future sustainable urban environments.

Author 1: Weiwei LIU
Author 2: Guifeng CHEN

Keywords: Sustainable cities; Internet of Things; image processing; urban monitoring; smart city

PDF

Paper 116: Exploring Enhanced Object Detection and Classification Methods for Alstroemeria Genus Morado

Abstract: As an important ornamental plant, the automatic detection and classification of the maturity of Alstroemeria Genus Morado flowers hold significant importance in precision agriculture. However, this task faces numerous challenges due to the diversity of morphological characteristics, complex growth environments, and factors such as occlusion and lighting variations. Currently, this field is relatively unexplored, necessitating innovative methods to overcome existing difficulties. To fill this research gap, this study developed a deep learning-based object detection framework, the Alstroemeria Genus Morado Network (AGMNet), specifically optimized for the detection and classification of Alstroemeria Genus Morado flowers. This convolutional neural network utilizes multi-scale feature fusion techniques and spatial attention mechanisms, along with a dual-path detection structure, significantly enhancing its capability for automatic maturity classification and detection of flowers. Notably, AGMNet addresses the issue of class imbalance in its design and employs advanced data augmentation techniques to enhance the model's generalization ability. In comparative experiments on the morado_5may dataset, AGMNet demonstrated superior performance in Precision, Recall, and F1-score, with a 3.8% improvement in the mAP metric over the latest YOLOv9 model, showcasing stronger generalization capabilities. AGMNet is expected to play a more significant role in enhancing agricultural production efficiency and automation levels.

Author 1: Yaru Huang
Author 2: Yangxu Wang

Keywords: Alstroemeria; object detection; maturity classification; multi-scale feature fusion; Convolutional Neural Network (CNN)

PDF

Paper 117: Enhanced Arachnid Swarm-Tuned Convolutional Neural Network Model for Efficient Intrusion Detection

Abstract: Digital systems in the connected world of today bring convenience but also complicated cyber security challenges. The inadequacies of conventional intrusion detection techniques are exposed by the constant adaptation and exploitation of vulnerabilities by advanced cyber threats. Identifying dangers in massive data flows gets more difficult as networks grow, necessitating innovative methods. With the aim of minimizing these concerns, a new ID model is created utilizing cutting-edge machine learning to proactively and flexibly combat dynamic cyber attacks, with regard to evolving cyber attackers, this model seeks to improve accuracy and protection systems. This research develops an arachnid swarm optimization-based Convolutional neural network (ASO opt CNN) model to improve ID performance. An improved modified residual CNN is employed in the model to lessen the vanishing and exploding gradient problems in deep networks and facilitates the optimization process, making it easier for deep networks to learn. The developed model is adjusted using arachnid swarm optimization (ASO), which is the hybridization particle swarm optimization (PSO) and social spider optimization (SSO). Utilizing test data, the model's efficacy is evaluated at last. This test data is also subjected to preprocessing, which leads to the creation of a robust detection model that can identify the presence of network attacks. Experimentation and comparison indicate the approach's effectiveness by attaining accuracies of 95.95%, 95.61%, and 95.00% for three datasets respectively. This highlights the developed model’s potential to detect intrusions more effectively.

Author 1: Nishit Patil
Author 2: Shubhlaxmi Joshi

Keywords: Intrusion Detection; arachnid swarm optimization; Convolutional Neural Network; pre-processing; arachnid swarm optimization

PDF

Paper 118: Elevating Offensive Language Detection: CNN-GRU and BERT for Enhanced Hate Speech Identification

Abstract: Upholding a secure and accepting digital environment is severely hindered by hate speech and inappropriate information on the internet. A novel approach that combines Convolutional Neural Network with GRU and BERT from Transformers proposed for enhancing the identification of offensive content, particularly hate speech. The method utilizes the strengths of both CNN-GRU and BERT models to capture complex linguistic patterns and contextual information present in hate speech. The proposed model first utilizes CNN-GRU to extract local and sequential features from textual data, allowing for effective representation learning of offensive language. Subsequently, BERT, advanced transformer-based model, is employed to capture contextualized representations of the text, thereby enhancing the understanding of detailed linguistic nuances and cultural contexts associated with hate speech. Fine tuning BERT model using hugging face transformer. To execute tests using datasets for hate speech identification that are made accessible to the public and show how well the method works to identify inappropriate content. By assisting with the continuing efforts to prevent the dissemination of hate speech and undesirable language online, the proposed framework promotes a more diverse and secure digital environment. The proposed method is implemented using python. The method achieves 98% competitive performance compared to existing approaches LSTM and RNN, CNN, LSTM and GBAT, showcasing its potential for real-world applications in combating online hate speech. Furthermore, it provides insights into the interpretability of the model's predictions, highlighting key linguistic and contextual factors influencing offensive language detection. The study contributes to advancing hate speech detection by integrating CNN-GRU and BERT models, giving a robust solution for enhancing offensive content identification in online platforms.

Author 1: M. Madhavi
Author 2: Sanjay Agal
Author 3: Niyati Dhirubhai Odedra
Author 4: Harish Chowdhary
Author 5: Taranpreet Singh Ruprah
Author 6: Veera Ankalu Vuyyuru
Author 7: Yousef A.Baker El-Ebiary

Keywords: Bidirectional encoder representations from transformers; convolutional neural network; Gated Recurrent Unit; hate speech; hugging face transformer

PDF

Paper 119: Optimizing Resource Allocation in Cloud Environments using Fruit Fly Optimization and Convolutional Neural Networks

Abstract: Cloud computing environments play a crucial role in modern computing infrastructures, offering scalability, flexibility, and cost-efficiency. However, optimizing resource utilization and performance in such dynamic and complex environments remains a significant challenge. This study addresses this challenge by proposing a novel framework that integrates Fruit Fly Optimization (FFO) with Convolutional Neural Networks (CNN) for task scheduling optimization. The background emphasizes the importance of efficient resource allocation and management in cloud computing to meet increasing demands for computational resources while minimizing costs and enhancing overall system performance. The objective of this research is to develop a comprehensive framework that leverages the complementary strengths of FFO and CNN to address the shortcomings of traditional task scheduling approaches. The novelty of the proposed framework lies in its integration of optimization techniques with advanced data analysis methods, enabling dynamic and adaptive task allocation based on real-time workload patterns. The proposed framework is thoroughly evaluated using historical workload data, and results demonstrate significant improvements over traditional methods. Specifically, the FFO-CNN framework achieves average response times ranging from 120 to 180 milliseconds, while maintaining high resource utilization rates ranging from 90% to 98%. These results highlight the effectiveness of the FFO-CNN framework in enhancing resource utilization and performance in cloud computing environments. This research contributes to advancing the state-of-the-art in cloud resource management by introducing a novel approach that combines optimization and data analysis techniques. The proposed framework offers a promising solution to the challenges of resource allocation and task scheduling in cloud computing environments, paving the way for more efficient and sustainable cloud infrastructures in the future.

Author 1: Taviti Naidu Gongada
Author 2: Girish Bhagwant Desale
Author 3: Shamrao Parashram Ghodake
Author 4: K. Sridharan
Author 5: Vuda Sreenivasa Rao
Author 6: Yousef A.Baker El-Ebiary

Keywords: Cloud computing; resource utilization; task scheduling; Fruit Fly Optimization; convolutional neural networks

PDF

Paper 120: Explainable Artificial Intelligence Method for Identifying Cardiovascular Disease with a Combination CNN-XG-Boost Framework

Abstract: Cardiovascular disease (CVD) is a globally significant health issue that presents with a multitude of risk factors and complex physiology, making early detection, avoidance, and effective management a challenge. Early detection is essential for effective treatment of CVD, and typical approaches involve an integrated strategy that includes lifestyle modifications like exercise and diet, medications to control risk factors like high blood pressure and cholesterol, interventions like angioplasties or bypass surgery in extreme cases, and ongoing surveillance to prevent complications and promote heart function. Traditional approaches often rely on manual interpretation, which is time-consuming and prone to error. In this paper, proposed study uses an automated detection method using machine learning. The CNN and XGBoost algorithms' greatest characteristics are combined in the hybrid technique. CNN is excellent in identifying pertinent features from medical images, while XGBoost performs well with tabular data. By including these strategies, the model's robustness and precision in predicting CVD are both increased. Furthermore, data normalization techniques are employed to confirm the accuracy and consistency of the model's projections. By standardizing the input data, the normalization procedure lowers variability and increases the model's ability to extrapolate across instances. This work explores a novel approach to CVD detection using a CNN/XGBoost hybrid model. The hybrid CNN-XGBoost and explainable AI system has undergone extensive testing and validation, and its performance in accurately detecting CVD is encouraging. Due to its ease of use and effectiveness, this technique may be applied in clinical settings, potentially assisting medical professionals in the prompt assessment and care of patients with cardiovascular disease.

Author 1: J Chandra Sekhar
Author 2: T L Deepika Roy
Author 3: K. Sridharan
Author 4: Natrayan L
Author 5: Dr.K.Aanandha Saravanan
Author 6: Ahmed I. Taloba

Keywords: Cardiovascular disease; CNN; XGBoost; traditional approaches; explainable AI

PDF

Paper 121: Utilizing Machine Learning Approach to Forecast Fuel Consumption of Backhoe Loader Equipment

Abstract: This study addresses the challenge of forecasting fuel consumption for various categories of construction equipment, with a specific focus on Backhoe Loaders (BL). Accurate predictions of fuel usage are crucial for optimizing operational efficiency in the increasingly technology-driven construction industry. The proposed methodology involves the application of multiple machine learning (ML) models, including Multiple Linear Regression (MLR), Support Vector Regression (SVR), and Decision Tree Regression (DT), to analyze historical data and key equipment characteristics. The results demonstrate that Decision Tree models outperform other techniques in terms of precision, as evidenced by comparative analysis of the coefficient of determination. These findings enable construction firms to make informed decisions about equipment utilization, resource allocation, and operational productivity, thereby enhancing cost efficiency and minimizing environmental impact. This study provides valuable insights for decision-makers in construction project cost estimation, emphasizing the significant influence of fuel consumption on overall project expenses.

Author 1: Poonam Katyare
Author 2: Shubhalaxmi Joshi
Author 3: Mrudula Kulkarni

Keywords: Machine learning; construction equipment; fuel consumption

PDF

Paper 122: Image Generation of Animation Drawing Robot Based on Knowledge Distillation and Semantic Constraints

Abstract: With the development of robot technology, animation drawing robots have gradually appeared in the public eye. Animation drawing robots can generate many types of images, but there are also problems such as poor quality of generated images and long image drawing time. In order to improve the quality of images generated by animation drawing robots, an animation face line drawing generation algorithm based on knowledge distillation was designed to reduce computational complexity through knowledge distillation. To further raise the quality of images generated by robots, the research also designed an unsupervised facial caricature image generation algorithm based on semantic constraints, which uses facial semantic labels to constrain the facial structure of the generated images. The outcomes denote that the max values of the peak signal-to-noise ratio and feature similarity index measurements of the line drawing generation model are 39.45 and 0.7660 respectively, and the mini values are 37.51 and 0.7483 respectively. The average values of the gradient magnitude similarity bias and structural similarity of the loss function used in this model are 0.2041 and 0.8669 respectively. The max and mini values of Fréchet Inception Distance of the face caricature image generation model are 81.60 and 71.32 respectively, and the max and mini time-consuming values are 15.21s and 13.24s respectively. Both the line drawing generation model and the face caricature image generation model have good performance and can provide technical support for the image generation of animation drawing robots.

Author 1: Dujuan Wang

Keywords: Knowledge distillation; semantic constraints; robot; image; generation

PDF

Paper 123: Integrating AI and IoT in Advanced Optical Systems for Sustainable Energy and Environment Monitoring

Abstract: The increasing demand for sustainable energy solutions and environmental monitoring necessitates advanced technologies. This work combines the capabilities of AI, in the form of a GRU-Auto encoder, with IoT-connected Advanced Optical Systems to create a comprehensive monitoring system. Current monitoring systems often face limitations in real-time analysis and adaptability. Conventional methods struggle to provide timely insights for sustainable energy and environmental management due to the complexity of data patterns and the lack of dynamic adaptability. Our proposed methodology introduces an optimized GRU-Auto encoder, which excels in learning complex temporal patterns, making it well-suited for dynamic environmental and energy data. The integration with Advanced Optical Systems ensures a continuous influx of high-quality real-time data through IoT, enabling more accurate and adaptive analysis. The study involves optimizing the GRU-Auto encoder through hyper parameter tuning and gradient clipping. The model is integrated into an IoT platform that connects with Advanced Optical Systems for seamless data flow. Real-time data from environmental and energy sensors are processed through the AI model, providing immediate insights. Performance is evaluated based on the system's ability to accurately predict environmental trends, optimize energy consumption, and adapt to dynamic changes. Comparative analyses with traditional methods show advantages of the suggested strategy in terms of efficiency and accuracy. This research presents a significant development in the field of study of sustainable energy and environment monitoring, offering a robust solution for real-time data analysis and adaptive decision-making. The integration of an optimized GRU-Auto encoder with IoT-connected Advanced Optical Systems showcases promising results in improving overall system performance and sustainability.

Author 1: Shamim Ahmad Khan
Author 2: Abdul Hameed Kalifullah
Author 3: Kamila Ibragimova
Author 4: Akhilesh Kumar Singh
Author 5: Elangovan Muniyandy
Author 6: Venubabu Rachapudi

Keywords: Auto encoder; artificial intelligence; Internet of Things; gated recurrent unit; sustainable energy; environmental monitoring

PDF

Paper 124: NLP-Based Automatic Summarization using Bidirectional Encoder Representations from Transformers-Long Short Term Memory Hybrid Model: Enhancing Text Compression

Abstract: When the amount of online text data continues to grow, the need for summarized text documents becomes increasingly important. Manually summarizing lengthy articles and determining the domain of the content is a time-consuming and tiresome process for humans. Modern technology can classify large amounts of text documents, identifying key phrases that serve as essential concepts or terms to be included in the summary. Automated text compression allows users to quickly identify the key points and generate the novel words of the document. The study introduces a NLP based hybrid approach for automatic text summarization that combines BERT-based extractive summarization with LSTM-based abstractive summarization techniques. The model aims to create concise and informative summaries. Trained on the BBC news summary dataset, a widely accepted benchmark for text summarization tasks, the model's parameters are optimized using Particle Swarm Optimization, a metaheuristic optimization technique. The hybrid model integrates BERT's extractive capabilities to identify important sentences and LSTM's abstractive abilities to generate coherent summaries, resulting in improved performance compared to individual approaches. PSO optimization enhances the model's efficiency and convergence during training. Experimental results demonstrate the evaluated accuracy scores of ROUGE 1 is 0.671428, ROUGE 2 is 0.56428 and ROUGE L is 0.671428 effectiveness of the proposed approach in enhancing text compression, producing summaries that capture the original text that minimizing redundancy and preserving key information. The study contributes to advancing text summarization tasks and highlights the potential of hybrid NLP-based models in this field.

Author 1: Ranju S Kartha
Author 2: Sanjay Agal
Author 3: Niyati Dhirubhai Odedra
Author 4: Ch Sudipta Kishore Nanda
Author 5: Vuda Sreenivasa Rao
Author 6: Annaji M Kuthe
Author 7: Ahmed I. Taloba

Keywords: Automated text compression; BERT-based extractive summarization; LSTM-based abstractive summarization; NLP-based hybrid approach; Particle Swarm Optimization

PDF

Paper 125: The Impact of Various Factors on the Convolutional Neural Networks Model on Arabic Handwritten Character Recognition

Abstract: Recognizing Arabic handwritten characters (AHCR) poses a significant challenge due to the intricate and variable nature of the Arabic script. However, recent advancements in machine learning, particularly through Convolutional Neural Networks (CNNs), have demonstrated promising outcomes in accurately identifying and categorizing these characters. While numerous studies have explored languages like English and Chinese, the Arabic language still requires further research to enhance its compatibility with computer systems. This study investigates the impact of various factors on the CNN model for AHCR, including batch size, filter size, the number of blocks, and the number of convolutional layers within each block. A series of experiments were conducted to determine the optimal model configuration for the AHCD dataset. The most effective model was identified with the following parameters: Batch Size (BS) = 64, Number of Blocks (NB) = 3, Number of Convolution Layers in Block (NC) = 3, and Filter Size (FS) = 64. This model achieved an impressive training accuracy of 98.29% and testing accuracy of 97.87%.

Author 1: Alhag Alsayed
Author 2: Chunlin Li
Author 3: Ahmed Fat’hAlalim
Author 4: Mohammed Hafiz
Author 5: Jihad Mohamed
Author 6: Zainab Obied
Author 7: Mohammed Abdalsalam

Keywords: Arabic Handwritten Character Recognition (AHCR); Optical Character Recognition (OCR); Deep Learning (DL); Convolutional Neural Network (CNN); Characters Recognition (CR)

PDF

Paper 126: Revolutionary AI-Driven Skeletal Fingerprinting for Remote Individual Identification

Abstract: This research aims to devise a distinct mathematical key for individual identification and recognition. This key, represented through signals, is constructed using Lagrange polynomials derived from the skeletal points. Consequently, we present this key as a novel fingerprint categorized within physiological fingerprints. It’s crucial to highlight that the primary application of this fingerprint is for remote individual identification, specifically excluding any bodily masking. Subsequently, we implement an artificial intelligence model, specifically a Convolutional Neural Network (CNN), for the automated detection of individuals. The proposed CNN is trained on an extensive dataset comprising 10000 real-world cases and augmented data. Our skeletal fingerprint recognition system demonstrates superior performance compared to other physiological fingerprints, achieving a remark-able 98% accuracy in detecting individuals at a distance.

Author 1: Achraf BERRAJAA
Author 2: Ayyoub El OUTMANI
Author 3: Issam BERRAJAA
Author 4: Nourddin SAIDOU

Keywords: Artificial intelligence; recognition of individuals; new fingerprint; lagrange polynomials; CNN

PDF

Paper 127: Deep Learning Enhanced Hand Gesture Recognition for Efficient Drone use in Agriculture

Abstract: The use of deep learning in unmanned aerial vehicles (UAVs), or drones, has greatly improved various technologies by making complex tasks easier, faster, and requiring less human help. This study looks into how artificial intelligence (AI) can be used in farming, especially through creating a system where drones can be controlled by hand gestures to support agricultural activities. By using a special type of AI called a Convolutional Neural Network (CNN) with an EfficientNet B3 model, this research developed a gesture recognition system. It was trained on 1,393 pictures of different hand signals taken under various light conditions and from three different people. The system was evaluated based on its training and testing performance, showing very high scores in terms of loss, accuracy, F1 score, and the Area Under the Curve (AUC), which means it can recognize gestures accurately and work well in different situations. This has big implications for farming, as it gives farmers an easy way to control drones for tasks like checking on crops and spraying them precisely, which also helps keep them safe. This study is an important step towards smarter farming practices. Moreover, the system’s ability to perform well in different settings shows it could also be useful in other areas like construction, where drones need to operate precisely and flexibly.

Author 1: Phaitoon Srinil
Author 2: Pattharaporn Thongnim

Keywords: Deep learning; Convolutional Neural Network; hand gesture recognition; drone; agriculture

PDF

Paper 128: Inclusive Smart Cities: IoT-Cloud Solutions for Enhanced Energy Analytics and Safety

Abstract: Securing smart cities in the evolving Internet of Things (IoT) demands innovative security solutions that extend beyond conventional theft detection. This study introduces temporal convolutional networks and gated recurrent units (TCGR), a pioneering model tailored for the dynamic IoT-SM dataset, addressing eight distinct forms of theft. In contrast to conventional techniques, TCGR utilizes Jaya tuning (TCGRJ), ensuring improved accuracy and computational efficiency. The technique employs ResNeXt for feature extraction to extract important patterns from IoT device-generated data and Edited Nearest Neighbors for data balancing. Empirical evaluations validate TCGRJ’s greater precision (96.7%) and accuracy (97.1%) in detecting theft. The model significantly aids in preventing theft-related risks and is designed for real-time Internet of Things applications in smart cities, aligning with the broader goal of creating safer spaces by reducing hazards associated with unauthorized electrical connections. TCGRJ promotes sustainable energy practices that benefit every resident, particularly those with disabilities, by discouraging theft and encouraging economical power consumption. This research underscores the crucial role of advanced theft detection technologies in developing smart cities that prioritize inclusivity, accessibility, and an enhanced quality of life for all individuals, including those with disabilities.

Author 1: Abdulwahab Ali Almazroi
Author 2: Faisal S. Alsubaei
Author 3: Nasir Ayub
Author 4: Noor Zaman Jhanjhi

Keywords: IoT Security; theft detection; smart cities; cloud computing; disability support

PDF

Paper 129: Enhancing Diabetes Prediction: An Improved Boosting Algorithm for Diabetes Prediction

Abstract: Diabetes is increasing gradually due to the inability to effectively use the human body’s insulin, which threatens public health. People with diabetes who go undiagnosed at early stages or who have diabetes have a high risk of heart disease, kidney disease, eye problems, stroke, and nerve damage for which diabetes diagnosis is crucial to prevent. Our advanced machine learning algorithm is the gateway to a revolutionary possibility of detecting whether the human body has diabetes. Developed this method based on machine learning with one lakh data and the main objective of creating a new and novel diabetes prediction model named moderated Ada-Boost(AB) that can accurately diagnose diabetes. About 10 different classification methods are applied in this research such as Random forest classifier (RF), logistic regression (LR), decision tree classifier (DT), support vector machine (SVM), Bayesian Classifier (BC) or Naive Bayes Classifier (NB), Bagging Classifier (BG), Stacking Classifier (ST), Moderated Ada-Boost(AB) Classifier, K Neighbors Classifier (KN) and Artificial Neural Network (ANN). The crucial contribution is to find out the appropriate values for the different models using the hyper-parameter tuning process. We have proposed a new boosting model named Moderated Ada-Boost(AB) which is the combination of the hyper-parameter tuned random forest model and Ada-boost model. Different evaluation metrics such as accuracy, precision, recall, f1 score, and others are used to evaluate the performance of the models. Our proposed new boosting algorithm named Moderated Ada-Boost(AB) provides better accuracy than other models whose training accuracy is 99.95% and testing accuracy is 98.14%.

Author 1: Md. Shahin Alam
Author 2: Most. Jannatul Ferdous
Author 3: Nishat Sarkar Neera

Keywords: Diabetes prediction; ensemble technique; machine learning; binary classification; Moderated-AdaBoost

PDF

Paper 130: Adaptive Learning Model for Detecting Wheat Diseases

Abstract: Nowadays, the wheat plant has been considered a crucial source of protein, energy, and micronutrients for people. The motivation behind this study comes from how to increase the wheat crop growth and prevent wheat diseases as this plant plays a significant impact on food security all over the world. Wheat plant diseases can be divided into fungal, bacterial, viral, nematode, insect pests, physiological and genetic anomalies, and mineral and environmental stress. Digital images containing the wheat plant disease are collected from different public sources like Kaggle and GitHub. In this study, an adaptive deep-learning model is developed to classify and detect various types of wheat diseases collected digitally in an efficient accurate manner. The dataset is split into two sets: approximately 80% of the data ( 8,946 images) for the training set and 20% (2,259 images) for the validation set. The training set is composed of 1445, 1478, 1557, 1510, 1424, and 1532 images of healthy, leaf rust, powdery mildew, septoria, stem rust, and stripe rust while the validation set contains 357, 360, 404, 402, 353 and 383 images respectively. The suggested method achieved 97.47% validation accuracy on the training set of images and a testing accuracy of 98.42%on the testing set. This study offers a method of training for the classification and detection of wheat diseases using a mix of recently established pre-trained convolutional neural networks (CNN), DenseNet, ResNet, and EfficientNet integrated with the one-fit cycle policy. In comparison to the current state of the art, the proposed model is accurate and efficient.

Author 1: Mohammed Abdalla
Author 2: Osama Mohamed
Author 3: Elshaimaa M. Azmi

Keywords: Food security; image recognition; deep learning; conventional neural networks; digital agriculture; agriculture sustainability

PDF

Paper 131: Detecting Digital Image Forgeries with Copy-Move and Splicing Image Analysis using Deep Learning Techniques

Abstract: The proliferation of digitally altered images across social media platforms has escalated the urgency for robust image forgery detection systems. Traditional detection methodologies, while varied, often fall short in addressing the multifaceted nature of image forgeries in the digital landscape. Recognizing the need for advanced solutions, this paper introduces a novel deep-learning approach that leverages the architectural strengths of GNNs, CNNs, VGG16, MobileNet, and ResNet50. Our method uniquely integrates these architectures to effectively detect and analyze multiple types of image forgeries, including image splicing and copy-move forgeries. This approach is groundbreaking as it adapts these networks to focus on identifying discrepancies in the compression quality between forged and original image regions. By examining the differences between the original and compressed image versions, our model constructs a feature-rich representation, which is then analyzed by a tailored deep-learning network. This network has been enhanced by removing its original classifier and implementing a new one specifically designed for binary forgery classification. Very few researchers have explored the application of deep learning techniques in copy-move and splice image analysis for detecting digital image forgeries, making our work particularly significant. A comprehensive comparative analysis with pre-trained models underscores the superiority of our method, with the GNN model achieving an impressive accuracy of 98.54 percent on the CASIA V1 dataset. This not only sets a new benchmark in the field but also highlights the efficiency of our model, which benefits from reduced training parameters and accelerated training times.

Author 1: Divya Prathana Timothy
Author 2: Ajit Kumar Santra

Keywords: Copy-move; splicing; deep learning; image forgery detection

PDF

Paper 132: An Improved Facial Expression Recognition using CNN-BiLSTM with Attention Mechanism

Abstract: In the recent years, Facial Expression Recognition is one of the hot research topics among the researchers and experts in the field of Computer Vision and Human Computer Interaction. Traditional deep learning models have found it difficult to process images that has occlusion, illumination and pose dimensional properties, and also imbalances of various datasets has led to large distinction in recognition rates, slow speed of convergence and low accuracy. In this paper, we pro-pose a hybrid Convolution Neural Networks-Bidirectional Long Short Term Memory along with point multiplication attention mechanism and Linear Discriminant analysis is incorporated to tackle aforementioned non-frontal image properties with the help of Median Filter and Global Contrast Normalization in data preprocessing. Following this, DenseNet and Softmax is used for reconstruction of images by enhancing feature maps with essential information for classifying the images in the undertaken input datasets i.e. FER2013 and CK+. The proposed model is compared with other traditional models such as CNN-LSTM, DSCNN-LSTM, CNN-BiLSTM and ACNN-LSTM in terms of accuracy, precision, recall and F1 score. The proposed network model achieved highest accuracy in classifying the facial images on FER2013 dataset with 95.12% accuracy which is 3.1% higher than CNN-LSTM, 2.7% higher than DSCNN-LSTM, 2% higher than CNN-BiLSTM and 3.7% higher than ACNN-LSTM network models, and the proposed model has achieved 98.98% of accuracy with CK+ in classifying the images which is 5.1% higher than CNN-LSTM, 5.7% higher than DSCNN-LSTM, 3.3% higher than CNN-BiLSTM and 6.9% higher than ACNN-LSTM network models in facial expression recognition.

Author 1: Samanthisvaran Jayaraman
Author 2: Anand Mahendran

Keywords: Facial expression recognition; occlusion; attention mechanism; convolution neural networks; bidirectional long short time memory

PDF

Paper 133: A Survey of Reversible Data Hiding in Encrypted Images

Abstract: The creation and application of multimedia has undergone a revolution in the last several years. This is a result of the rise in internet-based communications, which involves the exchange of digital data in the forms of text files, audio files, video files, and image files. For this reason, multimedia has emerged as a vital aspect of people’s everyday existence. Information security is crucial since there are several threats that target multime-dia integrity, confidentiality, and authentication.Multimedia data needs to be safeguarded, perhaps using encryption, in order to solve these numerous issues. Reversible data hiding in encrypted pictures (RDHEI) is investigate in this survey. (RDHEI) process, which functions by adding extra data to a picture, has surfaced. Employers and academics alike are becoming more interested in and focused on the RDHEI due to its vast range of applications. The purpose of this review is to introduce the various RDHEI schemes, identify the most important RDHEI techniques with varying embedding rates, and then examine the applications and future prospects of RDHEI. The main characteristics of each representative RDHEI Technique taken into consideration in this survey are enumerated in a comparison table.

Author 1: Ghadeer Asiri
Author 2: Atef Masmoudi

Keywords: Reversible data hiding; encrypted image

PDF

Paper 134: HybridGCN: An Integrative Model for Scalable Recommender Systems with Knowledge Graph and Graph Neural Networks

Abstract: Graph Neural Networks (GNNs) have emerged as a state-of-the-art approach in building modern Recommender Systems (RS). By leveraging the complex relationships among items, users, and their attributes, which can be represented as a Knowledge Graph (KG), these models can explore implicit semantic sub-structures within graphs, thereby enhancing the learning of user and item representations. In this paper, we propose an end-to-end architectural framework for developing recommendation models based on GNNs and KGs, namely Hy-bridGCN. Our proposed methodologies aim to address three main challenges: (1) making graph-based RS scalable on large-scale datasets, (2) constructing domain-specific KGs from unstructured data sources, and (3) tackling the issue of incomplete knowledge in constructed KGs. To achieve these goals, we design a multi-stage integrated procedure, ranging from user segmentation and LLM-supported KG construction process to interconnectedly propagating between the KG and the Interaction Graph (IG). Our experimental results on a telecom e-commerce domain dataset demonstrate that our approach not only makes existing GNN-based recommender baselines feasible on large-scale data but also achieves comparative performance with the HybridGCN core.

Author 1: Dang-Anh-Khoa Nguyen
Author 2: Sang Kha
Author 3: Thanh-Van Le

Keywords: Large-scale dataset processing; recommender systems; graph neural network; knowledge graph construction; data segmentation

PDF

Paper 135: Transformer Meets External Context: A Novel Approach to Enhance Neural Machine Translation

Abstract: Most neural machine translation (NMT) systems rely on parallel data, comprising text in the source language and its corresponding translation in the target language. While it’s acknowledged that context enhances NMT models, this work proposes a novel approach by incorporating external context, specifically explanations of source text meanings, akin to how human translators leverage context for comprehension. The suggested methodology innovatively addresses the challenge of incorporating lengthy contextual information into NMT systems. By employing state-of-the-art transformer-based models, external context is integrated, thereby enriching the translation process. A key aspect of the approach lies in the utilization of diverse text summarization techniques, strategically employed to efficiently distill extensive contextual details into the NMT framework. This novel solution not only overcomes the obstacle posed by lengthy context but also enhances the translation quality, marking a advancement in the field of NMT. Furthermore, the data-centric approach ensures robustness and effectiveness, yielding improvements in translation quality, as evidenced by a considerable boost in BLEU score points ranging from 0.46 to 1.87 over baseline models. Additionally, we make our dataset publicly available, facilitating further research in this domain.

Author 1: Mohammed Alsuhaibani
Author 2: Kamel Gaanoun
Author 3: Ali Alsohaibani

Keywords: Deep learning; transformers; context; NMT; neural machine translation; natural language processing systems

PDF

Paper 136: Mitigating Security Risks in Firewalls and Web Applications using Vulnerability Assessment and Penetration Testing (VAPT)

Abstract: In today’s digital age, both organizations and individuals heavily depend on web applications for a wide range of activities. However, this reliance on the web also opens up opportunities for attackers to exploit security weaknesses present in these applications. Web Application Firewalls (WAFs) are typically the first line of defense, protecting web apps by filtering and monitoring HTTP traffic. However, if these firewalls are not properly configured, they can be bypassed or compromised by attackers. The escalating number of attacks targeting web applications underscores the urgent need to enhance their security. This paper offers an in-depth review of existing research on web application Vulnerability Assessment and Penetration Testing (VAPT). Our unique contribution lies in the comprehensive synthesis and categorization of VAPT tools based on their optimal use cases, which provides a practical guide for selecting the appropriate tools for specific scenarios. Additionally, this study integrates emerging technologies such as artificial intelligence and machine learning into the VAPT framework, addressing the evolving nature of cyber threats. The paper also identifies common challenges encountered during the VAPT process and proposes actionable recommendations to overcome these obstacles. Furthermore, it discusses best practices such as secure coding practices and defense-in-depth strategies to improve the effectiveness and efficiency of VAPT efforts. By offering these insights, this paper aims to advance the current understanding and application of VAPT in enhancing the security of web applications and firewalls.

Author 1: Alanoud Alquwayzani
Author 2: Rawabi Aldossri
Author 3: Mounir Frikha

Keywords: Web Application Firewalls (WAFs); Vulnerability Assessment and Penetration Testing (VAPT); cybersecurity; security vulnerabilities; security misconfigurations; network scanning tools; vulnerability detection

PDF

Paper 137: A Deep Learning Approach to Convert Handwritten Arabic Text to Digital Form

Abstract: The recognition of Arabic words presents considerable difficulties owing to the complex characteristics of the Arabic script, which encompasses letters positioned both above and below the baseline, hamzas, and dots. In order to address these intricacies, we provide a structured approach for transforming handwritten Arabic text into a digital format. We employ a hybrid deep learning technique that combines Convolutional Neural Networks (CNNs), Bidirectional Long Short-Term Memory (BLSTM), and Connectionist Temporal Classification (CTC). We collected datasets that cover a wide range of Arabic text variations. We have also created a pre-processing pipeline. Our methodology successfully achieved an accuracy rate of 99.52%. At the level of recognizing the letters of the word, with an accuracy of 98.36% at the level of the full word. In order to evaluate the effectiveness of our suggested method for recognizing handwritten text, we utilize two essential metrics: Word Error Rate (WER) and Character Error Rate (CER) to compare its performance. The experimental research demonstrates a WER of 1.64 % and a CER of 0.48%.

Author 1: Bayan N. Alshahrani
Author 2: Wael Y. Alghamdi

Keywords: Deep learning; convolutional neural networks; bidirectional long short term memory; connectional temporal classification; Arabic handwriting recognition

PDF

Paper 138: User-Friendly Privacy-Preserving Blockchain-based Data Trading

Abstract: As the digital economy flourishes, the use of blockchain technology for data trading has seen a surge in popularity. Yet, previous approaches have frequently faltered in harmonizing security with user experience, culminating in suboptimal transactional efficiency. This study introduces a personalized local differential privacy framework, adeptly tackling data security concerns while accommodating the individual privacy preferences of data owners. Furthermore, the framework bolsters transaction flexibility and efficiency by catering to needs of data consumers for detailed queries and enabling data owners to effortlessly elevate their privacy budget to achieve greater financial returns. The efficacy of our approach is validated through a comprehensive series of experimental validations.

Author 1: Jiahui Cao
Author 2: Junyao Ye
Author 3: Junzuo Lai

Keywords: Data trading; blockchain; personalized local differential privacy; data security; user-friendly

PDF

Paper 139: AEGANB3: An Efficient Framework with Self-attention Mechanism and Deep Convolutional Generative Adversarial Network for Breast Cancer Classification

Abstract: Breast cancer remains a significant illness around the world, but it has become the most dangerous when faced with women. Early detection is paramount in improving prognosis and treatment. Thus, ultrasonography has appeared as a valuable diagnostic tool for breast cancer. However, the accurate interpretation of ultrasound images requires expertise. To address these challenges, recent advancements in computer vision such as using convolutional neural networks (CNN) and vision transformers (ViT) for the classification of medical images, which become popular and promise to increase the accuracy and efficiency of breast cancer detection. Specifically, transfer learning and fine-tuning techniques have been created to leverage pre-trained CNN models. With a self-attention mechanism in ViT, models can effectively feature extraction and learning from limited annotated medical images. In this study3, the Breast Ultrasound Images Dataset (Dataset BUSI) with three classes including normal, benign, and malignant was utilized to classify breast cancer images. Additionally, Deep Convolutional Generative Adversarial Networks (DCGAN) with several techniques were applied for data augmentation and preprocessing to increase robustness and address data imbalance. The AttentiveEfficientGANB3 (AEGANB3) framework is proposed with a customized EfficientNetB3 model and self-attention mechanism, which showed an impressive result in the test accuracy of 98.01%. Finally, Gradient-weighted Class Activation Mapping (Grad-CAM) for visualizing the model decision.

Author 1: Huong Hoang Luong
Author 2: Hai Thanh Nguyen
Author 3: Nguyen Thai-Nghe

Keywords: Breast cancer; classification; Convolutional Neural Network (CNN); Vision Transformer (ViT); fine-tuning; transfer learning; self-attention

PDF

Paper 140: An Optimal Knowledge Distillation for Formulating an Effective Defense Model Against Membership Inference Attacks

Abstract: A membership inference attack (MIA) on machine learning models aims to determine the sensitive data that has been used to train machine learning models. Machine learning-based applications (MLaaS—machine learning as a service) in finance, banking, healthcare, etc. are facing the risks of private data leaks by MIA. Several solutions have been proposed for mitigating MIA attacks, such as confidence score masking, regularization, knowledge distillation (KD), etc. However, the utility-privacy trade-off problem is still a major challenge for existing approaches. In this work, we explore the KD-based approach to defending against MIA attacks. This approach has received increasing attention in the research community on machine learning safety recently as it aims at effectively addressing the above-mentioned challenge of mitigating MIA attacks. An efficient KD-based defense framework that includes multiple teacher and student models is proposed in this work for alleviating MIA attacks. Three main phases are deployed in this framework: (1) teacher model training; (2) knowledge distillation from the teacher model to the student model based on prediction augmentation and aggregation from the teacher model; and (3) repeated knowledge distillation among student models. The experimental results on standard datasets show the outperforms in both model utility and privacy of the proposed framework compared to other state-of-the-art solutions for mitigating MIA.

Author 1: Thi Thanh Thuy Pham
Author 2: Huong-Giang Doan

Keywords: Knowledge distillation; membership inference attack; teacher model; student model; privacy-utility trade-off

PDF

Paper 141: Audio Watermarking: A Comprehensive Review

Abstract: Audio watermarking has emerged as a potent technology for copyright protection, content authentication, content monitoring, and tracking in the digital age. This paper offers a comprehensive exploration of audio watermarking principles, techniques, applications, and challenges. Initially, it presents the fundamental concepts of digital watermarking, elucidating its key characteristics and functionalities. After that, different audio watermarking methods in both the time and transform domains are explained, such as feature-based, parametric, and spread-spectrum methods, along with how they work, and their pros and cons. The paper further addresses critical challenges in maintaining key criteria such as imperceptibility, robustness, and payload capacity associated with audio watermarking. Additionally, it examines watermarking evaluation metrics, datasets, and performance findings under diverse signal-processing attacks. Finally, the review concludes by discussing future directions in audio watermarking research, emphasizing advancements in deep learning-based approaches and emerging applications.

Author 1: Mohammad Shorif Uddin
Author 2: Ohidujjaman
Author 3: Mahmudul Hasan
Author 4: Tetsuya Shimamura

Keywords: Audio watermarking; deep learning approach; spread-spectrum method; signal-processing attacks; time domain; transform domain

PDF

Paper 142: ACNGCNN: Improving Efficiency of Breast Cancer Detection and Progression using Adversarial Capsule Network with Graph Convolutional Neural Networks

Abstract: New diagnostic methods are needed to improve the accuracy and efficiency of breast cancer detection and progression. Although successful, current methods frequently lack precision, accuracy, and timeliness, especially in the early phases Of breast cancer progression. Our research proposes a new model using deep learning to improve breast cancer detection and classification, addressing constraints. Our breast cancer image and sample preprocessing approach combines a non-local means filter (NLM) and Generative Adversarial Networks (GAN). The model classifies datasets using LSTM with BiGRU-based Recurrent ShuffleNet V2, a highly efficient and accurate technique for sequential data samples. The integration of a Capsule Network with Graph Convolutional Neural Networks (CNGCNN) significantly improves breast cancer detection. This method was carefully tested on BreaKHis. The results were amazing, showing gains across multiple metrics: 4.9% greater precision, 3.5% higher accuracy, 3.4% higher recall, 2.5% higher AUC (Area Under the Curve), 1.9% higher specificity, and 3.4%decreased delay in the identification of breast cancer stages. Particularly striking was the model’s performance in diagnosing illness development, where it displayed 3.5% greater precision, 3.9% higher accuracy, 4.5% higher recall, 3.4% higher AUC, 2.9% higher specificity, and 1.5% lower latency. Significant clinical impacts result from this work. Our methodology enables early diagnosis and precise staging of breast cancer, enabling focused therapies to improve patient outcomes and survival rates. The greater precision and reduced time lag in diagnosing disease progression also allow for more effective monitoring and treatment modifications. Overall, this study marks a considerable improvement in the field of breast cancer diagnostics, delivering a more efficient, accurate, and reliable tool for healthcare providers in their fight against this ubiquitous disease.

Author 1: Srinivasa Rao Pallapu
Author 2: Khasim Syed

Keywords: Breast cancer detection; deep learning; image pre-possessing; disease progression; recurrent neural networks

PDF

Paper 143: Securing Networks: An In-Depth Analysis of Intrusion Detection using Machine Learning and Model Explanations

Abstract: As cyber threats continue to evolve in complexity, the need for robust intrusion detection systems (IDS) becomes increasingly critical. Machine learning (ML) models have demon-strated their effectiveness in detecting anomalies and potential intrusions. In this article, we delve into the world of intrusion detection by exploring the application of four distinct ML models: XGBoost, Decision Trees, Random Forests, and Bagging. And leveraging the interpretability tools LIME (Local Interpretable Model-agnostic Explanations) and SHAP (SHapley Additive ex-Planations) to explain the classification results. Our exploration begins with an in-depth analysis of each machine learning model, shedding light on their strengths, weaknesses, and suitability for intrusion detection. However, machine learning models often operate as ”black boxes” making it crucial to explain their inner workings. This article introduces LIME (Local Interpretable Model-agnostic Explanations) and SHAP (SHapley Additive ex-Planations) as indispensable tools for model interpretability. Throughout the article, we demonstrate the practical application of LIME and SHAP to explain and interpret the output of our intrusion detection models. By doing so, we gain valuable insights into the decision-making process of these models, enhancing our ability to identify and respond to potential threats effectively.

Author 1: Hoang-Tu Vo
Author 2: Nhon Nguyen Thien
Author 3: Kheo Chau Mui
Author 4: Phuc Pham Tien

Keywords: Intrusion detection systems; machine learning models; model interpretability; cybersecurity; LIME; SHAP; explain-able machine learning models

PDF

Paper 144: Log Clustering-based Method for Repairing Missing Traces with Context Probability Information

Abstract: In real business processes, low quality event logs due to outliers and missing values tend to degrade the performance of process mining related algorithms, which in turn affects the correct execution of decisions. In order to repair the missing values in event logs under the condition that the reference model of the process system is unknown, this paper proposes a method that can repair consecutive missing values. First, the event logs are divided according to the integrity of the trace, and then the cluster algorithm is applied to complete logs to generate homogeneous trace clusters. Then match the missing trace to the most similar sub log, generate candidate sequences according to the context of the missing part, calculate the context probability of each candidate sequence, and select the one with the highest probability as the repair result. When the number of missing items in the trace is 1, our method has the highest repair accuracy of 97.5 percent in the Small log and 93.3 percent in the real event logs bpic20. Finally, the feasibility of this method is verified on four event logs with different missing ratios and has certain advantages compared with existing methods.

Author 1: Huan Fang
Author 2: Wenjie Su

Keywords: Trace clustering; log repairing; process mining; context semantics; conditional probability

PDF

The Science and Information (SAI) Organization
BACK TO TOP

Computer Science Journal

  • About the Journal
  • Call for Papers
  • Submit Paper
  • Indexing

Our Conferences

  • Computing Conference
  • Intelligent Systems Conference
  • Future Technologies Conference
  • Communication Conference

Help & Support

  • Contact Us
  • About Us
  • Terms and Conditions
  • Privacy Policy

© The Science and Information (SAI) Organization Limited. All rights reserved. Registered in England and Wales. Company Number 8933205. thesai.org