The Science and Information (SAI) Organization
  • Home
  • About Us
  • Journals
  • Conferences
  • Contact Us

Publication Links

  • IJACSA
  • Author Guidelines
  • Publication Policies
  • Digital Archiving Policy
  • Promote your Publication
  • Metadata Harvesting (OAI2)

IJACSA

  • About the Journal
  • Call for Papers
  • Editorial Board
  • Author Guidelines
  • Submit your Paper
  • Current Issue
  • Archives
  • Indexing
  • Fees/ APC
  • Reviewers
  • Apply as a Reviewer

IJARAI

  • About the Journal
  • Archives
  • Indexing & Archiving

Special Issues

  • Home
  • Archives
  • Proposals
  • Guest Editors
  • SUSAI-EE 2025
  • ICONS-BA 2025
  • IoT-BLOCK 2025

Future of Information and Communication Conference (FICC)

  • Home
  • Call for Papers
  • Submit your Paper/Poster
  • Register
  • Venue
  • Contact

Computing Conference

  • Home
  • Call for Papers
  • Submit your Paper/Poster
  • Register
  • Venue
  • Contact

Intelligent Systems Conference (IntelliSys)

  • Home
  • Call for Papers
  • Submit your Paper/Poster
  • Register
  • Venue
  • Contact

Future Technologies Conference (FTC)

  • Home
  • Call for Papers
  • Submit your Paper/Poster
  • Register
  • Venue
  • Contact
  • Home
  • Call for Papers
  • Editorial Board
  • Guidelines
  • Submit
  • Current Issue
  • Archives
  • Indexing
  • Fees
  • Reviewers
  • Subscribe

IJACSA Volume 16 Issue 1

Copyright Statement: This is an open access publication licensed under a Creative Commons Attribution 4.0 International License, which permits unrestricted use, distribution, and reproduction in any medium, even commercially as long as the original work is properly cited.

View Full Issue

Paper 1: Advanced Machine Learning Approaches for Accurate Migraine Prediction and Classification

Abstract: Migraine is a neurovascular disorder with a prevalence that exceeds 1 billion individuals worldwide, but it has long been recognized to have unique diagnostic challenges due to its heterogeneous pathophysiology and dependence on subjective assessments. As has been extensively documented by a number of international law bodies, migraine in the workplace has been identified as a significant issue that requires urgent attention. Migraine defined by episodic, unilateral and debilitating symptoms including aura, nausea incurs a high socioeconomic burden in disability. Mechanisms such as altered cortical excitability and trigeminal system activation, although researched to a high extent, are still inadequately understood. Deep learning and machine learning (ML) hold tremendous potential for transforming diagnosis and classification of migraine. This study evaluates several machine learning (ML) models such as gradient boosting, decision tree, random forest, k-Nearest Neighbors (KNN), support vector machine (SVM), logistic regression, multi-layer perceptron (MLP), artificial neural network (ANN), and deep neural network (DNN) for multi-class classification of migraine. By employing advanced preprocessing techniques and publicly obtainable datasets, the study addresses the challenge of identifying different types of migraines that may share common variables. In this study, several machine learning (ML) models including gradient boosting, decision tree, random forest, k-Nearest Neighbors show that for multi-class migraine classification MLP and Gradient Boosting had good performance in most models, but did perform poorly in complex subcategories like Typical Aura with Migraine. Both attained high accuracies (96.4% and 97%, respectively). KNN and Logistic Regression, two traditional models, performed well at basic classifications but poorly at more complex situations; Neural networks (ANN and DNN) showed much flexibility towards data complexities. These results underscore how important it is to align model selection with data properties and provide avenues for improving performance through regularization and feature engineering. This strategy illustrates how AI-powered solutions can revolutionize the way we manage, treat, and prevent migraines across the globe.

Author 1: Chokri Baccouch
Author 2: Chaima Bahar

Keywords: Headache classification; migraine; migraine diagnosis; migraine classification

PDF

Paper 2: A Comparative Study of Predictive Analysis Using Machine Learning Techniques: Performance Evaluation of Manual and AutoML Algorithms

Abstract: In this study, we have compared manual machine learning with automated machine learning (AutoML) to see which performs better in predictive analysis. Using data from past football matches, we tested a range of algorithms to forecast game outcomes. By exploring the data, we discovered patterns and team correlations, then cleaned and prepped the data to ensure the models had the best possible inputs. Our findings show that AutoML, especially when using logistic regression can outperform manual methods in prediction accuracy. The big advantage of AutoML is that it automates the tricky parts, like data cleaning, feature selection, and tuning model parameters, saving time and effort compared to manual approaches, which require more expertise to achieve similar results. This research highlights how AutoML can make predictive analysis easier and more accurate, providing useful insights for many fields. Future work could explore using different data types and applying these techniques to other areas to show how adaptable and powerful machine learning can be.

Author 1: Karim Mohammed Rezaul
Author 2: Md. Jewel
Author 3: Anjali Sudhan
Author 4: Mifta Uddin Khan
Author 5: Maharage Roshika Sathsarani Fernando
Author 6: Kazy Noor e Alam Siddiquee
Author 7: Tajnuva Jannat
Author 8: Muhammad Azizur Rahman
Author 9: Md Shabiul Islam

Keywords: Machine learning; predictive analytics; sports forecasting; automated machine learning (AutoML); feature engineering; model evaluation; data pre-processing; algorithm comparison; football analytics; sports betting; team performance metrics; exploratory data analysis (EDA); cross-validation techniques

PDF

Paper 3: Detection of DDoS Cyberattack Using a Hybrid Trust-Based Technique for Smart Home Networks

Abstract: As Smart Home Internet of Things (SHIoT) continue to evolve, improving connectivity and security whilst offering convenience, ease, and efficiency is crucial. SHIoT networks are vulnerable to several cyberattacks, including Distributed Denial of Service (DDoS) attacks. The ever-changing landscape of Smart Home IoT threats presents many problems for current cybersecurity techniques. In response, we propose a hybrid Trust-based approach for DDoS attack detection and mitigation. Our proposed technique incorporates adaptive mechanisms and trust evaluation models to monitor device behaviour and identify malicious nodes dynamically. By leveraging real-time threat detection and secure routing protocols, the proposed trust-based mechanism ensures uninterrupted communication and minimizes the attack surface. Additionally, energy-efficient techniques are employed to safeguard communication without overburdening resource-constrained SHIoT devices. To evaluate the effectiveness of the proposed technique in efficiently detecting and mitigating DDoS attacks, we conducted several simulation experiments and compared the performance of the approach with other existing DDoS detection mechanisms. The results showed notable improvements in terms of energy efficiency, improved system resilience and enhanced computations. Our solution offers a targeted approach to securing Smart Home IoT environments against evolving cyber threats.

Author 1: Oghenetejiri Okporokpo
Author 2: Funminiyi Olajide
Author 3: Nemitari Ajienka
Author 4: Xiaoqi Ma

Keywords: Trust; smart home; IoT; DDoS; denial of service; DoS; cyber threats; techniques

PDF

Paper 4: Forecasting the Emergence of a Dominant Design by Classifying Product and Process Patents Using Machine Learning and Text Mining

Abstract: Forecasting the emergence of a dominant design in advance is important because the emergence of the dominant design can provide useful information about the external environment for the product launch. Although the emergence of the dominant design can only be determined as a result of the introduction of the product into the market, it may be possible to predict the emergence of the dominant design in advance by applying a solution based on patent analysis. In the newly proposed technique of separating patents, we can capture changes in the state of technological innovation and analyze the emergence of the dominant design, but there is a problem that it requires processing of large amounts of patent data, and that the processing involves subjective judgments by experts. This study focuses on analyzing technological innovation trends using an approach that separates product patents from process patents, investigates whether this approach can be applied to machine learning, and aims to develop a learning model that automatically classifies patents. We applied text mining to patent information to create structured data sets and compared nine different machine learning classification algorithms with and without dimensionality reduction. The approach was effectively applied to machine learning, and the Random Forest, AdaBoost and Support Vector Machine models achieved high classification performance of over 95%. By developing these learning models, it is possible to objectively forecast the emergence of a dominant design with high accuracy.

Author 1: Koji Masuda
Author 2: Yoshinori Hayashi
Author 3: Shigeyuki Haruyama

Keywords: Dominant design; patent analysis; technological innovation; machine learning; text mining; classification

PDF

Paper 5: Control Interface for Multi-User Video Games with Hand or Head Gestures in Directional Key-Based Games

Abstract: This paper describes the development and implementation of a hand or head gesture-based control interface for video games, enhanced for games that use directional keys. The objective is to develop an adaptive control system for a multiplayer video game that allows users to choose between the use of traditional directional keys or a gesture-based interface. The methodology used follows the Cross-Industry Standard Process for Data Mining (CRISP-DM) development model, which allows a structured integration of analysis, design, implementation and evaluation steps. Technologies such as OpenCV, MediaPipe and deep learning algorithms are used, translating hand movements into directional commands in real time. In addition, the system integrates a client-server architecture based on Node.js that supports multiple users, enabling an immersive gaming experience on PC and mobile platforms. The results highlight the accuracy of the system and its potential to improve accessibility, especially for users with motor disabilities by using their hands or head movements to control the directional keys. Concluding that the control interface for multi-user video games provides the necessary support to gamers in performing the task, promoting accessibility in the entertainment environment.

Author 1: Oscar Ramirez-Valdez
Author 2: César Baluarte-Araya
Author 3: Rodrigo Castillo-Lazo
Author 4: Italo Ccoscco-Alvis
Author 5: Alexander Valdiviezo-Tovar
Author 6: Alexander Villafuerte-Quispe
Author 7: Dylan Zuñiga-Huraca

Keywords: Control interface; video games; artificial vision; gesture-based interface; directional commands; human-computer interaction; deep learning algorithms; accessibility; real-time; pattern recognition

PDF

Paper 6: Teaching Programming in Higher Education: Analyzing Trends, Technologies, and Pedagogical Approaches Through a Bibliometric Lens

Abstract: In today’s information society, developing programming competencies is essential in higher education. Numerous studies have been conducted on effective strategies for fostering these skills. This study performs a bibliometric analysis of research on teaching strategies for programming in higher education, using data from the SCOPUS and Web of Science (WOS) databases between 2014 and 2023. The analysis identifies key trends, influential authors, and collaboration networks in this field. The most effective teaching strategies include project-based learning, flipped classrooms, and collaborative programming. Emerging technologies such as augmented reality and virtual reality are gaining prominence in programming education. Despite the growth of research in this area, challenges remain, such as the lack of longitudinal studies exploring the long-term impact of these methodologies and the need for greater geographic diversity in studies. This paper emphasizes the importance of exploring new technologies and interdisciplinary approaches and fostering international collaborations to enhance programming education. The findings guide researchers and educators on how to optimize programming learning in a global context.

Author 1: Mariuxi Vinueza-Morales
Author 2: Jorge Rodas-Silva
Author 3: Cristian Vidal-Silva

Keywords: Programming; higher education; teaching strategies; bibliometrics

PDF

Paper 7: Harnessing the Power of Federated Learning: A Systematic Review of Light Weight Deep Learning Protocols

Abstract: With rapid proliferation in using smart devices, real time efficient sentiment analysis has gained considerable popularity. These devices generate variety of data. However, for resource con-strained devices to perform sentiment analysis over multimodal data using conventional modals that are computationally complex and resource hungry, is challenging. This challenge may be addressed using a light weight but efficient modal specifically focused on sentiment analysis for contrained devices. in the literature, there are several modals that claims to be light weight however, the real sense and logic to determine if the modal may be termed as lightweight still requires further research. This paper reviews approaches to federated learning for multimodal sentiment analysis. Federated learning enables decen-tralized training without sharing data. Considering the review need to balance privacy concerns, performance, and resource usage, the review evaluates existing approaches to enhance accuracy in sentiment classification. The review identifies strengths and limitations in handling multimodal data. The search focused on studies in databases like IEEE Xplore and Scopus. Studies published in peer-reviewed journals over the past five years were included. The review covers 45 studies, mostly experimental, with some theoretical models. Key results show lightweight protocols improve efficiency and privacy in federated learning. They reduce computational demands while handling text, image, and audio data. There is a growing focus on resource-constrained devices in research. Trade-offs between model complexity and speed are commonly explored. The review addresses how these protocols balance accuracy and computational cost.

Author 1: Haseeb Khan Shinwari
Author 2: Riaz Ul Amin

Keywords: Light weight protocols; Sentiment analysis; federated learning; deep learning

PDF

Paper 8: SEC-MAC: A Secure Wireless Sensor Network Based on Cooperative Communication

Abstract: Wireless Sensor Networks (WSNs) are essential for a wide range of applications, from environmental monitoring to security systems. However, challenges such as energy efficiency, throughput, and packet delivery delay need to be addressed to enhance network performance. This paper introduces a novel Medium Access Control (MAC) protocol that utilizes cooperative communication strategies to improve these critical metrics. The proposed protocol enables source nodes to leverage intermediate nodes as relays, facilitating efficient data transmission to the access point. By employing a cross-layer approach, the protocol optimizes the selection of relay nodes based on factors like transmission time and residual energy, ensuring optimal end-to-end paths. The protocol's performance is rigorously evaluated using a simulation environment, demonstrating significant improvements over existing methods. Specifically, the protocol enhances throughput by 12%, boosts energy efficiency by 50%, and reduces average packet delivery delay by approximately 48% than IEEE 802.11b. These results indicate that the protocol not only extends the lifespan of sensor nodes by conserving energy but also improves the overall reliability and efficiency of the WSN, making it a robust solution for modern wireless sensor networks. Security in Wireless Sensor Networks (WSNs) is crucial due to vulnerabilities like eavesdropping, data tampering, and denial of service attacks. Our proposed MAC protocol addresses these challenges by incorporating authentication techniques, such as the handshaking protocol. These measures protect data integrity, confidentiality, and availability, ensuring reliable and secure data transmission across the network. This approach enhances the resilience of WSNs, making them more secure and trustworthy for critical applications such as healthcare and security monitoring.

Author 1: Yassmin Khairat
Author 2: Tamer O. Diab
Author 3: Ahmed Fawzy
Author 4: Samah Osama
Author 5: Abd El- Hady Mahmoud

Keywords: Wireless Sensor Networks (WSNs); energy efficiency; Media Access Control (MAC); cooperative communication; handshaking algorithm

PDF

Paper 9: Digital Twin Model from Freehanded Sketch to Facade Design, 2D-3D Conversion for Volume Design

Abstract: The article proposes a method for creating digital twins from freehand sketches for facade design, converting 2D designs to 3D volumes, and integrating these designs into real-world GIS systems. It outlines a process that involves generating 2D exterior images from sketches using generative AI (Gemini 1.5 Pro), converting these 2D images into 3D models with TriPo, and creating design drawings with SketchUp. Additionally, it describes a method for creating 3D exterior images using GauGAN, all for the purpose of construction exterior evaluation. The paper also discusses generating BIM data using generative AI, converting BIM data (in IFC file format) to GeoTiff, and displaying this information in GIS using QGIS software. Moreover, it suggests a method for generating digital twins with SketchUp to facilitate digital design information sharing and simulation within a virtual space. Lastly, it advocates for a cost-effective AI system designed for small and medium-sized construction companies, which often struggle to adopt BIM, to harness the advantages of digital twins.

Author 1: Kohei Arai

Keywords: BIM; AI; GIS; digital twins; metaverse; generative AI; GauGAN; TriPo; SketchUp; IFC format; GeoTiff

PDF

Paper 10: Marked Object-Following System Using Deep Learning and Metaheuristics

Abstract: This paper presents a deep learning methodology for a marked object-following system that incorporates the YOLOv8 (You Only Look Once version 8) object identification model and an inversely proportional distance estimation algorithm. The primary aim of this study is to develop a marked object-following algorithm capable of autonomously tracking a designated marker while maintaining a suitable distance through advanced computer vision techniques. In this study, a marked object is defined as an object that is explicitly labeled, tagged, or physically marked for identification, typically using visible markers such as QR codes, stickers, or distinct added features. Central to the system’s functionality is the YOLOv8 model, which detects objects and generates bounding boxes around identified target classes in real-time. The proposed marked object-following algorithm utilizes the distance estimation method, which leverages fluctuations in the bounding box width to determine the relative distance between the observed user and the camera. A pathfinding algorithm was created using tabu search and a-star to avoid obstacle and generate a path to continue following the marker object. Furthermore, the system’s efficacy was assessed using critical performance metrics, including the F1-score and Precision-Recall. The YOLOv8 model attained an F1-score of 0.95 at a confidence threshold of 0.461 and a mean Average Precision (mAP) of 0.961 at an IoU threshold of 0.5 for all target classes. These results indicate a high level of accuracy in object detection and tracking. However, it is important to note that this algorithm has close door and controlled environments.

Author 1: Ken Gorro
Author 2: Elmo Ranolo
Author 3: Lawrence Roble
Author 4: Rue Nicole Santillan
Author 5: Anthony Ilano
Author 6: Joseph Pepito
Author 7: Emma Sacan
Author 8: Deofel Balijon

Keywords: Object detection; YOLOv8; distance estimation; A-star; tabu search

PDF

Paper 11: Hawk-Eye Deblurring and Pose Recognition in Tennis Matches Based on Improved GAN and HRNet Algorithms

Abstract: In tennis matches, the Hawk-eye system causes blurry trajectory judgment and low accuracy in player posture recognition due to rapid movement and complex backgrounds. Therefore, the research improves the backbone network and iterative attention feature fusion mechanism of deblur generative adversarial network version. At the same time, Ghost, Sandglass module, and coordinate attention mechanism are used to optimize the high-resolution network, and a new model for deblurring and pose recognition of Hawk-eye images in tennis matches is proposed by integrating the improved generative adversarial network and high-resolution network. The new model achieved an information entropy value of 11.2, a peak signal-to-noise ratio of 29.74 decibels, a structural similarity of 0.89, a minimum parameter size of 4.53, and a running time of 0.25 seconds on the tennis tracking dataset and the Max Planck Society human posture dataset, which was superior to current advanced models. The highest accuracy of deblurring and pose recognition for the model under different lighting intensities was 92.44%, and the highest improvement rate of video frame quality was 18%. From this, the model has significant advantages in deblurring effect, posture recognition accuracy, parameter quantity, and running time, and has high practical application potential. It can provide an advanced theoretical reference for tennis match refereeing and technical training.

Author 1: Weixin Zhao

Keywords: DeblurGANv2; HRNet; tennis; hawk-eye system; deblurring; pose recognition

PDF

Paper 12: A Highly Functional Ensemble of Improved Chaos Sparrow Search Optimization Algorithm and Enhanced Sun Flower Optimization Algorithm for Query Optimization in Big Data

Abstract: Numerous systems have to provide the highest level of performance feasible to their users due to the present accessibility of enormous datasets and scalability needs. Efficiency in big data is measurable in terms of the speed at which queries are executed physically. It is too demanding on big data for queries to be executed on time to satisfy users' needs. The query optimizer, one of the critical parts of big data that selects the best query execution plan and subsequently influences the query execution duration, is the primary focus of this research. Therefore, a well-designed query enables the user to obtain results in the required time and enhances the credibility of the associated application. This research suggested an enhanced query optimizing method for big data (BD) utilizing the ICSSOA-ESFOA algorithm (Improved Chaos Sparrow Search Optimization Algorithm- Enhanced Sun Flower Optimization algorithm) with HDFS Map Reduce to avoid the challenges associated with the optimization of queries. The essential features are extracted by employing the ResNet50V2 approach. Effective data arrangement is necessary for making sense of large and complex datasets. For this purpose, we ensemble Density-Based Spatial Clustering of Applications with Noise (DBSCAN) and Improved Spectral Clustering (ISC). The experimental findings demonstrate a significant benefit of the proposed strategy over the present optimization of the queries paradigm, and the proposed approach obtains less execution time and memory consumption. The experimental results show that the proposed strategy significantly outperforms the current optimization paradigm, reaching 99.5% accuracy, 29.4 seconds of execution time, and 450 MB less memory use.

Author 1: Mursubai Sandhya Rani
Author 2: N. Raghavendra Sai

Keywords: Big data (BD); query optimization; Improved Chaos Sparrow Search Optimization Algorithm (ICSSOA); Enhanced Sun Flower Optimization Algorithm (ESOA); ResNet50V2; DBSCAN

PDF

Paper 13: IT Spin-Offs Challenges in Developing Countries

Abstract: IT-enabled spin-off ventures in developing countries’ higher learning institutions have the potential to transform academic research into commercially viable products, thereby fostering economic and technological progress. However, practical implementation faces significant challenges, particularly in conflict areas, such as limited resources, socio-political instability, skill gaps, weak intellectual property laws, and inadequate frameworks for protecting innovation. Objective: This study aims to mitigate these challenges by proposing a strategic framework that leverages universities' available resources to promote IT-enabled spin-offs. This framework addresses barriers and converts challenges into opportunities. Methods: This case study focused on higher learning institutions in developing countries. Specifically, this study examines the unique constraints faced by Palestinian higher learning institutions in conflict zones in order to design a tailored IT-enabled spin-off framework. Results: The proposed framework aligns with the National Development Plan and offers pathways for universities to overcome practical barriers. It emphasizes transforming research output into sustainable IT spin-off ventures that support entrepreneurship and innovation. Conclusions: This study highlights the critical need for a new strategic framework for higher learning institutions that incorporates IT-enabled spinoffs as a guiding principle to promote innovation and entrepreneurship. The proposed framework addresses current gaps and provides actionable solutions for advancing sustainable development in conflict-affected regions.

Author 1: Mahmoud M. Musleh
Author 2: Ibrahim Mohamed
Author 3: Hasimi Sallehudin
Author 4: Hussam F. Abushawish

Keywords: IT spin-off framework; higher learning; IT challenges; spin-off; framework; developing countries; entrepreneurship; innovation

PDF

Paper 14: Multi-Factors Analysis Using Visualizations and SHAP: Comprehensive Case Analysis of Tennis Results Forecasting

Abstract: Explainable Artificial Intelligence (XAI) enhances interpretability in data-driven models, providing valuable insights into complex decision-making processes. By ensuring transparency, XAI bridges the gap between advanced Artificial Intelligence (AI) techniques and their practical applications, fostering trust and enabling data-informed strategies. In the realm of sports analytics, XAI proves particularly significant, as it unravels the multifaceted nature of factors influencing athletic performance. This work uses a rich data analysis flow that includes descriptive, predictive, and prescriptive analysis for the tennis match outcomes. Descriptive analysis uses XAI techniques such as SHAP (SHapley Additive exPlanations) with diverse factors such as physical, geographical, surface level and skill disparities. Top players are ranked; the trend of country-wise winning is presented for the last many decades. Correlation analysis presents inter-dependence of factors. Correlation analysis presents inter-dependence of factors. Predictive analysis makes use of machine learning models, the highest overall accuracy of 80% according to the K-Nearest Neighbors classifier. Lastly, prescriptive analysis recommends specific details which can be helpful for players and coaches as well as for overall strategies planning and performance enhancement. The research underscores the significance of AI-driven insights in sports analytics, particularly for a fast-paced and strategic sport like tennis. By leveraging advanced data analytics methods, this study offers a nuanced understanding of the interplay between player attributes, match contexts, and historical trends, paving the way for enhanced performance and informed strategic planning in professional tennis.

Author 1: Yuan Zhang

Keywords: Artificial intelligence; data analytics; machine learning; match result prediction; XAI; SHAP

PDF

Paper 15: Exploring Diverse Conventional and Deep Linguistic Features for Sentiment Analysis of Online Content

Abstract: Social media has changed the world by providing the facility to common person to share their views and generate their own content, known as Users Generated Content (UGC). Due to huge volume of UGC data being created at great velocity, so to analysis this big data, latest AI (Artificial Intelligence) and its sub-domain NLP (Natural Language Processing) are being used. Sentiment analysis of online content is an active research area due to its vast applications in business for review analysis, social and political issues. In this research study, we aim to carry out sentiment analysis of online content by exploring conventional features like Term Frequency – Inverse Document Frequency (TF-IDF), Count-Vectorization, and state of the art word embeddings based word2vec. Extensive exploratory data analysis has been carried out using the latest data visualization approaches. The main novelty lies in the application of unique and diverse machine learning algorithms on social media datasets and the results evaluation using standard performance evaluation measures reveal that the word2vec using Quadratic Discriminant analysis-based classifier show optimal results.

Author 1: Yajun Tang

Keywords: Artificial intelligence; sentiment analysis; machine learning; word embeddings; natural language programming

PDF

Paper 16: An AI-Driven Approach for Advancing English Learning in Educational Information Systems Using Machine Learning

Abstract: In current era of globalization, English language learning is important as it has become a global language and helps people to communicate from various regions and languages. For vocational students whose main aim is to get skills and get employed, learning English for communication is important. We here present a proposed framework for learning English language which can become a foundation for a complete Artificial Intelligence (AI) based system for help and guidance to the educators. This study explores the use of diverse Natural Language Processing (NLP) techniques to predict various grammatical aspects of English language content especially focused on tense prediction which lay the foundation of English content. Textual features of Bag of words (BoW) which considers each word as a separate token and Term Frequency –Inverse Document Frequency (TF-IDF) are explored. For both diverse features, the shallow machine learning models of Support Vector Machine (SVM) and Multinomial Naïve Bayes are applied. Moreover, the ensemble models based on Bagging and Calibrated are applied. The results reveal that BoW model input for SVM and Bagging technique using TF-IDF shows optimal results with high accuracy of 90% and 89% respectively. This empirical analysis confirms that such models can be integrated with web or android based systems which can be helpful for learners of English language.

Author 1: Xue Peng
Author 2: Yue Wang

Keywords: Artificial intelligence; information system; machine learning; English language learning; natural language processing

PDF

Paper 17: Investigating Immersion and Presence in Virtual Reality for Architectural Visualization

Abstract: The architecture industry increasingly relies on virtual reality (VR) for architectural visualization, yet there is a critical issue of insufficient user involvement in the design process. This study investigates the sense of immersion and presence in the virtual environment among 60 Malaysian participants aged 20 to 40. The study utilized a 1000 sq. ft. apartment with three bedrooms and two bathrooms, was replicated in a 3D model based on real-world references. Our findings show that participants were moderately immersed in the virtual environment (M = 4.86), but the lack of sense of touch, lack of detail, and interactivity within the virtual environment affected their sense of immersion in VR for architectural visualization. This study has enhanced our understanding of human-computer interaction in VR, specifically for architectural visualization, and has emphasized the importance of improving these aspects to create more effective architectural visualization user experiences.

Author 1: Athira Azmi
Author 2: Sharifah Mashita Syed Mohamad

Keywords: Virtual environment; virtual reality; human-computer interaction; architectural visualization; sense of presence

PDF

Paper 18: Data Mining MRO-BP Network-Based Evaluation Effectiveness of Music Teaching

Abstract: This study addresses the need for data analysis in evaluating the teaching outcomes of higher music education. It proposes a solution using data-driven algorithms to measure and analyze these outcomes. This study focuses on the issue of measuring and evaluating the outcomes of music education teaching. It analyzes the process of measuring and assessing these outcomes, designs a program for doing so, and introduces key technologies such as music education teaching process analysis, measurement of music teaching outcomes, construction of an assessment model for music teaching outcomes, and application of the assessment model. The study selects teaching content, practical skills, and social practice ability as the three aspects to evaluate. The results demonstrate that this method achieves higher assessment accuracy and requires less time, effectively addressing the challenge of measuring and evaluating the teaching outcomes of higher music education using big data. The findings demonstrate that the technique exhibits a high level of assessment accuracy and is less time-consuming. Additionally, it effectively addresses the challenge of measuring and evaluating the teaching accomplishments in higher music education from the viewpoint of big data.

Author 1: Yifan Fan

Keywords: Mushroom propagation optimisation algorithm; BP neural network; higher music education teaching outcomes measurement; algorithm evaluation

PDF

Paper 19: Employing Data-Driven NOA-LSSVM Algorithm for Indoor Spatial Environment Design

Abstract: This study aims to enhance the precision and efficiency of indoor spatial design for college physical bookstores in the context of the new media environment. To achieve this, a novel intelligent analysis model was developed by integrating the Navigator Optimization Algorithm (NOA) with the Least Squares Support Vector Machine (LSSVM). The research analyzes the relationship between the new media environment and bookstore design, identifies key design principles, and establishes performance metrics. The proposed NOA-LSSVM model optimizes design parameters by utilizing a hybrid convergence-divergence search mechanism, achieving improved accuracy and computational efficiency. A case study of Jilin Jianzhu University's bookstore was conducted to evaluate the model's performance. The NOA-LSSVM model was compared with three other optimization algorithms: the Flower Pollination Algorithm (FPA), Whale Optimization Algorithm (WOA), and Sine Cosine Algorithm (SCA). Results showed that the NOA-LSSVM model achieved superior accuracy, with a Mean Absolute Percentage Error (MAPE) of 2.9, significantly lower than FPA (4.6), WOA (3.8), and SCA (4.2). Additionally, the model exhibited faster convergence and enhanced design efficiency, optimizing the bookstore's functional zones and spatial layout to balance dynamic and quiet areas effectively. In conclusion, the NOA-LSSVM model demonstrates a robust capability to optimize indoor spatial design in the new media environment, outperforming traditional methods in accuracy and practicality. This study provides valuable insights for integrating intelligent algorithms into spatial design processes, with the potential for broader applications in other commercial or educational spaces. Future research should focus on extending the model's generalizability and incorporating advanced media technologies for enhanced user experiences.

Author 1: Di Wang
Author 2: Hui Ma
Author 3: Tingting Lv

Keywords: New media environments; data-driven algorithms; indoor spatial environment design; mariner optimization method

PDF

Paper 20: Enhancing Customer Churn Prediction Across Industries: A Comparative Study of Ensemble Stacking and Traditional Classifiers

Abstract: Predicting customer churn is essential in sectors such as banking, telecommunications, and retail, where retaining existing customers is more cost-effective than acquiring new ones. This paper proposes an enhanced ensemble stacking methodology to improve the prediction performance of ensemble methods. Classic ensemble classifiers and individual models are undergoing enhancements to enhance their sector-wide generalisation. The proposed ensemble stacking method is compared with well-known ensemble classifiers, including Random Forest, Gradient Boosting Machines (GBMs), AdaBoost, and CatBoost, alongside single classifiers such as Logistic Regression (LR), Decision Trees (DT), Naive Bayes (NB), Support Vector Machines (SVM), and Multi-Layer Perceptron. Performance evaluation employs accuracy, precision, recall, and AUC-ROC metrics, utilising datasets from telecom, retail, and banking sectors. This study highlights the importance of investigating ensemble stacking within these three business entities, given that each sector presents distinct challenges and data patterns related to customer churn prediction. According to the results, when compared to other ensemble approaches and single classifiers, the ensemble stacking method achieves better generality and accuracy. The stacking method uses a meta-learner in conjunction with numerous base classifiers to improve model performance and make it adaptable to new domains. This study proves that the ensemble stacking method can accurately anticipate customer turnover and can be used in different industries. It gives firms a great way to keep their clients.

Author 1: Nurul Nadzirah bt Adnan
Author 2: Mohd Khalid Awang

Keywords: Customer churn; single classifier; ensemble classifier; stacking; accuracy

PDF

Paper 21: Hotspots and Insights on Quality Evaluation of Study Tours: Visual Analysis Based on Bibliometric Methodology

Abstract: In this paper, taking 474 articles about quality evaluation of study tours in Web of Science (WOS) database as the research object, quantitatively analyze them with the help of CiteSpace 6.3.R1 software and excel data statistics, and analyze the impact of the literature data, authors' cooperation network, issuing institutions, journal distribution, and keywords' co-occurrence, clustering, and emergence factors, combined with time interval in-depth analysis and prediction, so as to present the research results in the form of visualized knowledge map. The results of the study show that the field of quality evaluation of research and study tourism an interdisciplinary field involving innovative research with multidisciplinary integration. During the decade of 2015-2024, it has experienced three stages of starting and exploration (2015-2018), rapid growth and diversification (2019-2021), and adjustment and maturity (2022-2024). From the viewpoint of authors and issuing organizations, authors are mostly independent research and have not yet formed a clustering research network. Research hotspots from the theoretical system construction and model development, empirical analysis, gradually shifted to user behavior analysis and recommendation system research. The future tends to research on research and learning integration intelligent decision-making, research and learning industry economy, environmental tourism practice and risk management.

Author 1: Meihua Deng

Keywords: Research tourism; tourism quality evaluation; visualization analysis

PDF

Paper 22: Internet of Things (IoT) Driven Logistics Supply Chain Management Coordinated Response Mechanism

Abstract: This study explores the development of an IoT-driven logistics supply chain coordination and response mechanism aimed at achieving real-time information sharing, precise forecasting, and rapid decision-making among supply chain nodes. By employing a hierarchical system construction method, SQL database techniques for data management, and an evaluation model combining AHP and entropy methods, the study proposes a robust framework for improving supply chain efficiency and adaptability. The results demonstrate that IoT technology significantly enhances supply chain transparency, resource allocation, and operational efficiency while reducing risks and costs. The proposed mechanism facilitates dynamic adjustments to market changes and unexpected disruptions, fostering a resilient and collaborative supply chain network. This research provides a foundational basis for the integration of IoT in modern supply chains and offers insights into advancing intelligent logistics systems, with implications for improving global competitiveness in the evolving digital economy.

Author 1: Chong Li

Keywords: IoT; logistics supply chain; management coordination; response mechanism

PDF

Paper 23: Big Data Analytics of Knowledge and Skill Sets for Web Development Using Latent Dirichlet Allocation and Clustering Analysis

Abstract: Web development is a data-centric field and fundamental component of data science. The advent of big data analytics has significantly transformed the processes, knowledge domains, and competencies associated with Web development. Accordingly, educational programs must adjust to contemporary advancements by initially determining the abilities required for big data web developers to satisfy industry demands and adhere to current trends. This study aims to identify the knowledge areas and abilities essential for big data analytics and to create a taxonomy by correlating these competences with currently popular tools in web development. A mixed method consisting of semi-automatic and clustering methods is proposed for the semantic analysis of the text content of online job advertisements associated with the development of big data web applications. This methodology uses Latent Dirichlet Allocation (LDA), a probabilistic topic modeling tool, to uncover hidden semantic structures within a precisely specified textual corpus and average linkage hierarchical clustering as a clustering analysis technique for web developers. The results of this study are a web development competency map which is expected to help evaluate and improve the knowledge, qualifications and skills of IT professionals being hired. It helps to identify the roles and competencies of professionals in the company’s personnel recruitment process; and meet industry skill requirements through web development education programs. The competency map consists of knowledge domains, skills and essential tools for web development such as basic knowledge, frameworks, design and user experience, database design, web development, cloud computing and other soft skills. Furthermore, the proposed model can be extended to several types of jobs in the IT sector.

Author 1: Karina Djunaidi
Author 2: Dine Tiara Kusuma
Author 3: Rahma Farah Ningrum
Author 4: Puji Catur Siswipraptini
Author 5: Dina Fitria Murad

Keywords: Big data analytics; hierarchical clustering; Latent Dirichlet Allocation; web development; knowledge; skill

PDF

Paper 24: Optimizing Multi-Dimensional SCADA Report Generation Using LSO-GAN for Web-Based Applications

Abstract: This paper addresses the challenges of custom-generating multi-dimensional data SCADA (Supervisory Control And Data Acquisition) reports using web technologies. To improve efficiency, reduce maintenance costs, and enhance scalability, the paper proposes a custom generation method based on the LSO-GAN (Light Spectrum Optimizer - Generative Adversarial Network) model. The study begins by analyzing the requirements for multi-dimensional SCADA reports and proposes a web-based design scheme. The LSO algorithm is employed to optimize the GAN model, enabling efficient generation of customizable SCADA reports. The proposed LSO-GAN model was validated using relevant SCADA data, with experimental results showing that the method outperformed other models in terms of accuracy and generation efficiency. Specifically, the LSO-GAN model achieved an RMSE of 14.98 and a MAPE of 0.93, surpassing traditional models such as Conv-LSTM and FC-LSTM. The custom report generation method based on LSO-GAN significantly improves the customization and generation of multi-dimensional data SCADA reports, demonstrating superior performance in both accuracy and operational efficiency.

Author 1: Fanxiu Fang
Author 2: Guocheng Qi
Author 3: Haijun Cao
Author 4: He Huang
Author 5: Lingyi Sun
Author 6: Jingli Yang
Author 7: Yan Sui
Author 8: Yun Liu
Author 9: Dongqing You
Author 10: Wenyu Pei

Keywords: Web technologies; SCADA systems; report customisation; spectral optimisation algorithms; adversarial generative networks

PDF

Paper 25: Optimizing Decentralized Exam Timetabling with a Discrete Whale Optimization Algorithm

Abstract: In recent years, there has been increasing interest in intelligent optimization algorithms, such as the Whale Optimization Algorithm (WOA). Initially proposed for continuous domains, WOA mimics the hunting behavior of humpback whales and has been adapted for discrete domains through modifications. This paper presents a novel discrete Whale Optimization Algorithm approach, integrating the strengths of population-based and local-search algorithms for addressing the examination timetabling problem, a significant challenge many educational institutions face. This problem remains an active area of research and, to the authors' knowledge, has not been adequately addressed by the WOA algorithm. The method was evaluated using real-world data from the first semester of 2023/2024 for faculties at the Universiti of Sarawak, Malaysia. The problem incorporates standard and faculty-specified constraints commonly encountered in real-world scenarios, accommodating online and physical assessments. These constraints include resource utilization, exam spread, splitting exams for shared and non-shared rooms, and period preferences, effectively addressing the diverse requirements of faculties. The proposed method begins by generating an initial solution using a constructive heuristic. Then, several search methods were employed for comparison during the improvement phase, including three Variable Neighborhood Descent (VND) variations and two modified WOA algorithms employing five distinct neighborhoods. These methods have been rigorously tested and compared against proprietary heuristic-based software and manual methods. Among all approaches, the WOA integrated with the iterative threshold-based VND approach outperforms the others. Furthermore, a comparative analysis of the current decentralized approach, decentralized with re-optimization, and centralized approaches underscores the advantages of centralized scheduling in enhancing performance and adaptability.

Author 1: Emily Sing Kiang Siew
Author 2: San Nah Sze
Author 3: Say Leng Goh

Keywords: Examination timetabling; discrete whale optimization algorithm; variable neighborhood descent; capacitated; decentralized

PDF

Paper 26: Fusion of Multimodal Information for Video Comment Text Sentiment Analysis Methods

Abstract: Sentiment analysis of video comment text has important application value in modern social media and opinion management. By conducting sentiment analysis on video comments, we can better understand the emotional tendency of users, optimise content recommendation, and effectively manage public opinion, which is of great practical significance to the push of video content. Aiming at the current video comment text sentiment analysis methods problems such as understanding ambiguity, complex construction, and low accuracy. This paper proposes a sentiment analysis method based on the M-S multimodal sentiment model. Firstly, briefly describes the existing methods of video comment text sentiment analysis and their advantages and disadvantages; then it studies the key steps of multimodal sentiment analysis, and proposes a multimodal sentiment model based on the M-S multimodal sentiment model; finally, the efficiency of the experimental data from the Communist Youth League video comment text was verified through simulation experiments. The results show that the proposed model improves the accuracy and real-time performance of the prediction model, and solves the problem that the time complexity of the model is too large for practical application in the existing multimodal sentiment analysis task of the video comment text sentiment analysis method, and the interrelationships and mutual influences of the multimodal information are not considered.

Author 1: Jing Han
Author 2: Jinghua Lv

Keywords: Video commentary text sentiment analysis; multimodal information fusion; M-S multimodal sentiment model; convolutional neural network

PDF

Paper 27: Enhancing Stock Market Forecasting Through a Service-Driven Approach: Microservice System

Abstract: Predicting stock market is a difficult task that involves not just a knowledge of financial measures but also the ability to assess market patterns, investor sentiment, and macroeconomic factors that can affect the movement of stock prices. Traditional stock recommendation systems are built as monolithic applications, with all components closely coupled within a single codebase. While these systems are functional, yet they are difficult integrating several services and aggregating data from diverse sources due to their lack of scalability and extensibility. A service-driven approach is needed to manage the growing complexity, diversity, and speed of financial data processing. However, microservice architecture has become a useful solution across multiple sectors, particularly in stock systems. In this paper, we design and build a stock market forecasting system based on the microservice architecture that uses advanced analytical approaches such as machine learning, sentiment analysis, and technical analysis to anticipate stock prices and guide informed investing choices. The results demonstrate that the proposed system successfully integrates multiple financial analysis services while maintaining scalability and adaptability due to its microservice architecture. The system successfully retrieved financial metrics and calculated key technical indicators like RSI and MACD. Sentiment analysis detected positive sentiment in Saudi Aramco's Q3 2021 report, and the LSTM model achieved strong prediction results with an MAE of 0.26 and an MSE of 0.18.

Author 1: Asaad Algarni

Keywords: Stock market; microservice architecture; deep learning; technical indicators; sentiment analysis

PDF

Paper 28: Improved Whale Optimization Algorithm with LSTM for Stock Index Prediction

Abstract: After the COVID-19 pandemic, the global economy began to recover. However, stock market fluctuations continue to affect economic stability, making accurate predictions essential. This study proposes an Improved Whale Optimization Algorithm (IWOA) to optimize the parameters of the Long Short-Term Memory (LSTM) model, thereby enhancing stock index predictions. The IWOA improves upon the traditional Whale Optimization Algorithm (WOA) by integrating logistic chaotic mapping to increase population diversity and prevent premature convergence. Additionally, it incorporates a dynamic adjustment mechanism to balance global exploration and local exploitation, thus boosting optimization performance. Experiments conducted on five representative global stock indices demonstrate that the IWOA-LSTM model achieves higher accuracy and reliability compared to WOA-LSTM, LSTM, and RNN models. This highlights its value in predicting complex time-series data and supporting financial decision-making during economic recovery.

Author 1: Yu Sun
Author 2: Sofianita Mutalib
Author 3: Liwei Tian

Keywords: Long short-term memory network; chaotic mapping; dynamic adjustment mechanism; improved whale optimization algorithm; financial time series forecasting

PDF

Paper 29: Multinode LoRa-MQTT of Design Architecture and Analyze Performance for Dual Protocol Network IoT

Abstract: LoRaWAN networks and large places do not support Wi-Fi for multiple points. An architecture that offers dual networks to alter their supporting networks is needed for IoT device installation. The novelty in this research is that designing an architecture for multimode LoRa-MQTT with a mechanism for testing LoRa data transmission with different delays and Wireshark for testing Wi-Fi network QoS on MQTT is necessary. This hour-long LoRa network experiment shows that the End-Node can only receive one data at a time. One data set will be received if several data sets are obtained due to conflict. The second experiment showed data barely reached 70%. The signal strength or RSSI, and the node that sent the data initially decide the data received from a given node, some seconds apart, towards tested QOS with excellent packet loss, 21 ms delay, 50,616 bytes/s throughput, and 0.1426 jitter. Avoid data conflicts and loss by utilizing fewer nodes or adding end nodes in this experiment. The network service is excellent. According to this study, LoRa and MQTT can work well together. This approach could solve Internet of Things communication concerns, especially in large places that are LoRaWAN-inaccessible and Wi-Fi networks are limited.

Author 1: Rizky Rahmatullah
Author 2: Hongmin Gao
Author 3: Ryan Prasetya Utama
Author 4: Puput Dani Prasetyo Adi
Author 5: Jannat Mubashir
Author 6: Rachmat Muwardi
Author 7: Widar Dwi Gustian
Author 8: Hanifah Dwiyanti
Author 9: Yuliza

Keywords: LoRa; MQTT; multinode; QOS; LoRaWAN

PDF

Paper 30: Machine Learning-Based Fifth-Generation Network Traffic Prediction Using Federated Learning

Abstract: The rapid development and advancement of 5G technologies and smart devices are associated with faster data transmission rates, reduced latency, more network capacity, and more dependability over 4G networks. However, the networks are also more complex due to the diverse range of applications and technologies, massive device connectivity, and dynamic network conditions. The dynamic and complex nature of the 5G networks requires advanced and accurate traffic prediction methods to optimize resource allocation, enhance the quality of service, and improve network performance. Hence, there is a growing demand for training methods to generate high-quality predictions capable of generalizing to new data across various parties. Traditional methods typically involve gathering data from multiple base stations, transmitting it to a central server, and performing machine learning operations on the collected data. This work suggests a hybrid model of Long Short Term Memory (LSTM), Gated Recurrent Unit (GRU), and federated learning applied to 5G network traffic prediction. The model is assessed on one-step predictions, comparing its performance with standalone LSTM and GRU models within a federated learning environment. In evaluating the predictive performance of the proposed federated learning architecture compared to centralized learning, the federated learning approach results in lower Root Mean Square error (RMSE) and Mean Absolute Errors (MAE) and a 2.25 percent better Coefficient of Determination (R squared).

Author 1: Mohamed Abdelkarim Nimir Harir
Author 2: Edwin Ataro
Author 3: Clement Temaneh Nyah

Keywords: 5G Mobile network; machine learning; federated learning; parallel hybrid LSTM+GRU; network traffic prediction; centralized learning; dynamic network condition

PDF

Paper 31: CN-GAIN: Classification and Normalization-Denormalization-Based Generative Adversarial Imputation Network for Missing SMES Data Imputation

Abstract: Quality data is crucial for supporting the management and development of SMES carried out by the government. However, the inability of SMES actors to provide complete data often results in incomplete dataset. Missing values present a significant challenge to producing quality data. To address this, missing data imputation methods are essential for improving the accuracy of data analysis. The Generative Adversarial Imputation Network (GAIN) is a machine learning method used for imputing missing data, where data preprocessing plays an important role. This study proposes a new model for missing data imputation called the Classification and Normalization-Denormalization-based Generative Adversarial Imputation Network (CN-GAIN). The study simulates different patterns of missing values, specifically MAR (Missing at Random), MCAR (Missing Completely at Random), and MNAR (Missing Not at Random). For comparison, each missing value pattern is processed using both the CN-GAIN and the base GAIN methods. The results demonstrate that the CN-GAIN model outperforms GAIN in predicting missing values. The CN-GAIN model achieves an accuracy of 0.0801% for the MCAR category and shows a lower error rate (RMSE) of 48.78% for the MNAR category. The mean error (MSE) for the MAR category is 99.60%, while the deviation (MAE) for the MNAR category is 70%.

Author 1: Antonius Wahyu Sudrajat
Author 2: Ermatita
Author 3: Samsuryadi

Keywords: Missing values; GAIN method; normalization-denormalization; imputation; UMKM data

PDF

Paper 32: An Agile Approach for Collaborative Inquiry-Based Learning in Ubiquitous Environment

Abstract: The use of collaborative inquiry-based learning has been prevalent in educational contexts particularly in science education. Using such collaborative environments, learners can increase their engagement, knowledge and critical thinking skills about science. With the advancement of technologies, ubiquitous learning environments have been designed for facilitating learning in real-time contexts. Over the past few years, agile-based approaches have been implemented at higher education for inquiry-based learning activities. However, there is a lack of studies found that focuses on agile-based approach for ubiquitous collaborative inquiry learning activities at K-12 education level. Therefore, this study presents the ScrumBan Ubiquitous Inquiry Framework (SBUIF), for inquiry-based learning activities at K-12 education level. For this purpose, an application uASK has been developed on the proposed framework, SBUIF. For the evaluation purposes, computer-supported collaborative learning (CSCL) affordances along with micro and meso levels of the M3 evaluation framework has been applied. An experiment was conducted for the evaluation of uASK application in comparison with the Trello application, involving 205 to 127 seventh-grade students. Results demonstrated that uASK learners achieved higher scores as compared with Trello participants. Further, survey results indicated higher levels of engagement, satisfaction, and enjoyment among uASK users. The study concludes that uASK offers significant advantages over Trello in fostering collaborative inquiry-based learning activities in ubiquitous environment.

Author 1: Bushra Fazal Khan
Author 2: Sohaib Ahmed

Keywords: K12 education; agile; ubiquitous; collaborative learning; inquiry based learning

PDF

Paper 33: M-COVIDLex: The Construction of a Domain-Specific Mixed Code Sentiment Lexicon

Abstract: Sentiment lexicons serve as essential components in lexicon-based sentiment analysis models. Research on sentiment analysis based on the Malay lexicon indicates that most existing sentiment lexicons for this language are developed from official text corpora, general domain social media text corpora, or domain-specific social media text corpora. Nonetheless, none of the current sentiment lexicons adequately complement the corpus utilized in this study. The rationale is that words in established sentiment lexicons may convey different sentiments compared to those in this paper’s corpus, as the strength and sentiment of words are context-dependent, influenced by varying terminology or jargon across domains, and words may not share the same sentiment across multiple domains. This paper proposes the construction of a domain-specific mixed-code sentiment lexicon, termed M-COVIDLex, through the integration of corpus-based and dictionary-based techniques, utilizing seven Malay part-of-speech tags, and enhancing Malay part-of-speech tagging for social media text by introducing a new tag: FOR-POS. The constructed M-COVIDLex is evaluated using two distinct domains of social media text corpus: the specific domain and the general domain. The performance indicates that M-COVIDLex is more appropriate as a sentiment lexicon for analyzing sentiment in a domain-specific social media text corpus, providing valuable insights to governments in assessing the sentiment level regarding the analyzed topic.

Author 1: Siti Noor Allia Noor Ariffin
Author 2: Sabrina Tiun
Author 3: Nazlia Omar

Keywords: Malay social media text; mixed-code sentiment lexicon; sentiment analysis; domain-specific; lexicon-based; informal Malay; Malay part-of-speech; public health emergencies; COVID-19 Malaysia

PDF

Paper 34: Optimized Hybrid Deep Learning for Enhanced Spam Review Detection in E-Commerce Platforms

Abstract: Spam reviews represent a real danger to e-commerce platforms, steering consumers wrong and trashing the reputations of products. Conventional Machine learning (ML) methods are not capable of handling the complexity and scale of modern data. This study proposes the novel use of hybrid deep learning (DL) models for spam review detection and experiments with both CNN-LSTM and CNN-GRU architectures on the Amazon Product Review Dataset comprising 26.7 million reviews. One important finding is that 200k words vocabulary, with very little preprocessing improves the models a lot. Compared with other models, the CNN-LSTM model achieves the best performance with an accuracy of 92%, precision of 92.22%, recall of 91.73% and F1-score of 91.98%. This outcome emphasizes the effectiveness of using convolutional layers to extract local patterns and LSTM layers to capture long-term dependencies. The results also address how high constraints and hyperparameter search, as well as general-purpose represents such as BERT. Such advancements will help in creating more reliable and reliable spam detection systems to maintain consumer trust on e-commerce platforms.

Author 1: Abdulrahman Alghaligah
Author 2: Ahmed Alotaibi
Author 3: Qaisar Abbas
Author 4: Sarah Alhumoud

Keywords: Spam review detection; CNN-LSTM; CNN-RNN; CNN-GRU; big data; deep learning; amazon product review dataset

PDF

Paper 35: Optimization of Fourth Party Logistics Routing Considering Infection Risk and Delay Risk

Abstract: In the context of the rapid development of e-commerce and the increasing demands for logistics services, particularly in the face of challenges posed by public health emergencies, this paper explores how to integrate supply chain resources and optimize delivery processes. It provides an in-depth analysis of the characteristics of the Fourth Party Logistics Routing Optimization Problem (4PLROP) in complex environments, specifically focusing on the impacts of infection risk and delay risk, and proposes a new risk measurement tool. By constructing a mathematical model aimed at minimizing Conditional Value-at-Risk (CVaR) and improved Q-learning algorithm, the study addresses the 4PLROP while considering cost and risk constraints. This approach enhances the efficiency and service quality of the logistics industry, offers effective strategies for 4PL companies in the face of uncertainty, and provides customers with safer and more reliable logistics solutions, contributing to sustainable development.

Author 1: Guihua Bo
Author 2: Sijia Li
Author 3: Mingqiang Yin
Author 4: Mingkun Chen
Author 5: Xin Liu

Keywords: Logistics services; public health emergencies; logistics routing optimization; improved Q-learning algorithm; CVaR; infection risk

PDF

Paper 36: Convolutional Neural Network and Bidirectional Long Short-Term Memory for Personalized Treatment Analysis Using Electronic Health Records

Abstract: Correct precision techniques have far not been introduced for modeling the modality risk in Intensive Care Unit (ICU) patients. Traditional mortality risk prediction techniques effectively extract the data in longitudinal Electronic Health Records (EHRs), that ignore the difficult relationship and interactions among variables and time dependency in longitudinal records. The proposed work, developed the Convolutional Neural Network – Bidirectional-Long Short-Term Memory (CNN-Bi-LSTM) method for personalized treatment analysis using EHR data. The CNN extracts the significant features from relevant features, focused on spatial-based relationships. Then, the Bi-LSTM layer captured the sequential dependencies and temporal relationships in patient histories that are essential to understand the treatment results. The Circle Levy flight – Ladybug Beetle Optimization (CL-LBO) integrates the circle chaotic map and Levy flight process in traditional LBO to select relevant features for classification. The proposed method reached 99.85% accuracy, 99.60% precision, 99.50% recall, 99.55% f1-score, and 99.95% Area Under Curve (AUC) when compared to LSTM.

Author 1: Prasanthi Yavanamandha
Author 2: D. S. Rao

Keywords: Bidirectional-long short-term memory; circle chaotic map; convolutional neural network; electronic medical records; Intensive Care Unit (ICU); ladybug beetle optimization

PDF

Paper 37: Spam Detection Using Dense-Layers Deep Learning Model and Latent Semantic Indexing

Abstract: In the digital age, online shoppers heavily depend on product feedback and reviews available on the corresponding product pages to guide their purchasing decisions. Feedback is used in sentiment analysis, which is helpful for both customers and company management. Spam feedback can have a negative impact on high-quality products or a positive impact on low-quality products. In both cases, the matter is bothersome. Spam detection can be done with supervised or unsupervised learning methods. We suggested two direct methods to detect feedback orientation as ‘spam’ or "not spam", also called "ham," using the deep learning model and the LSI (Latent Semantic Indexing) technique. The first proposed model uses only dense layers to detect the orientation of the text. The second proposed model uses the concept of LSI, an effective information retrieval algorithm that finds the closest text to a provided query, i.e., a list containing spam words. Experimental results of both models using publicly available datasets show the best results (89% accuracy and 89% precision) when compared to their corresponding benchmarks.

Author 1: Yasser D. Al-Otaibi
Author 2: Shakeel Ahmad
Author 3: Sheikh Muhammad Saqib

Keywords: Spam; supervised learning methods; unsupervised learning methods; LSI; dense; deep learning

PDF

Paper 38: A Deep Learning for Arabic SMS Phishing Based on URLs Detection

Abstract: The increasing use of SMS phishing messages in Arab communities has created a major security threat, as attackers exploit these SMS services to steal users' sensitive and financial data. This threat highlights the necessity of designing models to detect SMS messages and distinguish between phishing and non-phishing messages. Given the lack of sufficient previous studies addressing Arabic SMS phishing detection, this paper proposes a model that leverages deep learning models to detect Arabic SMS messages based on the URLs they contain. The focus is on the URL aspect because it is one of the common indicators in phishing attempts. The proposed model was applied to two datasets that were in English, and one dataset was in Arabic. Two datasets were translated from English to Arabic. Three datasets included a number of Arabic SMS messages, mostly containing URLs. Three deep learning models—CNN, BiGRU, and GRU—were implemented and compared. Each model was evaluated using metrics such as precision, recall, accuracy, and F1 score. The results showed that the GRU model achieved the highest accuracy of 95.3% compared to other models, indicating its ability to capture sequential patterns in URLs extracted from Arabic SMS messages effectively. This paper contributes to designing a phishing detection model designed for Arab communities to enhance information security within Arab communities.

Author 1: Sadeem Alsufyani
Author 2: Samah Alajmani

Keywords: Phishing; URL phishing; SMS phishing; GRU; BiGRU; CNN

PDF

Paper 39: Jordanian Currency Recognition Using Deep Learning

Abstract: Automatic Currency Recognition (ACR) has a significant role in various domains, such as assessment of visually impaired people, banking transactions, counterfeit detection, digital transformation, currency exchange, vendor machines, etc. Therefore, developing an accurate ACR system enhances efficiency across several domains. The contribution of this paper is three-fold; it proposed a large dataset of 2799 images and seven denominations for Jordanian currency recognition. The second contribution proposed an efficient multiscale VGG net to recognize Jordanian currency. Third, popular CNN architectures on the proposed dataset will be evaluated, and the result will be compared with the proposed architectures. Four metrics were used in the evaluation. The experimental result showed the accuracy of the proposed Multiscale VGG outperformed VGG16, DenseNet121, ResNet50, and ResNet101 and achieved 99.88%, 99.88%, 99.89%, and 99.98% accuracy, precision, sensitivity, and specificity.

Author 1: Salah Alghyaline

Keywords: Automatic currency recognition; deep learning; VGG

PDF

Paper 40: Foreground Feature-Guided Camouflage Image Generation

Abstract: In the field of visual camouflage, generating a high-quality background image that seamlessly blends with complex foreground objects and diverse background environments is a critical task. When dealing with such complex scenes, the existing techniques have insufficient foreground feature extraction, resulting in insufficient fusion of the generated background image with the foreground objects, making it difficult to achieve the desired camouflage effect. In order to solve this problem and achieve the goal of higher quality visual camouflage effect, this paper proposes a new foreground feature-guided camouflage image generation method (Object Enhancement Module - Diffusion Refinement , OEM-DR), which generates camouflage images by enhancing the foreground features to guide the background. The method firstly designs a new object enhancement module to optimize the attention mechanism of the model, and eliminates the attention weights that have less influence on the output through pruning strategy, so that the model focuses more on the key features of the foreground objects, and thus guides the generation of the background more effectively. Second, a novel detail optimization framework based on diffusion strategy is constructed, which maintains the integrity of the global structure of the image while performing fine optimization processing on the local details of the image. In experiments on standard camouflaged image datasets, the proposed method in this study achieves significant improvement in both FID (Fréchet Inception Distance) and KID (Kullback-Leibler Divergence) evaluation metrics, which verifies the feasibility of the method. This suggests that by strengthening foreground features and detail optimization, the fusion between background images and foreground objects can be effectively improved to achieve higher quality visual camouflage effects.

Author 1: Yuelin Chen
Author 2: Yuefan An
Author 3: Yonsen Huang
Author 4: Xiaodong Cai

Keywords: Camouflage image; foreground features; object enhancement; detail optimization

PDF

Paper 41: Adoption of Generative AI-Enhanced Profit Sharing Digital Systems in MSMEs: A Comprehensive Model Analysis

Abstract: Adopting digital finance solutions is crucial for enhancing efficiency and competitiveness within the financial services industry, particularly for Micro, Small, and Medium Enterprises (MSMEs). This study examines the factors influencing the use and acceptance of a sharing-based digital system enhanced with a Generative AI website (E-Mudharabah), employing the Technology Acceptance Model (TAM) and the Unified Theory of Acceptance and Use of Technology (UTAUT). In this article, the Generative AI-enhanced profit-sharing digital systems called E-Mudharabah. It is a web-based management system facilitating capital management for financiers, consultants, and MSME actors. The research integrates key variables from both models, including Perceived Ease of Use, Perceived Usefulness, Performance Expectancy, Social Influence, Facilitating Conditions, Habit, and Technology Self-Efficacy, to assess their impact on Behavioral Intention and Actual Usage. The study utilizes a quantitative approach, gathering data through surveys and analyzing it using the Partial Least Squares Structural Equation Modeling (PLS-SEM) method. Results indicate significant positive effects of perceived usefulness, performance expectancy, and social influence on the behavioral intention to use E-Mudharabah. The findings underscore the role of user-friendly interfaces and societal acceptance in driving adoption. Perceived Usefulness was the most significant variable influencing Behavioral Intention and Actual Usage (p-value < 0.001). Additionally, Social Influence and Facilitating Conditions were shown to have substantial effects, highlighting the importance of user support and societal acceptance in technology adoption. The research also underscores the role of Technology Self-Efficacy in enhancing user confidence and engagement with the platform. These findings suggest that improving digital finance solutions' perceived benefits and ease of use while fostering a supportive environment can significantly boost their adoption rates.

Author 1: Mardiana Andarwati
Author 2: Galandaru Swalaganata
Author 3: Sari Yuniarti
Author 4: Fandi Y. Pamuji
Author 5: Edward R. Sitompul
Author 6: Kukuh Yudhistiro
Author 7: Puput Dani Prasetyo Adi

Keywords: Digital finance; Generative AI; TAM; UTAUT; MSMEs

PDF

Paper 42: DeepLabV3+ Based Mask R-CNN for Crack Detection and Segmentation in Concrete Structures

Abstract: In order to solve the problem of concrete structure crack detection and segmentation and improve the efficiency of detection and segmentation, this paper proposes a crack detection and segmentation method for concrete structure based on DeepLabV3+ and Mask R-CNN algorithm. Firstly, a crack detection and segmentation scheme is designed by analysing the crack detection and segmentation problem of concrete structure. Secondly, a crack detection method based on Mask R-CNN algorithm is proposed for the crack detection problem of concrete structure. Then, a crack segmentation method based on DeepLabV3+ algorithm is proposed for the crack segmentation problem of concrete structure. Finally, bridge crack image data is used for the crack detection and segmentation of concrete structure. Finally, the concrete structure crack detection and segmentation method is validated and analysed using bridge crack image data. The results show that the Mask R-CNN model has better performance in the localisation and identification of cracks, and the DeepLabV3+ model has higher accuracy and contour extraction integrity in solving the crack segmentation problem.

Author 1: Yuewei Liu

Keywords: DeepLabV3+; Mask R-CNN; concrete structure; crack detection and segmentation; deep learning algorithm

PDF

Paper 43: Multi-Objective Optimization of Construction Project Management Based on NSGA-II Algorithm Improvement

Abstract: In the building industry, one of the key components to ensuring a project's successful completion is multi-objective project management. However, due to its own limitations, the traditional multi-objective management approach for projects is no longer able to meet the requirements of building construction and urgently needs to be improved. This is because the construction industry is becoming more competitive and construction standards are improving. Traditional methods for multi-objective optimization typically involve simply summing multiple objectives with weights, overlooking the interdependencies among these objectives. These methods often get trapped in local optimal solutions and rely heavily on predefined models and parameters, limiting their adaptability to sudden changes during the construction process. Therefore, a multi-objective management approach based on multi-objective genetic algorithm for construction projects is proposed. It enables in-depth analysis and comprehensive optimization of the complex relationships between objectives, leading to more informed decisions. By facilitating rapid iteration and adaptation, it enables timely adjustments and optimizations to ensure that project goals remain consistent in complex and dynamic environments. In the experimental validation, the NSGA-II algorithm achieved a significant accuracy of 0.642 and success rate of 0.504 on the VOT dataset, both of which improved by about 1.0% and 0.6% compared to the comparison algorithm. Experimental results on the TrackingNet dataset revealed that the algorithm achieved an accuracy of 0.791 and a success rate of 0.763, while it still maintained an accuracy of 0.542 and a success rate of 0.763 in the face of occlusion. The enhanced multi-objective genetic algorithm had higher accuracy and success rates. This demonstrates the efficiency and excellence of the multi-objective management optimization approach suggested in this study for building projects. The research results have some application value in the multi-objective optimization of engineering projects.

Author 1: Yong Yang
Author 2: Jinrui Men

Keywords: NSGA-II algorithm; improvement strategy; construction engineering; project management; multi-objective optimization

PDF

Paper 44: Performance Evaluation of Efficient and Accurate Text Detection and Recognition in Natural Scenes Images Using EAST and OCR Fusion

Abstract: Scene texts refer to arbitrary text found in images captured by cameras in real-world settings. The tasks of text detection and recognition are critical components of computer vision, with applications spanning scene understanding, information retrieval, robotics, and autonomous driving. Despite significant advancements in deep learning methods, achieving accurate text detection and recognition in complex images remains a formidable challenge for robust real-world applications. Several factors contribute to these challenges. First, the diversity of text shapes, fonts, colors, and styles complicates detection efforts. Second, the myriad combinations of characters, often with unstable attributes, make complete detection difficult, especially when background interruptions obscure character strokes and shapes. Finally, effective coordination of multiple sub-tasks in end-to-end learning is essential for success. This research aimed to tackle these challenges by enhancing text discriminative representation. This study focused on two interconnected problems: Scene Text Recognition (STR), which involves recognizing text from scene images, and Scene Text Detection (STD), which entails simultaneously detecting and recognizing multiple texts within those images. This research focuses on implementing and evaluating the Efficient and Accurate Scene Text Detector (EAST) algorithm for text detection and recognition in natural scene images. The study aims to compare the performance of three prominent Optical Character Recognition (OCR) techniques—TesseractOCR, PaddleOCR, and EasyOCR. The EAST model was applied to a series of sample test images, and the results were visually represented with bounding boxes highlighting the detected text regions. The inference times for each image were recorded, highlighting the algorithm's efficiency, with average times of 0.446, 0.439, and 0.440 seconds for the respective test images. These results indicate that the EAST algorithm is accurate and operates in real-time, making it suitable for applications requiring immediate text recognition.

Author 1: Vishnu Kant Soni
Author 2: Vivek Shukla
Author 3: S. R. Tandan
Author 4: Amit Pimpalkar
Author 5: Neetesh Kumar Nema
Author 6: Muskan Naik

Keywords: Scene text recognition; optical character recognition; deep learning; feature extraction; scene text detection

PDF

Paper 45: AI-Powered Learning Pathways: Personalized Learning and Dynamic Assessments

Abstract: Integrating artificial intelligence (AI) in education has introduced innovative approaches, particularly in personalized learning and dynamic assessment. Conventional teaching models often struggle to address learners' diverse needs and abilities, underscoring the necessity for AI-driven flexible learning frameworks. This study explores how AI-aided smart learning paths and dynamic assessments enhance learning efficiency by improving knowledge acquisition, optimizing task completion time, and increasing student engagement. A six-week quasi-experimental study was conducted with 200 students, divided into an experimental group using an AI-based learning system and a control group following traditional methods. Pre- and post-tests and engagement analyses were used to evaluate outcomes. The experimental group demonstrated a 25% improvement in performance, completed tasks 25% faster, and showed a 15% increase in engagement compared to the control group. These findings highlight the potential of AI to deliver personalized learning experiences and timely feedback, significantly enhancing student outcomes. Future research should involve larger participant groups across higher educational levels and examine the long-term impact of AI-supported education on students’ knowledge retention and skill reinforcement.

Author 1: Mohammad Abrar
Author 2: Walid Aboraya
Author 3: Rawad Abdulghafor
Author 4: Kabali P Subramanian
Author 5: Yousuf Al Husaini
Author 6: Mohammed Al Husaini

Keywords: AI-powered learning; adaptive learning; dynamic assessments; education technology; personalized learning pathways; student engagement

PDF

Paper 46: Feature Reduction and Anomaly Detection in IoT Using Machine Learning Algorithms

Abstract: Anomaly detection in IoT is a hot topic in cybersecurity. Also, there is no doubt that the increased volume of IoT trading technology increases the challenges it faces. This paper explores several machine-learning algorithms for IoT anomaly detection. The algorithms used are Naïve Bayesian (NB), Support Vector Machine (SVM), Decision Tree (DT), XGBoost, Random Forest (RF), and K-nearest Neighbor (K-NN). Besides that, this research uses three techniques for feature reduction (FR). The dataset used in this study is RT-IoT2022, which is considered a new dataset. Feature reduction methods used in this study are Principal Component Analysis (PCA), Particle Swarm Optimization (PSO), and Gray Wolf Optimizer (GWO). Several assessment metrics are applied, such as Precision (P), Recall(R), F-measures, and accuracy. The results demonstrate that most machine learning algorithms perform well in IoT anomaly detection. The best results are shown in SVM with approximately 99.99% accuracy.

Author 1: Adel Hamdan
Author 2: Muhannad Tahboush
Author 3: Mohammad Adawy
Author 4: Tariq Alwada’n
Author 5: Sameh Ghwanmeh

Keywords: Machine learning; Internet of Things (IoT); anomaly detection; feature reduction; Naïve Bayesian (NB); Support Vector Machine (SVM); Decision Tree (DT); XGBoost; Random Forest (RF); K-Nearest Neighbor (K-NN)

PDF

Paper 47: Network Security Based on GCN and Multi-Layer Perception

Abstract: With the continuous progress of network technology, network security has become a critical issue at present. There are already many network security intrusion detection models, but these detection models still have problems such as low detection accuracy and long interception time of intrusion information. To address these drawbacks, this study utilizes graph convolutional network to optimize multi-layer perceptron. An optimization algorithm based on multi-layer perceptron is innovatively proposed to construct an intrusion detection model. Comparative experiments are conducted on the improved algorithm. The accuracy of the algorithm was 0.98, the F1 value was 0.97, and the detection time was 1.1s. The overall performance was much better than comparison algorithms. Subsequently, the intrusion detection model was applied to network security detection. The detection time was 0.1s, the accuracy was 0.98, and the overall performance outperformed other comparison algorithms. The results demonstrate that the intrusion detection method on the basis of optimized multi-layer perceptron can enhance the detection ability of illegal intrusion information. This study optimizes the performance of detecting illegal network intrusion information, providing a theoretical basis for further development of network security. However, the types of intrusion information in this study are limited and there is still uncertainty. In the future, data augmentation techniques can be used to oversample minority class samples, synthesize new minority class samples, expand sample size, increase detection information, and improve the overall detection performance of the model.

Author 1: Wei Yu
Author 2: Huitong Liu
Author 3: Yu Song
Author 4: Jiaming Wang

Keywords: Network security; graph convolutional network; multi-layer perceptron; intrusion detection model

PDF

Paper 48: An Ensemble Semantic Text Representation with Ontology and Query Expansion for Enhanced Indonesian Quranic Information Retrieval

Abstract: This study explores the effectiveness of an ensemble method for Quranic text retrieval, aimed at improving the relevance and accuracy of verses retrieved for specific themes. The ensemble approach integrates three semantic models—Word2Vec, FastText, and GloVe—through a voting mechanism that considers verse frequency and semantic alignment with the query topics. Testing was conducted on themes such as prayer, zakat, fasting, umrah, and eschatology, reflecting fundamental aspects of Quranic teachings. Results demonstrate that the ensemble method significantly outperforms non-ensemble approaches, achieving an average relevance rate of 88%, compared to individual models (Word2Vec: 75%, FastText: 80%, GloVe: 82%). The ensemble method effectively combines the unique strengths of each model. Word2Vec captures general semantic relationships, FastText handles morphological nuances, and GloVe identifies global contextual patterns. By combining these capabilities, the ensemble approach improves both the quantity and quality of retrieved verses, making it a robust tool for semantic analysis in Quranic studies. This research contributes to the field of computational Islamic studies by demonstrating the practical advantages of ensemble methods for religious text retrieval. It lays the foundation for further advancements, including the integration of deep learning techniques, dynamic query handling, and cross-linguistic analysis. The ensemble method offers a promising framework for supporting more accurate and contextually relevant Quranic studies, promoting a deeper understanding of Islamic teachings through data-driven methodologies.

Author 1: Liza Trisnawati
Author 2: Noor Azah Binti Samsudin
Author 3: Shamsul Kamal Bin Ahmad Khalid
Author 4: Ezak Fadzrin Bin Ahmad Shaubari
Author 5: Sukri
Author 6: Zul Indra

Keywords: Ensemble method; query expansion; ontology; Al-Quran; search engine

PDF

Paper 49: A Review of Reinforcement Learning Evolution: Taxonomy, Challenges and Emerging Solutions

Abstract: Reinforcement Learning (RL) has become a rapidly advancing field inside Artificial Intelligence (AI) and self-sufficient structures, revolutionizing the manner in which machines analyze and make selections. Over the past few years, RL has advanced notably with the improvement of more sophisticated algorithms and methodologies that address increasingly complicated actual-world troubles. This progress has been driven by using enhancements in computational power, the availability of big datasets, and improvements in machine-getting strategies, permitting RL to address challenges across a wide range of industries, from robotics and autonomous driving system to healthcare and finance. The effect of RL is evident in its capacity to optimize selection-making procedures in unsure and dynamic environments. By getting to know from interactions with the environment, RL agents can make decisions that maximize lengthy-time period rewards, adapting to converting situations and enhancing over time. This adaptability has made RL an invaluable tool in situations wherein traditional approaches fall brief, especially in complicated, excessive-dimensional spaces and behind-schedule remarks. This review aims to offer radical information about the current nation of RL, highlighting its interdisciplinary contributions and how it shapes the destiny of AI and autonomous technologies. It discusses how RL affects improvements in robotics, natural language processing, and recreation while exploring its deployment's ethical and practical demanding situations. Additionally, it examines key research from numerous fields that have contributed to RL's development.

Author 1: Ji Loun Tan
Author 2: Bakr Ahmed Taha
Author 3: Norazreen Abd Aziz
Author 4: Mohd Hadri Hafiz Mokhtar
Author 5: Muhammad Mukhlisin
Author 6: Norhana Arsad

Keywords: Artificial intelligence; autonomous systems; decision-making optimization; reinforcement learning; robotics

PDF

Paper 50: Towards Transparent Traffic Solutions: Reinforcement Learning and Explainable AI for Traffic Congestion

Abstract: This study introduces a novel approach to traffic congestion detection using Reinforcement Learning (RL) of machine learning classifiers enhanced by Explainable Artificial Intelligence (XAI) techniques in Smart City (SC). Conventional traffic management systems rely on static rules, and heuristics face challenges in dynamically addressing urban traffic problems' complexities. This study explains the novel Reinforcement Learning (RL) framework integrated with an Explainable Artificial Intelligence (XAI) approach to deliver more transparent results. The model significantly reduces the missing data rate and improves overall prediction accuracy by incorporating RL for real-time adaptability and XAI for clarity. The proposed method enhances security, privacy, and prediction accuracy for traffic congestion detection by using Machine Learning (ML). Using RL for adaptive learning and XAI for interpretability, the proposed model achieves improved prediction and reduces the missing data rate, with an accuracy of 98.10, which is better than the existing methods.

Author 1: Shan Khan
Author 2: Taher M. Ghazal
Author 3: Tahir Alyas
Author 4: M. Waqas
Author 5: Muhammad Ahsan Raza
Author 6: Oualid Ali
Author 7: Muhammad Adnan Khan
Author 8: Sagheer Abbas

Keywords: Reinforcement learning; Explainable Artificial Intelligence (XAI); Smart City (SC); IoT; Machine Learning (ML)

PDF

Paper 51: Strategic Supplier Selection in Advanced Automotive Production: Harnessing AHP and CRNN for Optimal Decision-Making

Abstract: This study presents a novel supplier selection methodology that integrates the Analytic Hierarchy Process (AHP) with a Convolutional Recurrent Neural Network (CRNN) to address the complexities of decision-making in dynamic industrial environments. The AHP component provides a systematic and transparent framework for evaluating many factors, ensuring consistency and minimizing subjective biases in supplier assessment. The Analytic Hierarchy Process (AHP) effectively combines expert knowledge with individual preferences, therefore embodying the human element of decision-making. The CRNN concurrently leverages its ability to process large sequential data, uncover hidden patterns, and assess supplier performance over time. This expertise enhances decision-making by transcending the limitations of traditional analytical methods in managing intricate, multidimensional data. The integration of AHP and CRNN offers a comprehensive evaluation framework, including both objective and subjective factors to enhance effective supplier selection decisions. This approach enhances the long-term sustainability of manufacturing operations by fostering reliable supplier relationships and ensuring access to high-performing suppliers. Experimental validations affirm the efficacy of the suggested approach in promoting sustainable manufacturing systems, highlighting its practical use. The findings demonstrate that the AHP-CRNN framework improves supplier selection criteria and offers prospects for future development and adaptation to address emerging challenges in complex manufacturing environments.

Author 1: Karim Haricha
Author 2: Azeddine Khiat
Author 3: Yassine Issaoui
Author 4: Ayoub Bahnasse
Author 5: Hassan Ouajji

Keywords: Supplier selection; analytic hierarchy process; convolutional recurrent neural network; sustainability; decision-making

PDF

Paper 52: Understanding Art Deeply: Sentiment Analysis of Facial Expressions of Graphic Arts Using Deep Learning

Abstract: Art serves as a profound medium for humans to express and present their thoughts, emotions, and experiences in aesthetically and captivating means. It is like a universal language transcending the limitations of language enabling communication of complex ideas and feelings. Artificial Intelligence (AI) based data analytics are being applied for research domains such as sentiment analysis in which usually text data is analyzed for opinion mining. In this research study, we take art work and apply deep learning (DL) algorithms to classify seven diverse facial expressions in graphics art. For empirical analysis, state of the art deep learning algorithms of Inceptionv3 and pre-trained model of ResNet have been applied on large dataset. Both models are considered revolutionary deep learning architecture allowing for the training of much deeper networks and thus enhancing model performance in various computer vision tasks such as image recognition and classification tasks. The comprehensive results analysis reveals that the proposed methods of ResNet and Inceptionv3 have achieved accuracy as high as 98% and 99% respectively as compared to existing approaches in the relevant field. This research contributes to the fields of sentiment analysis, computational visual art, and human-computer interaction by addressing the detection of seven diverse facial expressions in graphic art. Our approach enables enhanced understanding of user sentiments, offering significant implications for improving user engagement, emotional intelligence in AI-driven systems, and personalized experiences in digital platforms. This study bridges the gap between visual aesthetics and sentiment detection, providing novel insights into how graphic art influences and reflects human emotions by highlighting the efficacy of DL frameworks for real-time emotion detection applications in diverse fields such as human psychological assessment and behavior analysis.

Author 1: Fei Wang

Keywords: Artificial intelligence; deep learning; sentiment analysis; art detection; image processing; convolutional network

PDF

Paper 53: A Hybrid Transformer-ARIMA Model for Forecasting Global Supply Chain Disruptions Using Multimodal Data

Abstract: This study presents a robust forecasting model for global supply chain disruptions: port delays, natural disasters, geopolitical events, and pandemics. An integrated solution combining the help of transformer-based models for unstructured textual data preprocessing and ARIMA for structured time series analysis is referred to as a hybrid model. This model combines the insights from both approaches using a feature fusion mechanism. It evaluated the Hybrid Model using accuracy, precision, recall, and finally, F1 score, and it was found to perform much better, generally obtaining an overall accuracy of 94.2% and an overall weighted F1 score of 94.3%. Specifically, class-specific analysis demonstrated high precision in identifying disruptions such as pandemics (95.5%) and natural disasters (94.6%), showing the ability of a model to understand context and time. The proposed approach outperforms classic stand-alone statistical and deep learning models regarding scalability and adaptivity to real-life applications such as risk management and policy making. Future work could include making the weights for each cluster dynamic to optimize weights based on real-time trends and improving accuracy and resilience.

Author 1: Qingzi Wang

Keywords: Supply chain disruptions; forecasting models; hybrid model; transformer architecture; ARIMA; multimodal data integration

PDF

Paper 54: Marine Predator Algorithm and Related Variants: A Systematic Review

Abstract: The Marine Predators Algorithm (MPA) is classified under swarm intelligence methods based on its type of inspiration. It is a population-based metaheuristic optimization algorithm inspired by the general foraging behavior exhibited in the form of Levy and Brownian motion in ocean predators supported by the policy of optimum success rate found in the biological relationship between prey and predators. The algorithm is easy to implement and robust in searching, yielding better solutions to many real-world problems. It is attracting huge and growing interest. This paper provides a systematic review of the research progress and applications of the MPA by analyzing more than 100 articles sourced from Scopus and Web of Science databases using the PRISMA approach. The study expounded the classical MPA’s workflow. It also unveiled a steady upward trend in the use of the algorithm. The research presented different improvements and variants of MPA including parameter-tuning, enhancement of the balance between exploration and exploitation, hybridization of MPA with other techniques to harness the strengths of each of the algorithms towards complementing the weaknesses of the other, and more recently proposed advances. It further underscores the application of MPA in various areas such as Engineering, Computer Science, Mathematics, and Energy. Findings reveal several search strategies implemented to improve the algorithm’s performance. In conclusion, although MPA has been widely accepted, other areas remain yet to be applied, and some improvements are yet to be covered. These have been presented as recommendations for future research direction.

Author 1: Emmanuel Philibus
Author 2: Azlan Mohd Zain
Author 3: Didik Dwi Prasetya
Author 4: Mahadi Bahari
Author 5: Norfadzlan bin Yusup
Author 6: Rozita Abdul Jalil
Author 7: Mazlina Abdul Majid
Author 8: Azurah A Samah

Keywords: Exploitation-exploration; marine predator algorithm; metaheuristic algorithms; metaheuristic-hybridization; meta-heuristics; optimization; predator prey systems

PDF

Paper 55: The Current Challenges Review of Deep Learning-Based Nuclei Segmentation of Diffuse Large B-Cell Lymphoma

Abstract: Diffuse Large B-Cell Lymphoma stands as the most prevalent form of non-Hodgkin lymphoma worldwide, constituting approximately 30 percent of cases within this diverse group of blood cancers affecting the lymphatic system. This study addresses the challenges associated with the accurate DLBCL segmentation and classification, including difficulties in identifying and diagnosing DLBCL, manpower shortage, and limitations of manual imaging methods. The study highlights the potential of deep learning to effectively segment and classify DLBCL types. The implementation of such technology has the potential to extract and preprocess image patches, identify, and segment the nuclei in DLBCL images, and classify DLBCL severity based on segmented nuclei counting.

Author 1: Gei Ki Tang
Author 2: Chee Chin Lim
Author 3: Faezahtul Arbaeyah Hussain
Author 4: Qi Wei Oung
Author 5: Aidy Irman Yazid
Author 6: Sumayyah Mohammad Azmi
Author 7: Haniza Yazid
Author 8: Yen Fook Chong

Keywords: Deep learning; Diffuse Large B-Cell Lymphoma (DLBCL); lymphoma cancer; HoVerNet

PDF

Paper 56: User Interface Design of Digital Test Based on Backward Chaining as a Measuring Tool for Students’ Critical Thinking

Abstract: Assessing students' critical thinking skills is challenging due to the limitations of current measurement tools. Therefore, there is a need for a digital testing instrument that can effectively evaluate students' critical thinking abilities. The proposed digital test should be designed to present questions in a tiered manner, using a backward chaining approach that starts with general questions and progresses to more detailed ones. However, developing this measurement instrument requires careful planning. One of the initial steps in this process is to create a user interface design. The purpose of this study was to show the quality of the design of the user interface of a digital test based on backward chaining as a measuring tool for students’ critical thinking in a differentiated learning atmosphere. Design development used the Borg and Gall model and only focused on three stages. These stages include design planning, initial testing, and revision for the initial testing results. Data collection was through initial testing of the design. The tool used to collect data was a questionnaire. Respondents involved in the initial testing were 34 people. The location for the study was at several IT vocational high schools spread across six regencies in Bali. The data analysis technique compared the percentage comparison of the quality of the user interface design with the quality standards of the user interface design and referred to a five scale. The results of the study showed that the design quality of the digital test user interface based on backward chaining was included in the good category, as indicated by a quality percentage of 88.94%. Specifically, the impact of the results on the field of educational evaluation is to make it easier for evaluators to make accurate measurements. In general, the effect of this study on the field of informatics engineering education is the existence of innovations in realizing a test to measure critical thinking in the domain of differentiated learning.

Author 1: I Putu Wisna Ariawan
Author 2: P. Wayan Arta Suyasa
Author 3: Agus Adiarta
Author 4: I Komang Gede Sukawijana
Author 5: Nyoman Santiyadnya
Author 6: Dewa Gede Hendra Divayana

Keywords: User interface design; digital test; backward chaining; critical thinking; differentiated learning

PDF

Paper 57: Early Alzheimer’s Disease Detection Through Targeting the Feature Extraction Using CNNs

Abstract: Alzheimer's Disease (AD) is a persistent, irreversible, and degenerative neurological disorder of the brain that currently has no effective therapy. This condition is identified by pathological abnormalities in the hippocampal area, which may develop up to 10 years prior to the onset of clinical symptoms. Timely detection of pathogenic abnormalities is essential to impede the worsening of AD. Recent studies on neuroimaging have shown that the use of Deep Learning techniques to analyze multimodal brain scans may effectively and correctly detect AD. The main goal of this work is to design and develop an Artificial Intelligence (AI) based diagnostic framework that can accurately and promptly detect AD by analyzing Structural Magnetic Resonance Imaging (SMRI) data. This study presents a novel approach that combines a Directed Acyclic Graph 3D-CNN with an SVM classifier for timely detection and identification of AD by analyzing the Regions of Interest (RoI) like cerebral spinal fluid, white and gray matter, and the hippocampus in SMRI images. The proposed hybrid model combines Deep Learning for feature extraction and Machine Learning techniques for classification. The obtained results demonstrate its superior performance compared to earlier methods in accurately identifying individuals with early mild cognitive impairment (EMCI) from those with normal cognition (NC) using the Alzheimer's Disease Neuroimaging Initiative (ADNI) dataset. The model attains a classification accuracy of 97.67%, with precision at 94.12%, and sensitivity at 98.60%.

Author 1: D Prasad
Author 2: K Jayanthi
Author 3: Pradeep Tilakan

Keywords: Alzheimer's Disease (AD); convolutional neural networks (CNN); Support Vector Machine (SVM); Directed Acyclic Graph (DAG); Late Mild Cognitive Impairment (LMCI); Alzheimer's Disease Neuroimaging Initiative (ADNI)

PDF

Paper 58: Enhancement of Coastline Video Monitoring System Using Structuring Element Morphological Operations

Abstract: Coastal monitoring is vital in environmental management, disaster mitigation, and addressing climate change impacts. Traditional methods are time-consuming and error-prone, prompting the need for innovative systems. This study introduces the Coastal Video Monitoring System (CoViMos), a novel framework for real-time shoreline detection in tropical regions, specifically at Kedonganan Beach, Bali. The CoViMos framework utilizes advanced video monitoring and optimized morphological operations to address challenges such as environmental noise and dynamic shoreline behavior. Key innovations include Kapur’s entropy thresholding enhanced with the Grasshopper Optimization Algorithm (GOA) and structuring elements tailored to the beach’s unique features. Sensitivity analysis reveals that a structuring element size of five pixels offers optimal performance, balancing efficiency, and image fidelity. This configuration achieves peak values in quality metrics such as the Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index (SSIM), Complex Wavelet SSIM (CWSSIM), and Feature Similarity Index (FSIM) while minimizing Mean Squared Error (MSE) and reducing processing time. The results demonstrate significant improvements in shoreline detection accuracy, with PSNR increasing by 9.3%, SSIM by 1.4%, CWSSIM by 1.7%, and FSIM by 1.6%. Processing time decreased by 1.3%, emphasizing the system’s computational efficiency. These enhancements ensure more precise shoreline mapping, even in noisy and dynamic environments.

Author 1: I Gusti Ngurah Agung Pawana
Author 2: I Made Oka Widyantara
Author 3: Made Sudarma
Author 4: Dewa Made Wiharta
Author 5: Made Widya Jayantari

Keywords: Coastline detection; image processing; Video Monitoring System (CoViMoS); morphological operations

PDF

Paper 59: Application of MLP-Mixer-Based Image Style Transfer Technology in Graphic Design

Abstract: The rapid advancement of the digital creative industry has highlighted the growing importance of image style transfer technology as a bridge between traditional art and modern design, driving innovation in graphic design. However, conventional style transfer methods face significant challenges, including low computational efficiency and unnatural style transformation in complex image scenarios. This study addresses these limitations by introducing a novel approach to image style transfer based on the MLP-Mixer model. Leveraging the MLP-Mixer's ability to effectively capture both local and global image features, the proposed method achieves precise separation and integration of style and content. Experimental results demonstrate that the MLP-Mixer-based style transfer significantly enhances the naturalness and diversity of style transformation while preserving image clarity and detail. Additionally, the processing speed is improved by 50%, with style conversion accuracy and user satisfaction increasing by 30% and 35%, respectively, compared to traditional methods. These findings underscore the potential of the MLP-Mixer model for advancing efficiency and realism in graphic design applications.

Author 1: Qibin Wang
Author 2: Xiao Chen
Author 3: Huan Su

Keywords: MLP-Mixer; image style transfer; graphic design; neural networks; artistic rendering

PDF

Paper 60: Integrating Blockchain and Edge Computing: A Systematic Analysis of Security, Efficiency, and Scalability

Abstract: The integration of blockchain and edge computing presents a transformative potential to enhance security, computing efficiency, and data privacy across diverse industries. This paper begins with an overview of blockchain and edge computing, establishing the foundational technologies for this synergy. It explores the key benefits of their integration, such as improved data security through blockchain’s decentralized nature and reduced latency via edge computing's localized data processing. Methodologically, the paper employs a systematic analysis of existing technologies and challenges, emphasizing issues such as scalability, managing decentralized networks, and ensuring independence from cloud infrastructure. A detailed Ethereum-based case study demonstrates the feasibility and practical implications of deploying blockchain in edge computing environments, supported by a comparative analysis and an algorithmic approach to integration. The conclusion synthesizes the findings, addressing unresolved challenges and proposing future research directions to optimize performance and ensure the seamless convergence of these technologies.

Author 1: Youness Bentayeb
Author 2: Kenza Chaoui
Author 3: Hassan Badir

Keywords: Blockchain; edge computing; security; computing efficiency; data privacy

PDF

Paper 61: Enhancing COVID-19 Detection in X-Ray Images Through Deep Learning Models with Different Image Preprocessing Techniques

Abstract: The identification of COVID-19 using chest X-ray (CXR) images plays a critical role in managing the pandemic by providing a rapid, non-invasive, and accessible diagnostic tool. This study evaluates the impact of different image preprocessing techniques on the performance of deep learning models for COVID-19 classification based on COVID-19 Radiography Database, which includes 10,192 normal CXR images, 6012 lung opacity (non-COVID lung infection) images, and 1345 viral pneumonia images. Along with the images, corresponding lung masks are also included to aid in the segmentation and analysis of lung regions. Specifically, three convolutional neural network (CNN) models were developed, each using a distinct preprocessing method: Contrast Limited Adaptive Histogram Equalization (CLAHE), traditional histogram equalization, and no preprocessing. The results revealed that while the CLAHE-enhanced model achieved the highest training accuracy (93.26%) and demonstrated superior stability during training, it showed lower performance in the validation phase, with validation accuracy of 91.31%. In contrast, the model with no preprocessing, which exhibited slightly lower training accuracy (92.98%), outperformed the CLAHE model during validation, achieving the highest validation accuracy of 91.50% and the lowest validation loss. The histogram equalization model demonstrated performance similar to that of CLAHE but with slightly higher validation loss and accuracy compared to the unprocessed model. These findings suggest that while CLAHE excels in enhancing image details during training, it may lead to overfitting and reduced generalization ability. In contrast, the model without preprocessing showed the best generalization and stability, indicating that preprocessing techniques should be chosen carefully to balance feature enhancement with the need for generalization in real-world applications. This study underscores the importance of selecting appropriate image preprocessing techniques to enhance deep learning models' performance in medical image classification, particularly for COVID-19 detection. Histogram Equalization The results contribute to ongoing efforts to optimize diagnostic tools using AI and image processing.

Author 1: Ahmad Nuruddin bin Azhar
Author 2: Nor Samsiah Sani
Author 3: Liu Luan Xiang Wei

Keywords: X-ray; COVID-19; image enhancement; Contrast Limited Adaptive Histogram Equalization; Histogram Equalization

PDF

Paper 62: Deep Learning-Based Automatic Cultural Translation Method for English Tourism

Abstract: The general LSTM-based encoder-decoder model has the problems of not being able to mine the sentence semantics and translate long text sequences. This study presents a neural machine translation model utilizing LSTM with improved attention, incorporating multi-head attention and multi-skipping attention mechanisms into the LSTM baseline model. By adding multi-head attention computation, the syntactic information in different subspaces can be mined, and then attention can be paid to the semantic information in the sentence sequences, and then multiple attentions are computed on each head separately, which can effectively deal with the long-distance dependency problem and perform better in the translation of long sentences. The proposed model is analysed and compared using the WMT17 Chinese and English datasets, newsdev2017 and newstest2017, and the results show that the proposed model improves the BLEU score of the automatic translation of Tourism English Culture and solves the problem of low scores in long sentence translation.

Author 1: Jianguo Liu
Author 2: Ruohan Liu

Keywords: LSTM-based encoder-decoder model; tourism English culture; automatic translation; enhanced attention mechanism

PDF

Paper 63: A Novel Metric-Based Counterfactual Data Augmentation with Self-Imitation Reinforcement Learning (SIL)

Abstract: The inherent biases present in language models often lead to discriminatory predictions based on demographic attributes. Fairness in NLP refers to the goal of ensuring that language models and other NLP systems do not produce biased or discriminatory outputs that could negatively affect individuals or groups. Bias in NLP models often arises from training data that reflects societal stereotypes or imbalances. Robustness in NLP refers to the ability of a model to maintain performance when faced with noisy, adversarial, or out-of-distribution data. A robust NLP model should handle variations in input effectively without failing or producing inaccurate results. The proposed approach employs a novel metric called CFRE (Context-Sensitive Fairness and Robustness Evaluation) designed to measure both fairness and robustness of an NLP model under different contextual shifts. Next, it projected the benefits of this metric in terms of experimental parameters. Next, the work integrated counterfactual data augmentation with help of Self-Imitation Reinforcement Learning (SIL) to reinforce successful counterfactual generation by enabling the model to learn from its own high-reward experiences, fostering a more balanced understanding of language. The integration of SIL allows for efficient exploration of the action space, guiding the model to consistently produce unbiased outputs across different contexts. The proposed approach demonstrates the effectiveness of our method through extensive experimentation and compared the results of the proposed metric with that of WEAT and SMART testing, and showed a significant reduction in bias without compromising the model's overall performance. This framework not only addresses bias in existing models but also contributes to a more robust methodology for training fairer NLP systems. Both the proposed metric and SIL showed better results in experimental parameters.

Author 1: K. C. Sreedhar
Author 2: T. Kavya
Author 3: J. V. S. Rajendra Prasad
Author 4: V. Varshini

Keywords: Natural language processing; fairness; robustness; Word Embedding Association Test (WEAT); SMART testing

PDF

Paper 64: Segmentation of Nano-Particles from SEM Images Using Transfer Learning and Modified U-Net

Abstract: Nanomaterials, owing to their distinctive features, are crucial across numerous scientific domains, especially in materials science and nanotechnology. Precise segmentation of Scanning Electron Microscope (SEM) images is essential for evaluating attributes such as nanoparticle dimensions, morphology, and distribution. Conventional image segmentation techniques frequently prove insufficient for managing the intricate textures of SEM images, resulting in a laborious and imprecise process. In this research, a modified U-Net architecture is presented to tackle this challenge, utilizing a ResNet50 backbone pre-trained on ImageNet. This model utilizes the robust feature extraction abilities of ResNet50 alongside the effective segmentation performance of U-Net, hence improving both accuracy and computational efficiency in TiO2 nanoparticle segmentation. The suggested model was assessed using performance metrics including accuracy, precision, recall, IoU, and Dice Coefficient. The results indicated a high segmentation accuracy, demonstrated by a Dice score of 0.946 and an IoU of 0.897, with little variability reflected in standard deviations of 0.002071 and 0.003696, respectively, over 200 epochs. The comparison with existing methods demonstrates that the proposed model surpasses previous approaches by attaining enhanced segmentation accuracy. The modified U-Net design serves as an excellent technique for accurate nanoparticle segmentation in SEM images, providing substantial enhancements compared to traditional approaches. This progress indicates the model's potential for wider applications in nanomaterial research and characterization, where precise and efficient segmentation is essential for analysis.

Author 1: Sowmya Sanan V
Author 2: Rimal Isaac R S

Keywords: Nanomaterial; segmentation; ResNet 50; modified UNet; transfer learning; SEM

PDF

Paper 65: Application of Big Data Mining System Integrating Spectral Clustering Algorithm and Apache Spark Framework

Abstract: Spectral clustering algorithm is a highly effective clustering algorithm with broad application prospects in data mining. To improve the efficient data processing capability of big data mining systems, a big data mining system that integrates spectral clustering algorithm and Apache Spark framework is proposed. It is applied in the big data mining system by combining Hdoop, Spark framework, and spectral clustering algorithm. The research results indicated that after 300 iterations of spectral clustering algorithm, the error value tended to stabilize and drops to 0.123. In different datasets, different error values were displayed, indicating that spectral clustering algorithm had better performance in discrete data processing and smaller testing errors. The minimum time consumed by the comparative system was 37.83 seconds, the maximum time was 55.26 seconds, and the average time was 51.65 seconds. The minimum time consumed by the research system was 18.93 seconds, the maximum time consumed was 32.22 seconds, and the average time consumed was 28.14 seconds. Compared with the comparative system, the research system consumed less time, trained faster, and was more conducive to shortening the clustering running time. The algorithm framework and system raised in the research have good operational efficiency and clustering ability in data mining processing, which promotes the reliability and development of big data mining systems.

Author 1: Yuansheng Guo

Keywords: Spectral clustering algorithm; apache spark; big data; data mining

PDF

Paper 66: Large Language Models for Academic Internal Auditing

Abstract: This research examines the application of Artificial Intelligence in internal auditing, focusing on document management and information retrieval in academic institutions. The study proposes using Large Language Models to streamline document processing during audit preparation, addressing inefficiencies in traditional document handling methods. Through experimental evaluation of three embedding models (BGE-M3, Nomic-embed-text-v1, and CamemBERT) on a dataset of 300 academic regulatory queries, the research demonstrates BGE-M3's superior performance with an nDCG3 score of 0.90 and top-1 accuracy of 82.5%. The methodology incorporates query expansion using GPT-4 and Llama 3.1, revealing robust performance across varied query formulations. While highlighting AI's potential to transform internal auditing practices, particularly in Morocco's academic sector, the study acknowledges implementation challenges including institutional constraints and resistance to technological change. The conducted experiments and result analysis provide useful criteria that can be applied to similar information retrieval challenges in other fields and real-world applications.

Author 1: Houda CHAMMAA
Author 2: Rachid ED-DAOUDI
Author 3: Khadija BENAZZI

Keywords: Large language models; internal auditing; information retrieval; embedding models; academic institutions

PDF

Paper 67: Enhanced Facial Expression Recognition Based on ResNet50 with a Convolutional Block Attention Module

Abstract: Deep learning techniques are becoming increasingly important in the field of facial expression recognition, especially for automatically extracting complex features and capturing spatial layers in images. However, previous studies have encountered challenges such as complex data sets, limited model generalization, and lack of comprehensive comparative analysis of feature extraction methods, especially those involving attention mechanisms and hyperparameter optimization. This study leverages data science methodologies to handle and analyze large, intricate datasets, while employing advanced computer vision algorithms to accurately detect and classify facial expressions, addressing these challenges by comprehensively evaluating FER tasks using three deep learning models (VGG19, ResNet50, and InceptionV3). The convolutional block attention module is introduced to enhance feature extraction, and the performance of the model is further improved by hyperparameter tuning. The experimental results show that the accuracy of VGG19 model is the highest 71.7\% before the module is integrated, and the accuracy of ResNet50 is the highest 72.4\% after the module is integrated. The performance of all models was significantly improved through the introduction of attention mechanisms and hyperparameter tuning, highlighting the synergistic potential of data science and computer vision in developing robust and efficient in facial expression recognition systems.

Author 1: Liu Luan Xiang Wei
Author 2: Nor Samsiah Sani

Keywords: Data science; computer vision; deep learning; facial expression recognition

PDF

Paper 68: YOLO-WP: A Lightweight and Efficient Algorithm for Small-Target Detection in Weld Seams of Small-Diameter Stainless Steel Pipes

Abstract: To address the low detection efficiency and high computational resource demands of current welded pipe defect detection algo-rithms for small target defects, this paper proposes the YO-LO-WP algorithm based on YOLOv5s. The improvements of YOLO-WP are mainly reflected in the following aspects: First, an innovative GhostFusion architecture is introduced in the backbone network. By replacing the C3 modules with C2f mod-ules and integrating the Ghost CBS module inspired by Ghost convolution, cross-stage feature fusion is achieved, significantly enhancing computational efficiency and feature representation for small target defects. Second, the Slim-Neck lightweight de-sign based on GSConv is employed in the neck to further opti-mize the network structure and reduce the number of parame-ters. Additionally, the SimAM lightweight attention mechanism is incorporated to improve the network's ability to extract de-fect features, and the Focal-EIou loss is utilized to optimize CIou loss, thereby enhancing small object detection and accelerating loss convergence. The experimental results show that the AP(D1) and mAP@0.5 of the YOLO-WP model are improved by 5.3% and 3%, respectively, over the original model. In addi-tion, the number of model parameters and FLOPs are reduced by 40% and 45%, respectively, achieving a good balance be-tween performance and efficiency. We evaluated the perfor-mance of YOLO-WP using other datasets and showed that YOLO-WP exhibits excellent applicability. Compared to exist-ing mainstream detection algorithms, YOLO-WP is more ad-vanced. The YOLO-WP model significantly enhances produc-tion quality in industrial defect detection, laying the foundation for building compact, high-performance embedded weld pipe surface defect detection systems.

Author 1: Huaishu Hou
Author 2: Yukun Sun
Author 3: Chaofei Jiao

Keywords: Welded pipe; lightweight model; defect detection; deep learning; feature extraction; attention mechanism

PDF

Paper 69: Determination of Pre Coding Elements and Activities for a Pre Coding Program Model for Kindergarten Children Using the Fuzzy Delphi Method (FDM)

Abstract: Computational thinking (CT) skills are becoming increasingly crucial in education, particularly in early childhood education. Pre coding, which involves hands-on activities with real objects, has been shown to be quite effective in fostering kindergarten children's computational skills. Pre coding, on the other hand, is essential for boosting children’s CT skills, but teachers frequently lack the information necessary to teach these skills successfully. Their successful adoption is hampered by the early childhood education community's lack of interest in CT skills and the sparse application of pre coding techniques. In order to help kindergarten instructors incorporate pre coding into their teaching and learning, this study focuses on defining the elements and activities described in a pre-coding program model. The study reviewed and compiled a list of prior literature’s pre coding elements and activities. Subsequently, the Fuzzy Delphi Method (FDM) was utilised to refine and validate these elements and activities. Finally, the data collected from 11 selected experts relevant to this field of study were analysed using FDM to examine consensus. The results showed that the eight identified elements and 24 pre coding activities fulfilled the following required criteria: a threshold value (d) of lower than or equal to 0.2, an agreement percentage over 75%, and a fuzzy score value (A) higher than 0.5. These findings demonstrated the suitability of the identified pre coding elements and activities for integration into a pre coding program model for kindergarten children. In summary, this study provides valuable guidance for kindergarten teachers in implementing practical pre coding activities to enhance CT skills among children.

Author 1: Siti Naimah Rahman
Author 2: Norly Jamil
Author 3: Intan Farahana Abdul Rani
Author 4: Hafizul Fahri Hanafi

Keywords: Expert consensus; pre coding; element; activity; kindergarten children

PDF

Paper 70: A Novel Internet of Things and Cloud Computing-Driven Deep Learning Framework for Disease Prediction and Monitoring

Abstract: In smart cities, the e-healthcare systems aided by Internet of Things (IoT) technologies play a significant role in proficient health monitoring services. The sensitivity and number of users in health networks highlights the necessity of treating security attacks. In the era of rapid internet connectivity and cloud computing services, patient medical information is most sensitive, and its electronic representation poses privacy and security concerns. Moreover, it is challenging for the traditional classifier to process a massive amount of health data and classify patients' health statuses. To address this matter, this paper presents a novel healthcare model, IoT-CDLDPM, to estimate patients’ disease levels using original data and fuzzy entropy extracted from patients' remote locations. IoT-CDLDPM incorporates a deep learning classifier to analyze extensive patient-related data and provides efficient and accurate health status predictions. Furthermore, the proposed model presents the secured storage structure of the individual's health data in cloud servers. To give the authenticity of the health data, two new cryptographic algorithms are presented that encrypt and decrypt the data securely transmitted through the network. A comparison with existing methods reveals that the proposed system significantly reduces computation time, with a recorded time of 0.5 seconds, outperforming DSVS, PP-ESAP, and DRDA by up to 80%. Furthermore, the proposed cryptographic model enhances security levels, achieving a range between 99.4% and 99.8% across multiple experimental setups, surpassing other widely used encryption algorithms such as AES, RSA, and ECC-DH.

Author 1: Bo GUO
Author 2: Lei NIU

Keywords: IoT-driven healthcare; deep learning; fuzzy entropy; secure data storage; cryptography

PDF

Paper 71: Comparison of Artificial Neural Network and Long Short-Term Memory for Modelling Crude Palm Oil Production in Indonesia

Abstract: Indonesia is one of the largest producers and exporters of Crude Palm Oil (CPO), making CPO production crucial to the country's economic stability. Accurate forecasting of CPO production is essential for effective inventory management, export-import strategy, and economic planning. Traditional time series methods like ARIMA have limitations in modeling nonlinear data, leading to the adoption of machine learning approaches such as Artificial Neural Network (ANN) and Long Short-Term Memory (LSTM). This study compares the performance of ANN, a general neural network, and LSTM, a neural network specifically designed for time series data, in predicting CPO production in Indonesia. Data from 2003 to 2022 were used to train and evaluate both models with various hyperparameter tuning configurations. The results indicate that while both models provide excellent forecasting accuracy, with MAPE values below 10%, the LSTM model achieved a lower out-of-sample MAPE of 5.78% compared to ANN’s 6.87%, suggesting superior performance by LSTM in capturing seasonal patterns in CPO production. Consequently, LSTM is recommended as the preferred model for CPO production forecasting due to its enhanced ability to handle temporal dependencies and nonlinear patterns in the data.

Author 1: Brodjol Sutijo Suprih Ulama
Author 2: Robi Ardana Putra
Author 3: Fausania Hibatullah
Author 4: Mochammad Reza Habibi
Author 5: Mochammad Abdillah Nafis

Keywords: Artificial Neural Network (ANN); Crude Palm Oil (CPO); Long Short-Term Memory (LSTM)

PDF

Paper 72: Enhanced Jaya Algorithm for Quality-of-Service-Aware Service Composition in the Internet of Things

Abstract: The Internet of Things (IoT) has shifted how devices and services interact, resulting in diverse innovations ranging from health and smart cities to industrial automation. Nevertheless, at its core, IoT continues to face one of the major tough tasks of Quality of Service-aware Service Composition (QoS-SC), as these IoT settings are normally transient and unpredictable. This paper proposes an improved Jaya algorithm for QoS-SC and focuses on optimizing service selection with a balance between the main QoS attributes: execution time, cost, reliability, and scalability. The proposed approach was designed with adaptive mechanisms to avoid local optima stagnation and slow convergence and thus assure robust exploration and exploitation of the solution area. Incorporating these enhancements, the proposed algorithm outperforms prior metaheuristic approaches regarding QoS satisfaction and computational efficiency. Extensive experiments conducted over diverse IoT scenarios show the algorithm's scalability, demonstrating that it can achieve faster convergence with superior QoS optimization.

Author 1: Yan SHI

Keywords: Service composition; internet of things; quality of service; Jaya algorithm; optimization

PDF

Paper 73: Enhancing Facial Expressiveness in 3D Cartoon Animation Faces: Leveraging Advanced AI Models for Generative and Predictive Design

Abstract: An advanced system for facial landmark detection and 3D facial animation rigging is proposed, utilizing deep learning algorithms to accurately detect key facial points, such as the eyes, mouth, and eyebrows. These landmarks enable precise rigging of 3D models, facilitating realistic and controlled facial expressions. The system enhances animation efficiency and realism, providing robust solutions for applications in gaming, animation, and virtual reality. This approach integrates cutting-edge detection techniques with efficient rigging mechanisms. The AI-assisted rigging process reduces manual effort and ensures precise, dynamic animations. The study evaluates the system's accuracy in facial landmark detection, the efficiency of the rigging process, and its performance in generating consistent emotional expressions across animations. Additionally, the system's computational efficiency, scalability, and system performance are assessed, demonstrating its practicality for real-time applications. Pilot testing, emotion recognition consistency, and performance metrics reveal the system's robustness and effectiveness in producing realistic animations while reducing production time. This work contributes to the advancement of animation and virtual environments, offering a scalable solution for realistic facial expression generation and character animation. Future research will focus on refining the system and exploring its potential applications in interactive media and real-time animation.

Author 1: Langdi Liao
Author 2: Lei Kang
Author 3: Tingli Yue
Author 4: Aiting Zhou
Author 5: Ming Yang

Keywords: Facial landmark detection; 3D animation; deep learning; AI-assisted rigging; emotion recognition

PDF

Paper 74: A Lightweight Anonymous Identity Authentication Scheme for the Internet of Things

Abstract: With the rapid growth of Internet of Things (IoT) devices, many of which are resource-constrained and vulnerable to attacks, current identity authentication methods are often too resource-intensive to provide adequate security. This paper proposes an efficient identity authentication scheme that integrates Physical Unclonable Functions (PUFs), Chebyshev chaotic maps, and fuzzy extractors. The scheme enables mutual authentication and key agreement without the need for passwords or smart cards, while providing effective defense against various attacks. The security of the proposed scheme is formally analyzed using an improved BAN logic. A comparison with existing related protocols in terms of security features, computational overhead, and communication overhead demonstrates the security and efficiency of the proposed scheme.

Author 1: Zhengdong Deng
Author 2: Xuannian Lei
Author 3: Junyu Liang
Author 4: Hang Xu
Author 5: Zhiyuan Zhu
Author 6: Na Lin
Author 7: Zhongwei Li
Author 8: Jingqi Du

Keywords: Internet of Things; identity authentication; Physical Unclonable Functions; fuzzy extractors; chaotic maps

PDF

Paper 75: Comparative Analysis of Feature Selection Based on Metaheuristic Methods for Human Heart Sounds Classification Using PCG Signal

Abstract: Cardiovascular disease is a critical threat to human health, as most death cases are due to heart disease. Although several doctors employ stethoscopes to auscultate heart sounds to detect abnormalities, the accuracy of the approach is considerably dependent upon the experience and skills of the physician. Consequently, optimal methods are required to analyse and classify heart sounds with Phonocardiogram (PCG) signal-based machine learning methods. The current study formulated a binary classification model by subjecting PCG signals to hyper-filtering with low-pass and cosine filters. Subsequently, numerous features are extracted with the Wavelet Scattering Transform (WST) method. During the feature selection stage, several metaheuristic methods, including Harris Hawks Optimisation (HHO), Dragonfly Algorithm (DA), Grey Wolf Optimiser (GWO), Salp Swarm Algorithm (SSA), and Whale Optimisation Algorithm (WOA), are employed to compare the attributes separately and determine the ideal characteristics for improved classification accuracy. Finally, the selected features were applied as input for the Bidirectional Long Short-Term Memory (Bi-LSTM) algorithm, simplifying the classification process for distinguishing normal and abnormal heart sounds. The present study assessed three PCG datasets: PhysioNet 2016, Yaseen Khan 2018, and PhysioNet 2022, documenting 94.85%, 100%, and 66.87% accuracy rates with 127-SSA, 168-HHO, and 163-HHO, respectively. Based on the results of the PhysioNet 2016 and 2022 datasets, the proposed method with hyperparameters demonstrated superior performance to those with default parameters in categorising normal and abnormal heart sounds appropriately.

Author 1: Motaz Faroq A Ben Hamza
Author 2: Nilam Nur Amir Sjarif

Keywords: Cardiovascular Diseases (CVDs); Phonocardiogram (PCG) signal processing; Wavelet Scattering Transform (WST); Metaheuristic Methods; Harris Hawks Optimisation (HHO); Dragonfly Algorithm (DA); Grey Wolf Optimiser (GWO); Salp Swarm Algorithm (SSA); Whale Optimisation Algorithm (WOA); Bidirectional Long Short-Term Memory (Bi-LSTM)

PDF

Paper 76: Developing an Integrated Platform to Track Real Time Football Statistics for Somali Football Federation (SFF)

Abstract: The integration of technology in sports has revolutionized how stakeholders interact with and perceive the game. This thesis presents the development of an integrated platform aimed at tracking real-time football statistics for the Somali Football Federation (SFF). Football, being one of the most popular sports globally, relies heavily on accurate and up-to-date statistical data for player performance analysis, team strategies, and fan engagement. The SFF, like many other federations, faces challenges in collecting, managing, and utilizing football statistics effectively. The advent of digital technologies and the internet has revolutionized data collection and dissemination methods across various fields, including sports. Traditional methods of data collection and analysis, which are often manual and time-consuming, can no longer meet the demands of modern football analytics. The platform encompasses a mobile application for fans, an admin panel for administrators, and a backend system for data management. Leveraging modern technologies such as Flutter for mobile development, Node.js and MySQL for backend services, and React for the admin interface, the system ensures comprehensive coverage of match events, player statistics, and tournament standings. Real-time updates facilitated by Socket.IO enhance user engagement and decision-making capabilities for coaches and administrators.

Author 1: Bashir Abdinur Ahmed
Author 2: Husein Abdirahman Hashi
Author 3: Abdifatah Abdilatif Ahmed
Author 4: Abdikani Mahad Ali

Keywords: Real-time football statistics; Integrated sports platform; Somali Football Federation (SFF); user engagement; sports technology

PDF

Paper 77: Elevator Abnormal State Detection Based on Vibration Analysis and IF Algorithm

Abstract: Elevators play a crucial role in daily life, and their safety directly impacts the personal and property safety of users. To detect abnormal states of elevators and ensure people's personal safety, the acceleration signal of elevators is decomposed and Weiszfeld algorithm is used to estimate gravity acceleration. In addition, the study also introduces Kalman filtering to reduce error accumulation. To estimate the operating position of elevators, a method based on information fusion is studied and designed to construct a mapping relationship between elevator vibration energy and position, and to locate the height of elevator faults. Finally, an anomaly detection model combining vibration analysis and the Isolated Forest algorithm is developed. The results showed that the main distribution range of acceleration values in the horizontal direction was between 0.02 m2/s and -0.02 m2/s. The average estimation error and root mean square error of the research designed elevator position estimation method were 0.109 m and 0.113 m, respectively, which could solve the problem of accumulated position errors. The abnormal vibration energy and height corresponding to different operating conditions of elevators were different. The normal value ratios of the anomaly detection model under different sliding windows were 99.91% and 99.57%, respectively. The anomaly detection model designed for research has good performance and can provide technical support for the detection of elevator operation status.

Author 1: Zhaoxiu Wang

Keywords: Vibration analysis; IF algorithm; elevator; abnormal; detection

PDF

Paper 78: LFM Book Recommendation Based on Fusion of Time Information and K-Means

Abstract: To meet the growing demand in the field of book recommendation, the research focuses on meeting the personalized needs, behavioral patterns, and interests of readers. A book recommendation algorithm that combines K-means clustering with time information is proposed to provide more convenient and efficient book recommendation services and enhance readers' reading experience. The algorithm constructs a comprehensive user preference matrix by incorporating readers' borrowing time. Then, the K-means clustering is applied to group users with similar preferences and leverages a latent factor model to train and predict user ratings. The methodological integration of clustering and latent factor model ensures a more precise and dynamic recommendation process. The experimental results demonstrated that the proposed algorithm achieved a high average recommendation accuracy of 98.7%. Additionally, the algorithm maintained an average book popularity score of 8.2 after reaching stability, indicating its ability to suggest widely appreciated books. These outcomes validate the effectiveness of the algorithm in delivering accurate and popular book recommendations tailored to individual readers' needs. This study combines K-means clustering with time sensitive preference analysis and latent factor model to introduce an innovative method in the field of book recommendation systems. The findings provide valuable insights and practical applications for libraries seeking to enhance their personalized recommendation services, offering a significant contribution to the field of intelligent information retrieval.

Author 1: Dawei Ji

Keywords: Book recommendation; K-means; time information; latent factor model; preference matrix

PDF

Paper 79: A Proposed Approach for Agile IoT Smart Cities Transformation– Intelligent, Fast and Flexible

Abstract: Smart city architectures have been varying from one community to another. Each community leader develops their own perspective of smart cities. Some of these communities focus on data management, while others focus on provided services and infrastructure. In this research, an attempt to propose a clear, complete, and efficient perspective of smart cities is accomplished. The proposed generic architecture clarifies the full capabilities, requirements, and layers’ contribution to a successful smart city development. The proposed architecture utilizes Internet of Things tools as well as agile standards in the description of each layer. The research aims to discuss each layer in detail, the relationships among layers, the applied technology, and every aspect that leads to the success of using the recommended architecture. Although smart cities, IoT, and agile research have previously tackled the relation of each one of them with the other, up to the researchers’ knowledge, the three paradigms have not been previously considered as a unified collaborative approach. In order to reach the research target, the proposed architecture intelligently utilizes these paradigms and presents a robust architecture with high-quality standards.

Author 1: Othman Asiry
Author 2: Ayman E. Khedr
Author 3: Amira M. Idrees

Keywords: Smart city; agility; Internet of Things; cloud computing; intelligent systems

PDF

Paper 80: A Novel Optimization Strategy for CNN Models in Palembang Songket Motif Recognition

Abstract: Palembang Songket is an essential part of Indonesian cultural heritage, and its introduction and preservation present challenges, particularly in recognizing various motifs. This research introduces a novel strategy to optimize the performance of Convolutional Neural Networks (CNNs) by presenting a hierarchical integration of Ghost Module operations and Max Pooling, referred as Ghost Feature Maps. While the Ghost Module is effective in reducing parameters and enhancing feature extraction, it has limitations in filtering irrelevant information. To address this shortcoming, we propose a hierarchy in which Max Pooling works in conjunction with the Ghost Module, strengthening its performance by not only extracting dominant features but also eliminating excess, non-essential information. This hierarchical design enables more efficient feature extraction, thus enhancing the model's recognition accuracy. By combining Ghost Modules and Max Pooling in a structured manner, this approach advances established methodologies and offers a new perspective on feature representation within CNN architectures. Utilizing a dataset of 10 augmented classes of Palembang Songket motifs totaling 1000 images, we conducted experiments using varying ratios of Ghost Feature Maps. The results indicate that a ratio of 2 achieves an impressive accuracy of 0.98 with minimal parameter reduction. Additionally, a ratio of 3 results in a 34% decrease in parameters while maintaining a competitive accuracy of 0.95. Ratios of 4 and 5 continue to demonstrate robust performance, achieving accuracy levels of 0.93 while delivering over 60% reductions in model size and parameters. This research not only contributes to the optimization of CNN architectures but also supports the preservation of cultural heritage by improving the recognition capabilities of Palembang Songket motifs.

Author 1: Yohannes
Author 2: Muhammad Ezar Al Rivan
Author 3: Siska Devella
Author 4: Tinaliah

Keywords: Convolutional neural network; ghost module; palembang songket motif; recognition

PDF

Paper 81: A Novel Hybrid Algorithm Based on Butterfly and Flower Pollination Algorithms for Scheduling Independent Tasks on Cloud Computing

Abstract: Cloud computing is an Internet-based computing paradigm where virtual servers or workstations are offered as platforms, software, infrastructure, and resources. Task scheduling is considered one of the major NP-hard problems in cloud environments, posing several challenges to efficient resource allocation. Many metaheuristic algorithms have been extensively employed to address these task-scheduling problems as discrete optimization problems and have given rise to some proposals. However, these algorithms have inherent limitations due to local optima and convergence to poor results. This paper suggests a hybrid strategy for organizing independent tasks in heterogeneous cloud resources by incorporating the Butterfly Optimization Algorithm (BOA) and Flower Pollination Algorithm (FPA). Although BOA suffers from local optima and loss of diversity, which may cause an early convergence of the swarm, our hybrid approach outperforms such weaknesses by exploiting a mutualism-based mechanism. Indeed, the proposed hybrid algorithm outperforms existing methods while considering different task quantities with better scalability. Experiments are conducted within the CloudSim simulation framework with many task instances. Statistical analysis is performed to test the significance of the obtained results, which confirms that the suggested algorithm is effective at solving cloud-based task scheduling issues. The study findings indicate that the hybrid metaheuristic algorithm could be a promising approach to improving resource utilization and optimizing cloud task scheduling.

Author 1: Huiying SHAO

Keywords: Task scheduling; cloud computing; butterfly optimization algorithm; flower pollination algorithm; mutualism

PDF

Paper 82: Task Scheduling in Fog Computing-Powered Internet of Things Networks: A Review on Recent Techniques, Classification, and Upcoming Trends

Abstract: The Internet of Things (IoT) phenomenon influences daily activities by transforming physical equipment into smart objects. The IoT has achieved a wealth of technological innovations that were previously unimaginable. IoT application areas cover various sectors, including medical care, home automation, smart grids, and industrial operations. The massive growth of IoT applications causes network congestion because of the large volume of IoT tasks pushed to the cloud. Fog computing mitigates these transfers by placing resources near the edge. However, new challenges arise, such as limited computing power, high complexity, and the distributed characteristics of fog devices, negatively affecting the Quality of Service (QoS). Much research has been conducted to address these challenges in designing QoS-aware task scheduling optimization techniques. This paper comprehensively reviews task scheduling techniques in fog computing-powered IoT networks. We classify these techniques into heuristic-based, metaheuristic-based, and machine learning-based algorithms, evaluating their objectives, advantages, weaknesses, and performance metrics. Additionally, we highlight research gaps and propose actionable recommendations to address emerging challenges. Our findings offer a structured framework for researchers and practitioners to develop efficient, QoS-aware task scheduling solutions in fog computing environments.

Author 1: Dongge TIAN

Keywords: Internet of Things; task scheduling; fog computing; quality of service; network congestion; optimization

PDF

Paper 83: A System Dynamics Model of Frozen Fish Supply Chain

Abstract: The system dynamics methodology examines the intricate behaviors of complex systems through time, incorporating inventories, transfers, feedback cycles, lookup functions, and temporal delays. In fisheries systems, the interaction between resources and management entities is intricate, with the dynamics of fisheries significantly influencing the formulation of effective policies. Fisheries hold a vital position in Indonesia's economy, contributing to food security, nutrition, and the welfare of fishermen. Under Law Number 7 of 2016, the fisheries sector covers all activities, from resource management to the marketing of marine products. With its rich fishery resources, Indramayu Regency is a major contributor to West Java's fish production. TPI Karangsong, the hub of fishing activities in Indramayu, is a key player in the frozen fish supply chain, relying heavily on cold storage facilities to ensure product quality. Consequently, the system dynamics approach proves valuable in understanding the frozen fish supply chain by modeling the interactions between different variables and evaluating the impact of policies to improve fish quality. The system dynamics model in this study consists of six sub-models: fish at TPI, cold storage, refrigerated cabinets, total revenue, cash, and cold trucks. The simulation results provide policy recommendations to improve the quality of frozen fish at TPI Karangsong, namely the baseline scenario, cold truck scenario, cold truck scenario, truck and cold storage integration scenario, cold storage and fish catch drop integration scenario.

Author 1: Leni Herdiani
Author 2: Maun Jamaludin
Author 3: Iman Sudirman
Author 4: Widjajani
Author 5: Ismet Rohimat

Keywords: Supply chain; system dynamics; frozen fish; six sub-models; simulation; policy scenario

PDF

Paper 84: Methodological Review of Social Engineering Policy Model for Digital Marketing

Abstract: Social engineering attacks are recognized as human-based threats and continue to increase, despite studies focusing on prevention methods that do not rely on the human aspect. The impacts of these attacks are felt across various industries and organizations. To solve this issue, a social engineering policy model must be proposed for prevention in industrial settings, particularly emphasizing digital marketing activities, a crucial process in contemporary industries. However, hackers often exploit activities or information in these practices, necessitating an industry-specific policy to prevent these threats in digital marketing. As a result, a comprehensive review was conducted to identify critical method for develop social engineering policy model. The review uses Bryman's method to determine effective approaches for designing a social engineering policy model tailored for digital marketing. Consequently, this review provided a method for crafting effective social engineering policy, providing valuable insights for enhancing digital marketing security.

Author 1: Wenni Syafitri
Author 2: Zarina Shukur
Author 3: Umi Asma’ Mokhtar
Author 4: Rossilawati Sulaiman

Keywords: Digital marketing; social engineering attack prevention; review study; security policy model

PDF

Paper 85: Comprehensive Bibliometric Literature Review of Chatbot Research: Trends, Frameworks, and Emerging Applications

Abstract: This study aims to conduct a comprehensive bibliometric literature review of chatbot research by examining key trends, frameworks, and influential applications across various domains. It seeks to map the evolution of chatbot technologies, identify influential works, and analyze how the research focus has shifted over time, particularly towards AI-driven chatbot frameworks. An expanded dataset was compiled from the Scopus database, and bibliometric analyses were conducted using n-gram reference analysis, network mapping, and temporal trend visualization. The analysis was performed using R Studio with Biblioshiny, allowing for the identification of thematic clusters and the progression from rule-based to advanced retrieval and generative language model paradigms in chatbot research. Chatbot research has grown significantly from 2020 to 2024, with rising publication volumes and increased global collaboration, led by contributions from the USA, China, and emerging regions, such as Southeast Asia. Thematic analysis highlights a shift from foundational AI and NLP technologies to specialized applications such as mental health chatbots and e-commerce systems, emphasizing practical and user-centered solutions. Advances in chatbot architectures, including generative AI, have demonstrated the field's interdisciplinary nature and trajectory towards sophisticated, context-aware conversational systems. The analysis primarily used data from Scopus, which may limit the breadth of the included research. Future studies are encouraged to integrate data from other sources, such as the Web of Science (WoS) and PubMed, for a more comprehensive understanding of the field.

Author 1: Nazruddin Safaat Harahap
Author 2: Aslina Saad
Author 3: Nor Hasbiah Ubaidullah

Keywords: Chatbot research; bibliometric literature review; retrieval and generative; trend visualization

PDF

Paper 86: Application of Collaborative Filtering Optimization Algorithm Based on Semantic Relationships in Interior Design

Abstract: Due to the diversity of interior design, it is difficult for users to mine target data, so personalized recommendation systems for users are particularly important. Therefore, an optimized collaborative filtering recommendation system is proposed. Firstly, a random walk recommendation model based on category combination space is constructed, abandoning the traditional flat relationship connection and using Hasse diagram to achieve one-to-one mapping between items and types. The semantic relationship and distance are defined. Finally, a basic recommendation framework for random walks is established based on data such as jump behavior. Next, the potential semantic relationships between entities are explored, and a lightweight knowledge graph is proposed to define the social and explicit relationships between entities. Finally, the short-term features of the project are obtained using deep collaborative filtering technology, and a deep collaborative filtering temporal model based on semantic relationships is constructed. In subsequent validation, these experiments confirmed that under the vector dimension of 10, the average HR@K and NDCG@K were 6.9% and 12.9% higher than the other models. Therefore, the collaborative filtering recommendation model based on semantic relationships proposed in the study is reliable.

Author 1: Kai Zhao
Author 2: Lei Wang

Keywords: Semantic relationships; category combination space; random walks; collaborative filtering; temporal recommendations

PDF

Paper 87: Hybrid Clustering Framework for Scalable and Robust Query Analysis: Integrating Mini-Batch K-Means with DBSCAN

Abstract: Query clustering is a significant task in information retrieval. Research gaps still exist due to high-dimensional datasets, noise detection, and cluster interpretability. Solving these challenges will support large language models with faster and more efficient responses. This study aims to develop a hybrid clustering approach combining Mini-Batch K-means (MBK) and Density-Based Spatial Clustering of Application with Noise (DBSCAN) to cluster large-scale query datasets for information retrieval. The proposed method utilizes a preprocessing technique for data cleaning, extracts meaningful features, and scales all the features from the query dataset. The proposed hybrid clustering framework utilizes preprocessed data for clustering. The clustering algorithms MBK provide fast, scalable clustering, and DBSCAN delivers a precise, density-based refinement to efficiently process large-scale datasets while enhancing cluster boundaries to handle outliers. The proposed hybrid clustering framework effectively performs query analysis in information retrieval with a Silhouette score of 72.14 % and adjusted rand index of 78.23%. Thus, the hybrid clustering approach provides a robust and scalable solution for query analyzing tasks.

Author 1: Sridevi K N
Author 2: Rajanna M

Keywords: Hybrid clustering; information retrieval; mini-batch k-means; query analysis

PDF

Paper 88: Modified Moth-Flame Optimization Algorithm for Service Composition in Cloud Computing Environments

Abstract: Cloud computing service composition integrates services, distributed and diverse by nature, into an integrated entity that can meet a user's requirement with better effectiveness. However, some obstacles regarding high latency and suboptimal Quality of Service (QoS) still exist in a dynamic multi-cloud environment. This study addresses the limitations of traditional optimization algorithms in service composition, specifically the premature convergence and lack of population diversity in the Moth-Flame Optimization (MFO) algorithm. We propose the modified MFO algorithm with a new mechanism called Stagnation Finding and Replacement (SFR) to enhance the diversity of the population. It finds the static solutions based on a distance metric from globally optimal representative solutions and replaces them. MFO-SFR drastically improved all QoS metrics, such as response time, delay, and service stability. Empirical evaluations prove that MFO-SFR outperforms the baseline methods of multi-cloud service composition. It provides a computationally efficient and adaptive solution to cloud service composition problems, ensuring better resource utilization and higher user satisfaction in dynamic multi-cloud environments.

Author 1: Yeling YANG
Author 2: Miao SONG

Keywords: Cloud computing; quality of service; service composition; edge cloud; moth-flame optimization

PDF

Paper 89: Enhanced Task Scheduling Algorithm Using Harris Hawks Optimization Algorithm for Cloud Computing

Abstract: Amongst the most transformational technologies nowadays, cloud computing can provide resources such as CPU, memory, and storage over secure internet connections. Due to its flexibility and resource availability with guaranteed QoS, cloud computing allows comprehensive business and research adoptions. Despite the rapid development, resource management remains one of the significant challenges, especially handling task scheduling efficiently in this environment. Task scheduling strategically assigns tasks to available resources so that Quality of Service (QoS) metrics are effectively related to response time and throughput. This paper proposes an Enhanced Harris Hawks Optimization (EHHO) algorithm for scheduling cloud tasks to mitigate the common limitations found in existing algorithms. EHHO integrates a dynamic random walk strategy, enhancing exploration capabilities to avoid premature convergence and significantly improving scalability and resource allocation efficiency. Simulation outcomes reveal that EHHO minimizes makespan by up to 75%, memory usage by up to 60%, execution time by up to 39%, and cost by up to 66% compared to state-of-the-art algorithms. These benefits demonstrate that EHHO can optimize resource allocation while being highly scalable and reliable. Consistent performance over various stacks such as Kafka, Spark, Flink, and Storm further evidences the superiority of EHHO in handling complex scheduling challenges in dynamic cloud computing environments.

Author 1: Fang WANG

Keywords: Cloud computing; optimization; task scheduling; Harris Hawks Optimization; resource allocation; quality of service

PDF

Paper 90: PCE-BP: Polynomial Chaos Expansion-Based Bagging Prediction Model for the Data Modeling of Combine Harvesters

Abstract: With the rapid developments of measurement and monitoring techniques, massive amounts of in-situ data have been recorded and collected from the measurement system of combine harvesters in their working process and/or field experiments. However, the relationship between the operation parameters and the performance index such as clearing loss usually changes greatly in different sample subspaces, which makes it difficult for conventional prediction models to model the in-situ data, since most of them assume that the relationship is the same or similar throughout the whole sample space. Therefore, a polynomial chaos expansion-based bagging prediction model (PCE-BP) is proposed in this article. A polynomial chaos expansion-based decision tree is constructed to divide the sample space such that the relationship between the operation parameters and the performance index in the same part is more similar than the others, and bagging is used to ensemble the polynomial chaos expansion-based decision trees to reduce the perturbation and provide robust predictions. The experiments on the mathematical functions show that the proposed prediction model outperforms polynomial chaos expansion, polynomial chaos expansion-based decision tree, and the conventional bagging prediction model. The proposed prediction model is validated through two monitoring datasets from a combine harvester. The experimental results show that the PCE-BP model provides better cleaning loss and impurity rate prediction results than the other prediction models in most experiments, showing the advantages of sample space partitioning and bagging in the data modeling of combine harvesters.

Author 1: Liangyi Zhong
Author 2: Mengnan Deng
Author 3: Maolin Shi
Author 4: Ting Lou
Author 5: Shaoyang Zhu
Author 6: Jingwen Zhan
Author 7: Zishang Li
Author 8: Yi Ding

Keywords: Combine harvester; data modeling; polynomial chaos expansion; decision tree; bagging

PDF

Paper 91: Detecting Emotions with Deep Learning Models: Strategies to Optimize the Work Environment and Organizational Productivity

Abstract: This study proposes the implementation of a facial emotion recognition system based on Convolutional Neural Networks to detect emotions in real time, aiming to optimize the workplace environment and enhance organizational productivity. Six deep learning models were evaluated: Standard CNN, AlexNet, VGG16, InceptionV3, ResNet152 and DenseNet201, with DenseNet201 achieving the best performance, delivering an accuracy of 87.7% and recall of 96.3%. The system demonstrated significant improvements in key performance indicators (KPIs), including a 72.59% reduction in data collection time, a 63.4% reduction in diagnosis time, and a 66.59% increase in job satisfaction. These findings highlight the potential of Deep Learning technologies for workplace emotional management, enabling timely interventions and fostering a healthier, more efficient organizational environment.

Author 1: Cantuarias Valdivia Luis Alberto de Jesús
Author 2: Gómez Human Javier Junior
Author 3: Sierra-Liñan Fernando

Keywords: Facial recognition; real-time emotions; convolutional neural networks; work environment; artificial intelligence in human resources

PDF

Paper 92: Sentiment and Emotion Analysis with Large Language Models for Political Security Prediction Framework

Abstract: The increasing spread of textual content on social media, driven by the rise of Large Language Models (LLMs), has highlighted the importance of sentiment analysis in detecting threats, racial abuse, violence, and implied warnings. The subtlety and ambiguity of language present challenges in developing effective frameworks for threat detection, particularly within the political security domain. While significant research has explored hate speech and offensive content, few studies focus on detecting threats using sentiment analysis in this context. Leveraging advancements in Natural Language Processing (NLP), this study employs the NRC Emotion Lexicon to label emotions in a political-domain social media dataset. TextBlob is used to extract sentiment polarity, identifying potential threats where anger and fear intensities exceed a threshold alongside negative sentiment. The Bidirectional Encoder Representations from Transformers (BERT) was applied to enhance threat detection accuracy. The proposed framework achieved an Area Under the ROC Curve (AUC) of 87%, with the BERT model achieving 91% accuracy, 90.5% precision, 81.3% recall and F1-score of 91%, outperforming baseline models. These findings demonstrate the effectiveness of sentiment and emotion-based features in improving threat detection accuracy, providing a robust framework for political security applications.

Author 1: Liyana Safra Zaabar
Author 2: Adriana Arul Yacob
Author 3: Mohd Rizal Mohd Isa
Author 4: Muslihah Wook
Author 5: Nor Asiakin Abdullah
Author 6: Suzaimah Ramli
Author 7: Noor Afiza Mat Razali

Keywords: Political security; large language models; sentiment analysis; emotion analysis; BERT; threat prediction

PDF

Paper 93: Text-to-Image Generation Method Based on Object Enhancement and Attention Maps

Abstract: In the task of text-to-image generation, common issues such as missing objects in the generated images often arise due to the model's insufficient learning of multi-object category information and the lack of consistency between the text prompts and the generated image contents. To address these challenges, this paper proposes a novel text-to-image generation approach based on object enhancement and attention maps. First, a new object enhancement strategy is introduced to improve the model’s capacity to capture object-level features. The core idea is to generate difficult samples by processing the object mask maps of tokens, followed by dynamic weighting of the attention map using latent image embeddings. Second, to enhance the consistency between the text prompts and the generated image contents, we enforce similarity constraints between the cross-attention maps and the attention-weighted mask feature maps, penalizing inconsistencies through a loss function. Experimental results demonstrate that the Stable Diffusion v1.4 model, optimized using the proposed method, achieves significant improvements on the COCO instance dataset and the ADE20K instance dataset. Specifically, the MG metrics are improved by an average of 12.36% and 6.55%, respectively, compared to state-of-the-art models. Furthermore, the FID metrics show a 0.84% improvement over the state-of-the-art model on the COCO instance validation set.

Author 1: Yongsen Huang
Author 2: Xiaodong Cai
Author 3: Yuefan An

Keywords: Multi-object category; text-to-image generation; object enhancement; attention maps

PDF

Paper 94: Enhanced Traffic Congestion Prediction Using Attention-Based Multi-Layer GRU Model with Feature Embedding

Abstract: Intelligent Transportation Systems (ITS) are crucial for managing urban mobility and addressing traffic congestion, which poses significant challenges to modern cities. Traffic congestion leads to increased travel times, pollution, and fuel consumption, impacting both the environment and quality of life. Traditional traffic management solutions often fall short in predicting and adapting to dynamic traffic conditions. This study proposes an efficient deep learning (DL) model for predicting traffic congestion, utilizing the strengths of an attention-based multilayer Gated Recurrent Unit (GRU) network. The dataset used for this study includes 48,120 hourly vehicle counts across four junctions and additional weather data. Temporal and lagged features were engineered to capture daily and historical traffic trends and categorical data were considered by employing feature embedding. The attention-based GRU model integrates an attention mechanism to focus on relevant historical data, improving predictive performance by selectively emphasizing crucial time steps. This model architecture, consisting of two hidden layers and attention mechanisms, allows for nuanced traffic predictions by handling temporal dependencies and variations effectively. The performance was evaluated using various error metrics. The results demonstrate the model’s ability to predict traffic congestion with MSE of 0.9678, MAE of 0.4322, R² of 0.8686, MAPE of 6% offering valuable insights for traffic management and urban planning.

Author 1: Sreelekha M
Author 2: Midhunchakkaravarthy Janarthanan

Keywords: Intelligent transportation system; traffic congestion; urban mobility; deep learning; gated recurrent unit

PDF

Paper 95: Robust Joint Detection of Coronary Artery Plaque and Stenosis in Angiography Using Enhanced DCNN-GAN

Abstract: Timely detection and diagnosis of coronary artery segment plaque and stenosis in X-ray angiography is of great significance, however, the image quality variation, noise, and artifacts in the original image cause definitive difficulties to the current algorithms. These problems pose a challenge to meaningful analysis via traditional approaches, which compromises the efficiency of detection algorithms. To overcome these drawbacks, the current study presents a new integrated deep learning technique that integrates Deep Convolutional Neural Network (DCNN) with Generative Adversarial Network (GAN) in dual conditional detection. Detailed feature learning extracted from X-ray angiography images are performed through DCNN where it considers vascular structure and automatic pathologic regions detection. The use of GANs is to further enrich the dataset with synthetic images, distortions, and visual noise, which will make the model more immune to various conditions of images. Both approaches combined help in better classification of normal and pathological areas and less sensitiveness to quality of the obtained images. The proposed method therefore has shown an improvement of the diagnostic accuracy as a solid foundation for clinical decision making in cardiovascular systems. The efficacy of the suggested approach has been demonstrated by the following evaluation metrics: 97.9% F1 score, 98.7% accuracy, 98.2% precision, and 98% recall. The results prove higher sensitivity and accuracy of the plaque and stenosis identification comparing to the traditional methods, which confirms the efficiency of using the proposed DCNN-GAN method for considering the real-world fluctuations in the medical imaging. It reveals a decisive advancement in the ability to use algorithms for cardiovascular assessment by providing better results in difficult imaging environments.

Author 1: M. Jayasree
Author 2: L. Koteswara Rao

Keywords: DCNN-GAN; angiography; coronary artery plaque; stenosis; joint conditional detection

PDF

Paper 96: Design and Research of Accounting Automation Management System Based on Swarm Intelligence Algorithm and Deep Learning

Abstract: In the current research, the application verification of traditional algorithms in actual accounting management is insufficient, and deep learning data processing capabilities need to be fully optimized in complex accounting scenarios. Given the challenges of efficiency and accuracy faced by the current accounting industry in the context of big data, this study creatively combines the swarm intelligence algorithm and deep learning technology to design and implement an efficient and accurate accounting automation management system. The research aims to investigate the potential of swarm intelligence algorithms and deep learning techniques in developing an automated accounting management system, with a focus on improving efficiency, accuracy, and scalability. Key research questions include exploring the optimal configuration of swarm intelligence algorithms for accounting tasks and assessing the performance of deep learning models in automating various accounting processes. Through experimental verification, the system is tested with the financial data of a large enterprise for three consecutive years. The results show that the system can significantly shorten the time of financial statement generation by 65%, reduce the error rate to less than 0.5%, and increase the accuracy of abnormal data recognition by as much as 90%. These data not only reflect the significant improvement of the efficiency and accuracy of the system but also prove its great potential in early warning of financial risk, providing intelligent and automated solutions for the accounting industry.

Author 1: Dan Gui
Author 2: Wei Ma
Author 3: Wanfei Chen

Keywords: Swarm intelligence algorithm; deep learning; accounting; automation management

PDF

Paper 97: Enhancing Road Safety: A Multi-Modal Drowsiness Detection System for Drivers

Abstract: Driver drowsiness is a major contributing factor in road accidents, emphasizing the need for enhanced detection measures to improve car safety. This paper describes a multi-modal fatigue detection system that uses data from an internal camera, a front camera, and vehicle factors to reliably assess driver alertness. The technology outperforms traditional methods in terms of detection accuracy by utilizing powerful machine learning algorithms. Simulation and real-world tests show considerable improvements in reliability and performance. This integrated strategy offers a promising alternative for reducing the dangers associated with driver weariness and improving overall traffic safety.

Author 1: Guirrou Hamza
Author 2: Mohamed Zeriab Es-Sadek
Author 3: Youssef Taher

Keywords: Component; fatigue detection; drowsiness monitoring; ADAS

PDF

Paper 98: Decoding Face Attributes: A Modified AlexNet Model with Emphasis on Correlation-Heterogeneity Relationship Between Facial Attributes

Abstract: Face attribute estimation has several applications in computer vision, biometric systems, face verification /identification and image retrieval. The performance of face attribute estimation has been improved by using machine learning algorithms. In recent years, most algorithms have addressed this problem in multiple binary problem. Specifically, CNN-based approaches, which we can divide them into two classes; shared features and parts-based approaches. In shared features approach, the model uses two types of CNNs: one for feature extraction succeed by another one, for attribute classification. In the parts-based approaches, the approaches split the face image into multiple parts according to the geometric position of each attribute and train a CNN model for each part of the face. However, the shared features approach can handle attributes correlation but ignored attribute heterogeneity and gain in training time. On the other hand, the parts-based approaches can handle attributes heterogeneity but ignore attributes correlation and need more time in the training set compared with a shared feature approach. In this work, we propose a face attribute estimation method, which combined shared features and a parts-based approach into one model. Our model splits the input face image into five parts: whole image part, face part, face upper part, lower part, and nose part. In the same manner, the face attributes are subdivided into five groups according to the geometric position in the face image. We train shared feature model for each part, and we proposed an algorithm for feature selection task followed by AdaBoost algorithm to handle attribute classification task. Through a set of experiments using the LFWA and IIITM Face Emotion datasets, we demonstrate that our approach shows higher efficiency of face attribute estimation compared with the state-of-the art methods.

Author 1: Abdelaali Benaiss
Author 2: Otman Maarouf
Author 3: Rachid El Ayachi
Author 4: Mohamed Biniz
Author 5: Mustapha Oujaoura

Keywords: Face attribute estimation; biometrics; Convolutional Neural Network (CNN); face verification; computer vision

PDF

Paper 99: Evaluation of Eye Movement Features and Visual Fatigue in Virtual Reality Games

Abstract: VR games make people happy physically and mentally, but also lead to eye health problems. At present, the existing VR systems lack fatigue detection technology, which makes it difficult to help users use their eyes reasonably. In order to improve the user experience of VR gamers, this paper proposes a visual fatigue detection algorithm based on eye movement features, which uses the relationship between the lateral and longitudinal displacements of the human head and the displacement of the center point of the human eye to locate the position of the human eye. Moreover, in this paper, the human eye position tracking model is input into the three-frame difference algorithm to detect eye movement features. In addition, for tiny motion interference such as eyebrows, the image opening operation of eroding first and then expanding is used to remove it. Through experiments, it is found that the eye movement feature detection method adopted in this paper can greatly improve the detection speed with less accuracy loss, meet the sensitivity requirements of eye movement feature capture, improve the real-time performance of the system, and effectively improve the real-time analysis of player status. Therefore, integrating this algorithm into the virtual game system can help players adjust their own state, which has a positive effect on improving the game experience and reducing eye damage.

Author 1: Yuwei Ji

Keywords: Virtual reality; games; eye movement features; visual fatigue

PDF

Paper 100: High-Accuracy Vehicle Detection in Different Traffic Densities Using Improved Gaussian Mixture Model with Cuckoo Search Optimization

Abstract: Background subtraction plays a critical role in computer vision, particularly in vehicle detection and tracking. Traditional Gaussian Mixture Models (GMM) face limitations in dynamic traffic scenarios, leading to inaccuracies. This study proposes an Improved GMM with adaptive time-varying learning rates, exponential decay, and outlier processing to enhance performance across light, moderate, and heavy traffic densities. The model's parameters are automatically optimized using the Cuckoo Search algorithm, improving adaptability to varying environmental conditions. Validated on the ChangeDetection.net 2014 dataset, the Improved GMM achieves superior precision, recall, and F-measure compared to existing methods. Its consistent performance across diverse traffic scenarios highlights its effectiveness for real-time traffic flow analysis and vehicle detection applications.

Author 1: Nor Afiqah Mohd Aris
Author 2: Siti Suhana Jamaian

Keywords: Gaussian mixture model; vehicle detection; adaptive time-varying learning rate; exponential decay; outlier processing; cuckoo search optimization

PDF

Paper 101: PSR: An Improvement of Lightweight Cryptography Algorithm for Data Security in Cloud Computing

Abstract: Data security in cloud storage is a pressing concern as organizations increasingly rely on cloud computing services. Transitioning to cloud-based solutions underscores the need to safeguard sensitive information against data breaches and unauthorized access. Traditional cryptography algorithms are vulnerable to brute-force attacks and mathematical breakthroughs, necessitating large key sizes for security. Moreover, they lack resilience against emerging quantum computing threats, posing a significant risk to encryption. To tackle these issues, this study presents a novel lightweight cryptography algorithm named as PSR which is aimed at encryption so as to improve data security before storage in cloud systems. The proposed system converts 128 bit plaintext to cipher by employing techniques such as substitution, ASCII and hexadecimal conversions, block-wise transformations including Rail Fence, Grey Code, and XOR operations with random prime numbers. Notably, the proposed algorithm demonstrates superior performance with minimal runtime and memory usage, satisfying the avalanche effect criterion with a noteworthy efficacy in all executions and resistant to brute force attack.

Author 1: P. Sri Ram Chandra
Author 2: Syamala Rao
Author 3: Naresh K
Author 4: Ravisankar Malladi

Keywords: Cryptography; cloud security; PSR; encryption; decryption; avalanche effect

PDF

Paper 102: Optimizing Feature Selection in Intrusion Detection Systems Using a Genetic Algorithm with Stochastic Universal Sampling

Abstract: The current study presents a hybrid framework integrating the Genetic optimization algorithm with Stochastic Universal Sampling (GA-SUS) for feature selection and Deep Q-Networks (DQN) for fine-tuning an ensemble of classifiers to enhance network intrusion detection. The proposed method enhances genetic algorithms with stochastic universal sampling (GA-SUS) combined with recursive feature elimination (RFE). An ensemble of machine learning methods which includes gradient boosting and XG boost as base learners and subsequently logistic regression as meta learner is developed. A deep Q-networks (DQN) is used to optimize the base algorithms XG boost and gradient boost. The suggested method attains an accuracy of 97.60% on the popular NSL-KDD dataset and proficiently detects several attack types, such as probe attacks and Denial of Service (DoS), while tackling the issue of class imbalance. The multi-objective optimization approach is evident in anomaly detection and enhances model generalization by diminishing susceptibility to fluctuations in training data. Nonetheless, the model's efficacy regarding infrequent attack types, such as User to Root (U2R), remains inadequate due to their sparse representation in the dataset.

Author 1: RadhaRani Akula
Author 2: GS Naveen Kumar

Keywords: GA-SUS; anomaly detection; IDS; RFE; DQN

PDF

Paper 103: Optimizing Route Planning for Autonomous Electric Vehicles Using the D-Star Lite Algorithm

Abstract: Every vehicle, including autonomous vehicles, requires a route to navigate its journey. Route planning is a critical aspect of autonomous vehicle operations, as these vehicles rely on guided paths or sequential steps to move effectively. Ensuring that the route is optimal is a key consideration. This study tests the D-Star Lite algorithm to determine the most efficient route. In simulation tests, the D-Star Lite algorithm was compared with the A-Star algorithm. The results showed that D-Star Lite outperformed A-Star, achieving an average distance reduction of 124 meters. Real-time testing involved finding a route from node 36 to node 0, resulting in a total distance of 803 meters. Additional tests focused on route replanning in real-time scenarios. For instance, the initial route passing through nodes 36 →37→38→39→40→41→42→43→44→45→0 was adjusted to an alternative route: 36→37→38→46→26→11→2→4→1→0. Based on the results, the D-Star Lite algorithm proves effective in identifying the best route for autonomous electric vehicles while also enabling real-time route replanning.

Author 1: Bhakti Yudho Suprapto
Author 2: Suci Dwijayanti
Author 3: Desi Windisari
Author 4: Gatot Aria Pratama

Keywords: Autonomous vehicle; D-Star Lite; path planning; realtime; replanning route; optimal route

PDF

Paper 104: Stacking Regressor Model for PM2.5 Concentration Prediction Based on Spatiotemporal Data

Abstract: This study presents the development of a predictive model for PM2.5 concentrations resulting from forest and peatland fires in Riau Province, utilizing the stacking regressor technique within an ensemble learning framework. The model integrates spatiotemporal data from remote sensing and ground-based sensors at a resolution of 1 km x 1 km, demonstrating its effectiveness in capturing the intricate patterns of PM2.5 concentrations. By combining Random Forest, Gradient Boosting Machine (GBM), and XGBoost, with RidgeCV as a meta-learner, the model attained optimal performance, achieving R² = 0.851, MAE = 0.045 µg/m³, and MSE = 0.003 µg/m³. The incorporation of temporal feature engineering techniques, including lag and rolling window methods, significantly enhanced prediction accuracy, enabling the model to effectively capture seasonal variations and temporal dynamics. Key variables, such as air temperature, evapotranspiration, and Aerosol Optical Depth (AOD), were found to exhibit strong correlations with PM2.5 concentrations. The findings from this research contribute to the formulation of data-driven policies for air quality management and pollution mitigation, with the potential for broader application in regions encountering similar environmental challenges.

Author 1: Mitra Unik
Author 2: Imas Sukaesih Sitanggang
Author 3: Lailan Syaufina
Author 4: I Nengah Surati Jaya

Keywords: Ensemble learning; PM2.5 prediction; remote sensing; stacking regressor; spatio-temporal data

PDF

Paper 105: Feature Substitution Using Latent Dirichlet Allocation for Text Classification

Abstract: Text classification plays a pivotal role in natural language processing, enabling applications such as product categorization, sentiment analysis, spam detection, and document organization. Traditional methods, including bag-of-words and TF-IDF, often lead to high-dimensional feature spaces, increasing computational complexity and susceptibility to overfitting. This study introduces a novel Feature Substitution technique using Latent Dirichlet Allocation (FS-LDA), which enhances text representation by replacing non-overlapping high-probability topic words. FS-LDA effectively reduces dimensionality while retaining essential semantic features, optimizing classification accuracy and efficiency. Experimental evaluations on five e-commerce datasets and an SMS spam dataset demonstrated that FS-LDA, combined with Hidden Markov Models (HMMs), achieved up to 95% classification accuracy in binary tasks and significant improvements in macro and weighted F1-scores for multiclass tasks. The innovative approach lies in FS-LDA's ability to seamlessly integrate dimensionality reduction with feature substitution, while its predictive advantage is demonstrated through consistent performance enhancement across diverse datasets. Future work will explore its application to other classification models and domains, such as social media analysis and medical document categorization, to further validate its scalability and robustness.

Author 1: Norsyela Muhammad Noor Mathivanan
Author 2: Roziah Mohd Janor
Author 3: Shukor Abd Razak
Author 4: Nor Azura Md. Ghani

Keywords: Feature extraction; feature selection; Latent Dirichlet Allocation; text classification; Hidden Markov Model; dimensionality reduction

PDF

Paper 106: Multilabel Classification of Bilingual Patents Using OneVsRestClassifier: A Semiautomated Approach

Abstract: In response to the increasing complexity and volume of patent applications, this research introduces a semiautomated system to streamline the literature review process for Indonesian patent data. The proposed system employs a synthesis of multilabel classification techniques based on natural language processing (NLP) algorithms. This methodology focuses on developing an iterative and modular system, with each step visualised in detailed flowcharts. The system design incorporates data collection and preprocessing, multilabel classification model development, model optimisation, query and prediction, and results presentation modules. Experimental results demonstrate the promising potential of the multilabel classification model, achieving a micro F1 score of 0.6723 and a macro F1 score of 0.6009. The OneVsRestClassifier model with LinearSVC as the base classifier shows reasonably good performance in handling a bilingual dataset comprising 15,097 patent documents. The optimal model configuration uses TfidfVectorizer with 20,000 features, including bigrams, and an optimal C parameter of 0.1 for LinearSVC. Performance analysis reveals variations across IPC classes, indicating areas for further improvement. The discussion highlights the implications of the proposed system for researchers, patent examiners and industry professionals by facilitating efficient searches within patent databases. This study acknowledges the potential of semiautomated systems to enhance the efficiency of patent analysis while emphasising the need for further research to address identified challenges, such as class imbalance and performance variations across patent categories. This research paves the way for further developments in the field of automated patent classification, aiming to improve efficiency and accuracy in international patent systems while recognising the crucial role of human experts in the patent classification process.

Author 1: Slamet Widodo
Author 2: Ermatita
Author 3: Deris Stiawan

Keywords: Multilabel patent classification; Natural Language Processing (NLP); OneVsRestClassifier; TF–IDF vectorisation; bilingual patent analysis

PDF

Paper 107: Dolphin Inspired Optimization for Feature Extraction in Augmented Reality Tracking

Abstract: Feature extraction has the prominent role in Augmented Reality (AR) tracking. AR tracking monitor the position and orientation to overlay the 3D model in real-world environment. This approach of AR tracking, encouraged to propose the optimum feature extraction model by embedding the dolphin grouping system. We implemented dolphin grouping algorithm to extract the features effectively without compromising the accuracy. In addition, to prove the stability of the proposed model, we have included the affine transformation images such as rotation, blur image and light variation for the analysis. The Dolphin model obtained the average precision of 0.92 and recall score of 0.84. Whereas, the computation time of dolphin model is identified as 2ms which is faster than the other algorithm. The comparative result analysis reveals that accuracy and the efficiency of the proposed model surpasses the existing descriptors.

Author 1: Indhumathi S
Author 2: Christopher Clement J

Keywords: Feature descriptor; dolphin optimization; feature extraction; augmented reality tracking

PDF

Paper 108: Empirical Analysis of Variations of Matrix Factorization in Recommender Systems

Abstract: Recommender systems recommend products to users. Almost all businesses utilize recommender systems to suggest their products to customers based on the customer’s previous actions. The primary inputs for recommendation algorithms are user preferences, product descriptions, and user ratings on products. Content-based recommendations and collaborative filtering are examples of traditional recommendation systems. One of the mathematical models frequently used in collaborative filtering is matrix factorization (MF). This work focuses on discussing five variants of MF namely Matrix Factorization, Probabilistic MF, Non-negative MF, Singular Value Decomposition (SVD), and SVD++. We empirically evaluate these MF variants on six benchmark datasets from the domains of movies, tourism, jokes, and e-commerce. MF is the least performing and SVD is the best-performing method among other MF variants in terms of Root Mean Square Error (RMSE).

Author 1: Srilatha Tokala
Author 2: Murali Krishna Enduri
Author 3: T. Jaya Lakshmi
Author 4: Koduru Hajarathaiah
Author 5: Hemlata Sharma

Keywords: Recommendations; matrix factorization; content-based; collaborative filtering; RMSE

PDF

Paper 109: Efficient Tumor Detection in Medical Imaging Using Advanced Object Detection Model: A Deep Learning Approach

Abstract: Timely and accurate tumor detection in medical imaging is crucial for improving patient outcomes and reducing mortality rates. Traditional methods often rely on manual image interpretation, which is time-intensive and prone to variability. Deep learning, particularly convolutional neural networks (CNNs), has revolutionized tumor detection by automating the process and achieving remarkable accuracy. The present paper investigates the use of YOLOv11, a powerful object detection model, for tumor detection in several medical imaging modalities, such as CT scans, MRIs, and histopathological images. YOLOv11 incorporates architectural advancements, including enhanced feature pyramids and attention processes, allowing accurate identification of tumors with diverse sizes and complexity. The model’s real-time detection capabilities and lightweight architecture render it appropriate for use in clinical settings and resource-limited contexts. Experimental findings indicate that the fine-tuned YOLOv11 attains exceptional accuracy and efficiency, exhibiting an average precision of 91% and a mAP of 68%. This research highlights YOLOv11’s significance as a transformational instrument in the integration of AI in medical imaging, aimed at optimizing diagnostic processes and improving healthcare delivery.

Author 1: Taoufik Saidani

Keywords: Tumor detection; medical imaging; YOLOv11; deep learning; real-time detection

PDF

Paper 110: Efficient Anomaly Detection Technique for Future IoT Applications

Abstract: Internet of Things (IoT) provides smart wireless connectivity and is the basis of many future applications. IoT nodes are equipped with sensors that obtain application-related data and transmit to the servers using IEEE 802.15.4-based wire-less communications, thus forming a low-rate wireless personal area network. Security is a major challenge in IoT networks as malicious users can capture the network and waste the available bandwidth reserved for legitimate users, thus significantly reducing the Quality of Service (QoS) in terms of transmitted data and transmission delay. In this work, an Anomaly Detection Mechanism for IEEE 802.15.4 standard (ADM15.4) to improve the QoS of the IoT Nodes is proposed. ADM15.4 also proposes a mechanism to block the malicious nodes without affecting the overall performance of the medium. The performance of ADM15.4 is compared with the standard when there is no such anomaly detection is present. The results are obtained for different values of SO and for different sets of GTS requesting nodes and are compared with the standard in the presence and absence of malicious nodes. The simulation results show that the ADM15.4 improves data transmission up to 19.5% from IEEE 802.15.4 standard without attacks and up to 52% when there is malicious attacks. Furthermore, ADM15.4 transmits data 33% reduced time and accommodate 56% more GTS requesting legitimate nodes as compared to the standard in the presence of the malicious attacks.

Author 1: Ahmad Naseem Alvi
Author 2: Muhammad Awais Javed
Author 3: Bakhtiar Ali
Author 4: Mohammed Alkhathami

Keywords: Anomaly detection; IoT networks; security

PDF

Paper 111: GRACE: Graph-Based Attention for Coherent Explanation in Fake News Detection on Social Media

Abstract: Detecting fake news on social media is a critical challenge due to its rapid dissemination and potential societal impact. This paper addresses the problem in a realistic scenario where the original tweet and the sequence of users who retweeted it, excluding any comment section, are available. We propose a Graph-based Attention for Coherent Explanation (GRACE) to perform binary classification by determining if the original tweet is false and provide interpretable explanations by highlighting suspicious users and key evidential words. GRACE integrates user behaviour, tweet content, and retweet propagation dynamics through Graph Convolutional Networks (GCNs) and a dual co-attention mechanism. Extensive experiments conducted on Twitter15 and Twitter16 datasets demonstrate that GRACE out-performs baseline methods, achieving an accuracy improvement of 2.12% on Twitter15 and 1.83% on Twitter16 compared to GCAN. Additionally, GRACE provides meaningful and coherent explanations, making it an effective and interpretable solution for fake news detection on social platforms.

Author 1: Orken Mamyrbayev
Author 2: Zhanibek Turysbek
Author 3: Mariam Afzal
Author 4: Marassulov Ussen Abdurakhimovich
Author 5: Ybytayeva Galiya
Author 6: Muhammad Abdullah
Author 7: Riaz Ul Amin

Keywords: Graph neural network; dual attention; NLP; semantics; social network

PDF

Paper 112: Intelligent Fault Diagnosis for Elevators Using Temporal Adaptive Fault Network

Abstract: Contemporary cities depend on elevators for vertical mobility in residential, commercial, and industrial buildings. However, elevator system malfunctions may cause operational interruptions, economic losses, and safety dangers, requiring advanced tools for detection. High-dimensional sensor data, temporal interdependence, and fault dataset imbalances are common problems in fault detection algorithms. These restrictions reduce fault diagnostic accuracy and reliability, especially in real-time applications. This paper presents a Temporal Adaptive Fault Network (TAFN) to overcome these issues. The system uses Temporal Convolution Layers to capture sequential dependencies, Adaptive Feature Refinement Layers to dynamically improve feature relevance, and a Fault Decision Head for correct classification. For reliable performance, the Weighted Divergence Analyzer and innovative data processing methods are used for feature selection. Experimental findings show that the TAFN model outperforms state-of-the-art fault classification approaches with an F1-score of 98.5% and an AUC of 99.3%. The model’s capacity to handle unbalanced datasets and complicated temporal patterns makes it useful in real life. The paper also proposes the Fault Temporal Sensitivity Index (FTSI) to assess fault prediction temporal consistency. The results demonstrate that TAFN may revolutionize elevator problem detection, improving reliability, downtime, and safety. This technique advances predictive maintenance tactics for critical infrastructure.

Author 1: Zhiyu Chen

Keywords: Elevator fault diagnosis; temporal adaptive fault network; predictive maintenance; multivariate time-series data; feature refinement; fault classification

PDF

Paper 113: High-Precision Multi-Class Object Detection Using Fine-Tuned YOLOv11 Architecture: A Case Study on Airborne Vehicles

Abstract: The widespread adoption of airborne vehicles, including drones and UAVs, has brought significant advancements to fields such as surveillance, logistics, and disaster response. Despite these benefits, their increasing use poses substantial challenges for real-time detection and classification, particularly in multi-class scenarios where precision and scalability are essential. This paper proposes a high-performance detection framework based on YOLOv11, specifically tailored for identifying airborne vehicles. YOLOv11 integrates innovative features, such as anchor-free detection and enhanced attention mechanisms, to deliver superior accuracy and speed. The proposed framework is tested on a comprehensive airborne vehicle dataset featuring diverse conditions, including variations in altitude, occlusion, and environmental factors. Experimental results demonstrate that the fine-tuned YOLOv11 model exceeds the performance of existing models. Additionally, its ability to operate in real-time makes it ideal for critical applications like air traffic management and security monitoring.

Author 1: Nasser S. Albalawi

Keywords: Airborne vehicles; YOLOv11; object detection; surveillance

PDF

Paper 114: AI-Driven Image Recognition System for Automated Offside and Foul Detection in Football Matches Using Computer Vision

Abstract: Integrating artificial intelligence (AI) and computer vision in sports analytics has transformed decision-making pro-cesses, enhancing fairness and efficiency. This paper proposes a novel AI-driven image recognition system for automatically detecting offside and foul events in football matches. Unlike conventional methods, which rely heavily on manual intervention or traditional image processing techniques, our approach utilizes a hybrid deep learning model that combines advanced object tracking with motion analysis to deliver real-time, precise event detection. The system employs a robust, self-learning algorithm that leverages spatiotemporal features from match footage to track player movements and ball dynamics. By analyzing the continuous flow of video data, the model detects offside positions and identifies foul types such as tackles, handballs, and dangerous play—through a dynamic pattern recognition process. This multi-tiered approach overcomes traditional methods’ limitations by accurately identifying critical events with minimal latency, even in complex, high-speed scenarios. In experiments conducted on diverse datasets of live match footage, the system achieved an overall accuracy of 99.85% for offside detection and 98.56%for foul identification, with precision rates of 98.32% and 97.12%, respectively. The system’s recall rates of 97.45% for offside detection and 96.85% for foul recognition demonstrate its reliability in real-world applications. It’s clear from these results that the proposed framework can automate and greatly enhance the accuracy of match analysis, making it a useful tool for both referees and broadcasters. The system’s low computational overhead and growing ability make connecting to existing match broadcasting infrastructure easy. This establishes an immediate feedback loop for use during live games. This work marks a significant step forward in applying AI and computer vision for sports, introducing a powerful method to enhance the objectivity and precision of officiating in football.

Author 1: Qianwei Zhang
Author 2: Lirong Yu
Author 3: WenKe Yan

Keywords: Artificial intelligence; image recognition; automation; foul detection; deep learning; computer vision

PDF

Paper 115: Deep Q-Learning-Based Optimization of Path Planning and Control in Robotic Arms for High-Precision Computational Efficiency

Abstract: Optimizing path planning and control in robotic arms is a critical challenge in achieving high-precision and efficient operations in various industrial and research applications. This study proposes a novel approach leveraging deep Q-learning (DQL) to enhance robotic arm movements’ computational efficiency and precision. The proposed framework effectively ad-dresses key challenges such as collision avoidance, path smooth-ness, and dynamic control by integrating reinforcement learning techniques with advanced kinematic modelling. To validate the effectiveness of the proposed method, a simulated environment was developed using a 6-degree-of-freedom robotic arm, where the DQL model was trained and tested. Results demonstrated a significant performance improvement, achieving an average path optimization accuracy of 98.76% and reducing computational overhead by 22.4% compared to traditional optimization methods. Additionally, the proposed approach achieved real-time response capabilities, with an average decision-making latency of 0.45 seconds, ensuring its applicability in time-critical scenarios. This research highlights the potential of deep Q-learning in revolutionizing robotic arm control by combining precision and computational efficiency. The findings bridge gaps in robotic path planning and pave the way for future advancements in autonomous robotics and industrial automation. Further studies can explore the scalability of this approach to more complex and real-world environments, solidifying its relevance in emerging technological domains.

Author 1: Yuan Li
Author 2: Byung-Won Min
Author 3: Haozhi Liu

Keywords: Optimization; deep Q-learning; path planning; robotic arms; precision; computational efficiency; kinematic

PDF

Paper 116: Android Malware Detection Through CNN Ensemble Learning on Grayscale Images

Abstract: With Android’s widespread adoption as the leading mobile operating system, it has become a prominent target for malware attacks. Many of these attacks employ advanced obfuscation techniques, rendering traditional detection methods, such as static and dynamic analysis, less effective. Image-based approaches provide an alternative for effective detection that addresses some limitations of conventional methods. This re-search introduces a novel image-based framework for Android malware detection. Using the CICMalDroid 2020 dataset, Dalvik Executable (DEX) files from Android Package (APK) files are extracted and converted into grayscale images, with dimensions scaled according to file size to preserve structural characteristics. Various Convolutional Neural Network (CNN) models are then employed to classify benign and malicious applications, with performance further enhanced through a weighted voting ensemble optimized by Bayesian Optimization to balance the contribution of each model. An ablation study was conducted to demonstrate the effectiveness of the six-model ensemble, showing consistent improvements in accuracy as models were added incrementally, culminating in the highest accuracy of 99.3%. This result surpasses previous research benchmarks in Android malware detection, validating the robustness and efficiency of the proposed methodology.

Author 1: El Youssofi Chaymae
Author 2: Chougdali Khalid

Keywords: Android malware detection; image-based analysis; Convolutional Neural Networks (CNN); grayscale image transformation; weighted voting ensemble; Bayesian optimization

PDF

Paper 117: Cross-Domain Health Misinformation Detection on Indonesian Social Media

Abstract: Indonesia is among the world’s most prolific countries in terms of internet and social media usage. Social media serves as a primary platform for disseminating and accessing all types of information, including health-related data. However, much of the content generated on these platforms is unverified and often falls into the category of misinformation, which poses risks to public health. It is essential to ensure the credibility of the information available to social media users, thereby helping them make informed decisions and reducing the risks associated with health misinformation. Previous research on health misinformation detection has predominantly focused on English-language data or has been limited to specific health crises, such as COVID-19. Consequently, there is a need for a more comprehensive approach which not only focus on single issue or domain. This study proposes the development of a new corpus that encompasses various health topics from Indonesian social media. Each piece of content within this corpus will be manually annotated by expert to label a social media post as either misinformation or fact. Additionally, this research involves experimenting with machine learning models, including traditional and deep learning models. Our finding shows that the new cross-domain dataset is able to achieve better performance compared to those trained on the COVID dataset, highlighting the importance of diverse and representative training data for building robust health misinformation detection system.

Author 1: Divi Galih Prasetyo Putri
Author 2: Savitri Citra Budi
Author 3: Arida Ferti Syafiandini
Author 4: Ikhlasul Amal
Author 5: Revandra Aryo Dwi Krisnandaru

Keywords: Health misinformation; machine learning; social media

PDF

Paper 118: Comparison of Machine Learning Algorithms for Malware Detection Using EDGE-IIoTSET Dataset in IoT

Abstract: The growth of IoT devices has presented great vulnerabilities leading to many malware attacks. Existing IoT malware detection methods face many challenges; including: device heterogeneity, device resource restrictions, and the complexity of encrypted malware payloads, thus leading to less effective conventional cybersecurity techniques. This study’s objective is to reduce these gaps by assessing the results obtained from testing five machine learning algorithms that are used to detect IoT malware by applying them on the EDGE-IIoTSET dataset. Key preprocessing steps include: cleaning data, extracting features, and encoding network traffic. Several algorithms used these include: Logistic Regression, Decision Tree, Na¨ıve Bayes, KNN, and Random Forest. The Decision Tree model achieved perfect accuracy at 100%, making it the best-performing model for this analysis. In contrast, Random Forest delivered a strong performance with an accuracy of 99.9%, while Logistic Regression performed at 27%, Na¨ıve Bayes at 57%, and KNN with moderate performance. Hence, the results have shown the effectiveness of machine learning techniques to enhance the security IoT systems regarding real-time malware detection with high accuracy. These findings are useful input for policymakers, cybersecurity practitioners, and IoT developers as they develop better mechanisms for handling dynamic IoT malware attack incidents.

Author 1: Jawaher Alshehri
Author 2: Almaha Alhamed
Author 3: Mounir Frikha
Author 4: M M Hafizur Rahman

Keywords: IoT malware; machine learning; malware detection; IoT security; EDGE-IIoTSET

PDF

Paper 119: Building Detection from Satellite Imagery Using Morphological Operations and Contour Analysis over Google Maps Roadmap Outlines

Abstract: One such research area is building detection, which has a high influence and potential impact in urban planning, disaster management, and construction development. Classifying buildings using satellite images is a difficult task due to building designs, shapes, and complex backgrounds which lead to occlusion between buildings. The current study introduces a new method for constructing recognition and classification globally based on Google Maps contour trace detection and an evolved image processing technique, seeking synergies with a systematic methodology. We first extract the building outlines by taking the image from the ¨Roadmap¨view in Google Maps, converting it to gray scale, thresholding it to create binary boundaries,and finally applying morphological operations to facilitate noise removal and gap filling. These binary outlines are overlaid on colorful satellite imagery, which aids in identifying buildings. Machine learning techniques can also be used to improve aspect ratio analysis and improve overall detection accuracy and performance.

Author 1: Arbab Sufyan Wadood
Author 2: Ahthasham Sajid
Author 3: Muhammad Mansoor Alam
Author 4: Mazliham MohD Su’ud
Author 5: Arshad Mehmood
Author 6: Inam Ullah Khan

Keywords: Building detection; satellite imagery; urban planning; disaster response; image processing; machine learning; morphological operations; contour detection; aspect ratio

PDF

Paper 120: Exploring Machine Learning in Malware Analysis: Current Trends and Future Perspectives

Abstract: Sophisticated cyberattacks are an increasing concern for individuals, businesses, and governments alike. Detecting malware remains a significant challenge, particularly due to the limitations of traditional methods in identifying new or unexpected threats. Machine Learning (ML) has emerged as a powerful solution, capable of analyzing large datasets, recognizing complex patterns, and adapting to rapidly changing attack strategies. This paper reviews the latest advancements in machine learning for malware analysis, shedding light on both its strengths and the challenges it faces. Additionally, it explores the current limitations of these approaches and outlines future research directions. Key recommendations include improving data preprocessing techniques to reduce information loss, utilizing distributed computing for greater efficiency, and maintaining balanced, up-to-date datasets to enhance model reliability. These strategies aim to improve the scalability, accuracy, and resilience of ML-driven malware detection systems.

Author 1: Noura Alyemni
Author 2: Mounir Frikha

Keywords: Machine learning; malware analysis; cybersecurity

PDF

Paper 121: SM9 Key Encapsulation Mechanism for Power Monitoring Systems

Abstract: The boundaries of the new power system network are blurred, and data privacy and security are threatened. Although the SM9 algorithm is widely used in power systems to protect data security, its efficiency and security remain the main issues in application. Therefore, an SM9 key encapsulation mechanism (OSM9-KEM-CRF) was proposed to support outsourced decryption and cryptographic reverse firewall. In order to resist the backdoor attacks, we deployed cryptographic reverse firewalls at the terminals and proved that the proposed OSM9-KEM-CRF is ID-IND-CCA2 secure. The cryptographic reverse firewalls maintain functionality, weakly retain security, and weakly resist penetration, thereby enhancing the security of the scheme. In addition, considering the limited computing resources of terminal devices, decryption operations are outsourced to cloud servers in order to reduce the computational burden on the terminals. Compared with other SM9-KEMs, the proposed mechanism not only reduces computational and communication overhead, but also lowers energy consumption. Therefore, the proposed mechanism is more suitable for power monitoring systems.

Author 1: Chao Hong
Author 2: Peng Xiao
Author 3: Pandeng Li
Author 4: Zhenhong Zhang
Author 5: Yiwei Yang
Author 6: Biao Bai

Keywords: SM9; Outsourced decryption; cryptographic reverse firewall; power monitoring systems

PDF

Paper 122: A Review of Analyzing Different Agricultural Crop Yields Using Artificial Intelligence

Abstract: The advancement of Artificial Intelligence (AI), in particular Deep Learning (DL), has made it possible to interpret gathered data more quickly and effectively in this new digital era. To draw attention to development advancements in deep learning across many industries. Agriculture has been one of the most affected areas in recent advancements of the current globalized world agriculture plays a vital role and makes significant contributions. Over the years, agriculture has faced several difficulties in meeting the growing demands of the global people, which has creased over the last 50 years. Different forecasts have been made regarding this extraordinary population expansion which is expected to grasp almost 9 billion persons worldwide by 2050. More than a century ago, different technologies were brought into agriculture to solve issues related to crop cultivation. Many mechanical technologies are accessible today, and they are evolving at an amazing rate. To support their demands and help them optimize their crop yields based on data and task automation need innovative techniques to aid farmers. This will transform the agricultural industry into a new dimension. Therefore, this study’s primary goal was to present a thorough summary of the most current developments based on research interconnected with the digitization of agriculture for crop yields including fruit counting, crop management, water management, weed identification, soil management, seed categorization, disease detection, yield forecasting and harvesting of yields based on Artificial Intelligence Techniques.

Author 1: Vijaya Bathini
Author 2: K. Usha Rani

Keywords: Agriculture; artificial intelligence; deep learning; crop yields; management

PDF

Paper 123: LMS-YOLO11n: A Lightweight Multi-Scale Weed Detection Model

Abstract: With the advancement of precision agriculture, efficient and accurate weed detection has emerged as a pivotal task in modern crop management. Current weed detection methods face dual challenges: inadequate extraction of detailed features and edge information, coupled with the necessity for real-time performance. To address these issues, this paper pro-poses a lightweight multi-scale weed detection model based on YOLOv11n (You-only-look-once-11). Our approach incorporates three innovative components: (1) A fast-gated lightweight unit combined with C3K2 to enhance local and global interaction capabilities of weed features. (2) An adaptive hierarchical feature fusion network based on HSFPN, which improves the extraction of weed edge information. (3) A lightweight group convolution detection head module that captures multi-scale feature details while maintaining a lightweight structure. Experimental validation on two public datasets, CottonWeedDet3 and CottonWeed2, demonstrates that our model achieves an mAP50 improvement of 2.5% on CottonWeedDet3 and 1.9% on CottonWeed2 compared to YOLOv11n, with a 37% reduction in parameters and a 26%decrease in computational effort.

Author 1: YaJun Zhang
Author 2: Yu Xu
Author 3: Jie Hou
Author 4: YanHai Song

Keywords: You-only-look-once-11; weed; lightweight; group convolution

PDF

Paper 124: DBYOLOv8: Dual-Branch YOLOv8 Network for Small Object Detection on Drone Image

Abstract: Object detection based on drone platforms is a valuable yet challenging research field. Although general object detection networks based on deep learning have achieved breakthroughs in natural scenes, drone images in urban environments often exhibit characteristics such as a high proportion of small objects, dense distribution, and significant scale variations, posing significant challenges for accurate detection. To address these issues, this paper proposes a dual-branch object detection algorithm based on YOLOv8 improvements. Firstly, an auxiliary branch is constructed by extending the YOLOv8 backbone to aggregate high-level semantic information within the network, enhancing the feature extraction capability. Secondly, a Multi-Branch Feature Enhancement (MBFE) module is designed to enrich the feature representation of small objects and enhance the correlation of local features. Third, Spatial-to-Depth Convolution (SPDConv) is utilized to mitigate the loss of small object information during downsampling, preserving more small object feature information. Finally, a dual-branch feature pyramid is designed for feature fusion to accommodate the dual-branch input. Experimental results on the VisDrone benchmark dataset demonstrate that DBYOLOv8 outperforms state-of-the-art object detection methods. Our proposed DBYOLOv8s achieve mAP@0.5 of 49.3% and mAP@0.5:0.95 of 30.4%, which are 2.8% and 1.5%higher than YOLOv9e, respectively.

Author 1: Yawei Tan
Author 2: Bingxin Xu
Author 3: Jiangsheng Sun
Author 4: Cheng Xu
Author 5: Weiguo Pan
Author 6: Songyin Dai
Author 7: Hongzhe Liu

Keywords: Drone images; dual-branch; small object detection; YOLOv8

PDF

Paper 125: Eagle Framework: An Automatic Parallelism Tuning Architecture for Semantic Reasoners

Abstract: Parallel semantic reasoners use parallel architectures to improve the efficiency of reasoning tasks. Studies in semantic reasoning rely on manual tuning to configure the degree of parallelism. However, manual tuning becomes increasingly challenging as ontologies become massive and complex. Studies in related fields have developed automatic tuning frameworks using optimization search methods. Although these methods offer performance gains, reducing search time and space size is still an open problem. This study aims to bridge the gap in semantic reasoning and the problem in existing search methods. To achieve these aims, we propose Eagle Framework (EF), an innovative automatic tuning framework designed to improve the performance of parallel semantic reasoners. EF automatically configures the degree of parallelism and calculates the performance data. It incorporates a novel search space and algorithm, inspired by the AVL tree, that efficiently identifies the optimal degree of parallelism. In a case study, EF completed the tuning processes in seconds to a few minutes, achieving performance gains up to 65 times faster than common search methods. The reliability findings, with ICC scores ranging from 0.90 to 0.99, confirmed the consistency of the performance data calculated by EF. The regression analysis revealed the effectiveness of EF in identifying the factors that affect reasoning scalability, with the conclusion that the size of the ontology is the dominant factor. The study underscores the need for adaptive approaches to tune the degree of parallelism based on the size of the ontology.

Author 1: Haifa Ali Al-Hebshi
Author 2: Muhammad Ahtisham Aslam
Author 3: Kawther Saeedi

Keywords: Automatic tuning; parallel semantic reasoning; performance optimization; ontology; high performance computing

PDF

Paper 126: Imbalance Datasets in Malware Detection: A Review of Current Solutions and Future Directions

Abstract: Imbalanced datasets are a significant challenge in the field of malware detection. The uneven distribution of malware and benign samples is a challenge for modern machine learning based detection systems, as it creates biased models and poor detection rates for malicious software. This paper provides a systematic review of existing approaches for dealing with imbalanced datasets in malware detection such as data-level, algorithm-level, and ensemble methods. We explore different techniques including Synthetic Minority Oversampling Technique, deep learning techniques including CNN and LSTM hybrids, Genetic Programming for feature selection, and Federated Learning. Furthermore, we assesses the strengths, weakness, and areas of application of each approach. Computational complexity, scalability, and the practical applicability of these techniques remains as challenges. Finally, the paper summarizes promising directions for future research like lightweight models and advanced sampling strategies to further improve the robustness and practicality of malware detection systems in dynamic environments.

Author 1: Hussain Almajed
Author 2: Abdulrahman Alsaqer
Author 3: Mounir Frikha

Keywords: Malware detection; machine learning; imbalance datasets; oversampling; SMOTE

PDF

Paper 127: Artificial Intelligence in Financial Risk Early Warning Systems: A Bibliometric and Thematic Analysis of Emerging Trends and Insights

Abstract: With the continuous development of financial markets worldwide, there has been increasing recognition of the importance of financial risk management. To mitigate financial risk, financial risk early warning serves as a risk uncovering mechanism enabling companies to anticipate and counter potential disruptions. The present review paper aims to identify the bibliometric analysis for exploring the growth and academic evolution of financial risk, financial risk management, and financial risk early warning concepts. Academic literature is surveyed from the Scopus database during the period 2010-2024. The network analysis, conceptual structure, and bibliographic analysis of the selected articles are employed using VOSviewer and Bibliometric R Package. The biblioshiny technique based on the bibliometric R package was used to draw journal papers’ performance and scientific contributions by displaying distinctive features from the bibliometric method used in prior studies. The data was extracted from Scopus databases. In addition, this study comprehensively analyzes the evolution of financial risk early warning systems, highlighting significant trends and future directions. Thematic evaluation across 2010-2015, 2016-2021, and 2022-2024 reveals a shift from traditional statistical methods to advanced machine learning and AI techniques, with neural networks, random forests, and XGBoost being pivotal. Innovations like attention mechanisms and LSTM models improve prediction accuracy. The integration of sustainability factors, such as carbon neutrality and renewable energy, reflects a trend towards incorporating environmental considerations into risk management. The study underscores the need for interdisciplinary collaborations and advanced data analytics for comprehensive financial systems. Policy implications include promoting AI adoption, integrating environmental factors, fostering collaborations, and developing advanced data analytics frameworks.

Author 1: Muhammad Ali Chohan
Author 2: Teng Li
Author 3: Suresh Ramakrishnan
Author 4: Muhammad Sheraz

Keywords: Artificial intelligence; deep learning; financial risk management; early warning systems; bibliometrics analysis

PDF

Paper 128: DBFN-J: A Lightweight and Efficient Model for Hate Speech Detection on Social Media Platforms

Abstract: Hate speech on social media platforms like YouTube, Facebook, and Twitter threatens online safety and societal harmony. Addressing this global challenge requires innovative and efficient solutions. We propose DBFN-J (DistillBERT-Feedforward Neural Network with Jaya optimization), a lightweight and effective algorithm for detecting hate speech. This method combines DistillBERT, a distilled version of the Bidirectional Encoder Representations from Transformers (BERT), with a Feedforward Neural Network. The Jaya algorithm is employed for parameter optimization, while aspect-based sentiment analysis further enhances model performance and computational efficiency. DBFN-J demonstrates significant improvements over existing methods such as CNN BERT (Convolutional Neural Network BERT), BERT-LSTM (Long Short-Term Memory), and ELMo (Embeddings from Language Models). Extensive experiments reveal exceptional results, including an AUC (Area Under the Curve) of 0.99, a log loss of 0.06, and a balanced F1-score of 0.95. These metrics underscore its robust ability to identify abusive content effectively and efficiently. Statistical analysis further confirms its precision (0.98) and recall, making it a reliable tool for detecting hate speech across diverse social media platforms. By outperforming traditional algorithms in both performance and resource utilization, DBFN-J establishes a new benchmark for hate speech detection. Its lightweight design ensures suitability for large-scale, resource-constrained applications. This research provides a robust framework for protecting online environments, fostering healthier digital spaces, and mitigating the societal harm caused by hate speech.

Author 1: Nourah Fahad Janbi
Author 2: Abdulwahab Ali Almazroi
Author 3: Nasir Ayub

Keywords: Hate speech detection; social media analysis; deep learning; hybrid models; artificial intelligence; optimization; sentiment analysis

PDF

Paper 129: Exploring the Best Machine Learning Models for Breast Cancer Prediction in Wisconsin

Abstract: This research focuses on predicting Wisconsin Breast Cancer Disease using machine learning algorithm, employs a dataset offered by UCI repository (WBCD) dataset. The under- gone substantial preparation, includes managing missing values, normalization, outlier elimination, increase data quality. The Synthetic Minority Oversampling Technique (SMOTE) is used to alleviate class imbalance and to enable strong model training. Machine learning models, include SVM, kNN, Neural Networks, and Naive Bayes, were built and verified using Key performance metrics and K-Fold cv. included as recall, accuracy, F1-score, precision and AUC- ROC were employed to analyze the models. Among these, the Neural Network model emerged the most effective, obtaining a prediction accuracy 98.13%, precision 98.21%, recall 98.00%, F1Score of 97.96%, AUC-ROC score 0.9992. Study underscores promise of ML boosting the diagnosis and treatment of WBCD illnesses, giving scalable and accurate ways for early detection and prevention.

Author 1: Abdullah Al Mamun
Author 2: Touhid Bhuiyan
Author 3: Md Maruf Hassan
Author 4: Shahedul Islam Anik

Keywords: Wisconsin breast cancer disease prediction; ML; SVM; KNN; AUC-ROC; Naive Bayes

PDF

Paper 130: A Machine Learning-Based Analysis of Tourism Recommendation Systems: Holistic Parameter Discovery and Insights

Abstract: Tourism is a cornerstone of the global economy, fostering cultural exchange and economic growth. As travelers increasingly seek personalized experiences, recommendation systems have become vital in guiding decision-making and enhancing satisfaction. These systems leverage advanced technologies such as IoT and machine learning to provide tailored suggestions for destinations, accommodations, and activities. This paper explores the transformative role of tourism recommendation systems (TRS) by analyzing data from 3,013 research articles published between 2000 and 2024 using a BERT-based methodology for semantic text representation and clustering. A robust software framework, integrating tools such as UMAP for dimensionality reduction and HDBSCAN for clustering, facilitated data modeling, cluster analysis, visualization, and the identification of key parameters in TRS. We discover a comprehensive taxonomy of 16 TRS parameters grouped into 4 macro-parameters. These include Personalized Tourism; Sustainability, Health and Resource Awareness; Adaptability & Crisis Management; and Social Impact & Cultural Heritage. These macro-parameters align with all three dimensions of the triple bottom line (TBL) -- social, economic, and environmental sustainability. The findings reveal key trends, highlight underexplored areas, and provide research-informed recommendations for developing more effective TRS. This paper synthesizes existing knowledge, identifies research gaps, and outlines directions for advancing TRS to support sustainable, personalized, and innovative travel solutions.

Author 1: Raniah Alsahafi
Author 2: Rashid Mehmood
Author 3: Saad Alqahtany

Keywords: Recommendation Systems (RS); Tourism Recommendation Systems (TRS); big data analytics; machine learning; unsupervised learning; social; economic and environmental sustainability; Bidirectional Encoder Representations from Transformers (BERT); SDGs; literature review

PDF

The Science and Information (SAI) Organization
BACK TO TOP

Computer Science Journal

  • About the Journal
  • Call for Papers
  • Submit Paper
  • Indexing

Our Conferences

  • Computing Conference
  • Intelligent Systems Conference
  • Future Technologies Conference
  • Communication Conference

Help & Support

  • Contact Us
  • About Us
  • Terms and Conditions
  • Privacy Policy

© The Science and Information (SAI) Organization Limited. All rights reserved. Registered in England and Wales. Company Number 8933205. thesai.org