The Science and Information (SAI) Organization
  • Home
  • About Us
  • Journals
  • Conferences
  • Contact Us

Publication Links

  • IJACSA
  • Author Guidelines
  • Publication Policies
  • Digital Archiving Policy
  • Promote your Publication
  • Metadata Harvesting (OAI2)

IJACSA

  • About the Journal
  • Call for Papers
  • Editorial Board
  • Author Guidelines
  • Submit your Paper
  • Current Issue
  • Archives
  • Indexing
  • Fees/ APC
  • Reviewers
  • Apply as a Reviewer

IJARAI

  • About the Journal
  • Archives
  • Indexing & Archiving

Special Issues

  • Home
  • Archives
  • Proposals
  • Guest Editors
  • SUSAI-EE 2025
  • ICONS-BA 2025
  • IoT-BLOCK 2025

Future of Information and Communication Conference (FICC)

  • Home
  • Call for Papers
  • Submit your Paper/Poster
  • Register
  • Venue
  • Contact

Computing Conference

  • Home
  • Call for Papers
  • Submit your Paper/Poster
  • Register
  • Venue
  • Contact

Intelligent Systems Conference (IntelliSys)

  • Home
  • Call for Papers
  • Submit your Paper/Poster
  • Register
  • Venue
  • Contact

Future Technologies Conference (FTC)

  • Home
  • Call for Papers
  • Submit your Paper/Poster
  • Register
  • Venue
  • Contact
  • Home
  • Call for Papers
  • Editorial Board
  • Guidelines
  • Submit
  • Current Issue
  • Archives
  • Indexing
  • Fees
  • Reviewers
  • Subscribe

IJACSA Volume 12 Issue 11

Copyright Statement: This is an open access publication licensed under a Creative Commons Attribution 4.0 International License, which permits unrestricted use, distribution, and reproduction in any medium, even commercially as long as the original work is properly cited.

View Full Issue

Paper 1: Performance Evaluation of SNMPv1/2c/3 using Different Security Models on Raspberry Pi

Abstract: The Simple Network Management Protocol (SNMP) is one of the dominant protocols for network monitoring and configuration. The first two versions of SNMP (v1 and v2c) use the Community-based Security Model (CSM), where the community is transferred in clear text, resulting in a low level of security. With the release of SNMPv3, the User-based Security Model (USM) and Transport Security Model (TSM) were proposed, with strong authentication and privacy at different levels. The Raspberry Pi family of Single-Board Computers (SBCs) is widely used for many applications. To help their integration into network management systems, it is essential to study the impact of the different versions and security models of SNMP on these SBCs. In this work, we carried out a performance analysis of SNMP agents running in three different Raspberry Pis (Pi Zero W, Pi 3 Model B, and Pi 3 Model B+). Our comparisons are based on the response time, defined as the time required to complete a request/response exchange between a manager and an agent. Since we did not find an adequate tool for our assessments, we developed our own benchmarking tool. We did numerous experiments, varying different parameters such as the type of requests, the number of objects involved per request, the security levels of SNMPv3/USM, the authentication and privacy protocols of SNMPv3/USM, the transport protocols, and the versions and security models of SNMP. Our experiments were executed with Net-SNMP, an open-source and comprehensive distribution of SNMP. Our tests indicate that SNMPv1 and SNMPv2c have similar performance. SNMPv3 has a longer response time, due to the overhead caused by the security services (authentication and privacy). The Pi 3 Model B and Pi 3 Model B+ have comparable performance, and significantly outperform the Pi Zero W.

Author 1: Eric Gamess
Author 2: Sergio Hernandez

Keywords: Simple network management protocol; SNMP; performance evaluation; benchmarks; raspberry pi

PDF

Paper 2: Digital Economy and its Importance in the Development of Small and Medium Innovative Enterprises

Abstract: A single universally accepted definition, levels and interconnections of digital economy with other economies is not yet developed. Thus, various definitions of the digital economy have been investigated, as well as various approaches to describing the process of transformation of the digital economy for the correct establishment of these relationships. The article observes the relationship between the state of the digital economy, innovative small and medium enterprises, the development of small and medium businesses in general. The stage of transformation of the digital economy of Russia is determined at the second, intermediate stage of development and the main barriers to moving to the third level are pointed out. The dual role of the digital economy in the development of small and medium innovative enterprises is determined based on the selected model of R. Bukht & R. Heeks, the two directions of influence being the SMEs provision with necessary tools and the digital economy becoming the object of innovative development of SMEs. Finally, the assessment of the state of digital economy in Russia is given and the recommendations for its further implementation are given.

Author 1: Tatiana Korsakova
Author 2: Lyudmila Dubanevich
Author 3: Oleg Drozdov
Author 4: Anna Mikhailova
Author 5: Ekaterina Kamchatova

Keywords: Business; SMEs; entrepreneurship; Russia; digitalization

PDF

Paper 3: Human Action Recognition in Video Sequence using Logistic Regression by Features Fusion Approach based on CNN Features

Abstract: Human Action recognition (HAR) gains too much attention due to its wide range of real world applications, such as video surveillance, robotics and computer vision. In video surveillance systems security cameras are placed to monitor activities and motion, generate alerts in undesirable situations. Due to such importance of video surveillance in daily life, HAR becomes the primary and key factor of video surveillance systems. Many researchers worked on human action recognition but HAR still a challenging problem, due to large variation among human to human and human actions in daily life, which make human recognition very challenging and makes surveillance system difficult to outperform. In this article a novel method is proposed by features fusion of pre-trained convolution neural network (CNN) features. Initially pre-trained CNN VGG 19 weights are exploited to extract fully connected 7th layer (FC7) of the selected dataset, subsequently pre-trained fully connected 8th layer features (FC8) extracted by employing pre-trained weights of the same neural network. However the resultant feature fused vector further optimized by employing two statistical features selection techniques, chi-square test and mutual information to select best features among them to reduced redundancy and increase performance accuracy of human action, a threshold value used for selecting best features. Furthermore the best features are fused, then grid search with 10 fold cross validation is applied for tuning hyper parameter to select best k fold and the resulting best parameter are feed to Logistic regression (LR) classifier for recognition. The proposed technique used You Tube 11 action dataset and achieved 98.49% accuracy. Lastly the proposed method compares with the existing state of the art methods which show dominance performance.

Author 1: Tariq Ahmad
Author 2: Jinsong Wu
Author 3: Imran Khan
Author 4: Asif Rahim
Author 5: Amjad Khan

Keywords: Human action recognition; logistic regression; deep learning; convolution neural network; features fusion

PDF

Paper 4: Application-based Framework for Analysis, Monitoring and Evaluation of National Open Data Portals

Abstract: Open Government Data (OGD) portals are considered of significant national importance towards transparency and accountability improvement. The continuous publication of data in OGD portals introduces the need for high-quality data and the qualitative portal itself. This paper aims to address the data quality issues through a framework composed of several components aimed at measuring and monitoring the OGD portals in an automated way. Through this proposed framework, is intended to monitor and evaluate OGD quality, respectively OGD portals, and to show their progress/regress based on accumulated scores for different periods. The advantage of the proposed framework is the compatibility with any OGD Portal due to its flexibility of integration. The integration interface consists of only a few basic metrics but is necessary that almost the OGD portal possesses and can produce very compressive results. The other advantage is the possibility of extraction of collected data for further analysis and the introduction of artificial intelligence (AI) for prediction purposes to point out how the OGD portals will stand in the next period.

Author 1: Vigan Raca
Author 2: Goran Velinov
Author 3: Betim Cico
Author 4: Margita Kon-Popovska

Keywords: Open data; government; datasets; evaluation; portals; framework

PDF

Paper 5: Deep Learning for Arabic Image Captioning: A Comparative Study of Main Factors and Preprocessing Recommendations

Abstract: Captioning of images has been a major concern for the last decade, with most of the efforts aimed at English captioning. Due to the lack of work done for Arabic, relying on translation as an alternative to creating Arabic captions will lead to accumulating errors during translation and caption prediction. When working with Arabic datasets, preprocessing is crucial, and handling Arabic morphological features such as Nunation requires additional steps. We tested 32 different variables combinations that affect caption generation, including preprocessing, deep learning techniques (LSTM and GRU), dropout, and features extraction (Inception V3, VGG16). Moreover, our results on the only publicly avail-able Arabic Dataset outperform the best result with BLEU-1=36.5, BLEU-2=21.4, BLEU-3=12 and BLEU4=6.6. As a result of this study, we demonstrated that using Arabic preprocessing and VGG16 image features extraction enhanced Arabic caption quality, but we saw no measurable difference when using Dropout or LSTM instead of GRU.

Author 1: Hani Hejazi
Author 2: Khaled Shaalan

Keywords: Deep learning; NLP; Arabic image captioning; Arabic text preprocessing; LSTM; VGG16; INCEPTION V3

PDF

Paper 6: Hardware Architecture for Adaptive Dual Threshold Filter and Discrete Wavelet Transform based ECG Signal Denoising

Abstract: The ECG signal, like all signals obtained when instrumenting a data acquisition system, is affected by noises of physiological and technical sources such as Electromyogram (EMG) and power line interferences, which can deteriorate its morphology. To overcome this issue, it’s subjected to apply a preprocessing step to remove these noises. Filtring techniques are complex computations becoming more common in medical applications, which must be completed in real-time. As a result, these applications are geared at integrating high-performance embedded architectures. This paper presents an FPGA (Field Programmable Gate Array) embedded architecture designed for an ECG denoising hybrid technique based on the Discrete Wavelet transform (DWT) and the Adaptive Dual Threshold Filter (ADTF), dedicated to handle with noises affecting ECG signals. The architecture was designed following a hardware-software codesign using a high-level description language and synthetized to be implemented on different FPGAs due to the structural description flexibility. The global architecture was divided into a set of functional blocks to allow parallel processing of ECG data. The simulation results confirm the high performance of the system in noise reduction without affecting the morphology of the signal. The process takes 0.3 ms with an acquisition frequency of 360 Hz. The whole architecture requires a small area in different FPGAs in terms of resources utilization. It uses less than 1% of the total registers for all FPGA devices which represents a total of 292 registers for Cyclone III LS, Cyclone IV GX, Cyclone IV E, and Arria II GX; and a total of 329 registers for Cyclone V. The logic elements occupancy varies between 3% using Cyclone V and 60% using Cyclone IV GX freeing up space for other parallel processing tasks.

Author 1: Safa MEJHOUDI
Author 2: Rachid LATIF Wissam JENKAL
Author 3: Amine Saddik
Author 4: Abdelhafid EL OUARDI SATIE

Keywords: ECG signal; DWT; ADTF; hybrid technique; hardware-software codesign; FPGA

PDF

Paper 7: Machine Learning based Forecasting Systems for Worldwide International Tourists Arrival

Abstract: The international tourist movement has overgrown in recent decades, and travelers are considered a significant source of income to the tourism economy. When tourists visit a place, they spend considerable money on their enjoyment, travel, and hotel accommodations. In this research, tourist data from 2010 to 2020 have been extracted and extended with depth analysis of different dimensions to identify valuable features. This research attempts to use machine learning regression techniques such as Support Vector Regression (SVR) and Random Forest Regression (RFR) to forecast and predict worldwide international tourist arrivals and achieved forecasting accuracy using SVR is 99.4% and using RFR is 84.7%. The study also analyzed the forecasting deadlock condition after covid-19 in the sudden drop of international visitors due to lockdown enforcement by all countries.

Author 1: Ram Krishn Mishra
Author 2: Siddhaling Urolagin
Author 3: J. Angel Arul Jothi
Author 4: Nishad Nawaz
Author 5: Haywantee Ramkissoon

Keywords: Tourists; forecasting; machine learning; Covid-19

PDF

Paper 8: Analyzing the Sentiments of Jordanian Students Towards Online Education in the Higher Education Institutions

Abstract: Sentiment analysis and opinion polling are two areas that have grown significantly over the past decade. Opinion research and sentiments analysis in the online education environment can truly reflect the learning state of students and educators and experts in the field; providing the theoretical basis needed to further review educational procedure and conduct. This study aims to shed light on identifying and visualizing students' objective feelings based on an exploration of the subject matter and materials of learning and gathering sentiments from university Facebook groups at various levels and layers in detail. The proposed method is a qualitative descriptive research method that includes data pre-processing, subject discovery, sentiment analysis, and visualization. In relative terms, 39.7% of text messages were positive and 52.3% of text messages were negative and understanding the narrative of these feelings and their impact on the online learning environment.

Author 1: Bayan Alfayoumi
Author 2: Mohammad Alshraideh
Author 3: Saleh Al-Sharaeh
Author 4: Martin Leiner
Author 5: Iyad Muhsen AlDajani

Keywords: Online education; students; sentiment analysis; online education; online environment; online social media

PDF

Paper 9: Comparative Study of Flooding Area Detection with SAR Images based on Thresholding and Difference Images Acquired Before and After the Flooding

Abstract: Comparative study of flooding area detection with Synthetic Aperture Radar (SAR) images based on thresholding and difference images acquired before and after the flooding is conducted. Method for flooding, landslide and sediment disaster area detections with SAR is proposed. The following two different methods for flooding detection are common. It is not so easy to determine a threshold for the thresholding method while subtraction method between before and after images of a disaster occurrence has the disadvantage that false disaster areas are detected due to a variation of ground cover targets. Therefore, a comparative study between both methods is required. Its application is demonstrated for the disaster which is occurred in Saga Prefecture, Japan due to a long term of heavy rain during from the begging of August to the middle of August in 2021. Through experiments with Sentinel-1 SAR imagery data, it is found that the proposed method works well for the detection of the disaster.

Author 1: Kohei Arai

Keywords: Flooding; landslide; sediment disaster; heavy rain; image quality; Synthetic Aperture Radar; SAR; sentinel-1 SAR; thresholding; difference images between before and after disaster

PDF

Paper 10: Developing of Middleware and Cross Platform Chat Application

Abstract: The rapid development of technology has resulted in many new innovations on social media platforms. Now-a-days, there are many chat applications available, namely Whatsapp, Telegram, LINE, Viber, and many others. This in turn forces users to juggle between many chat applications as different applications can’t communicate with each other. This research aims to develop a chat application which serves as a middleware to make communication between developed chat application and two conventional chat applications possible (Telegram and LINE). Several tests are done to ensure that the message exchange process (in text, picture, video, and file type) works well between the developed chat application as well as Telegram or LINE.

Author 1: Danny Sebastian
Author 2: Restyandito
Author 3: Kristian Adi Nugraha

Keywords: Telegram API; line API; chat application; flutter; middleware

PDF

Paper 11: A Survey on Deep Learning Face Age Estimation Model: Method and Ethnicity

Abstract: Face age estimation is a type of study in computer vision and pattern recognition. Designing an age estimation or classification model requires data as training samples for the machine to learn. Deep learning method has improved estimation accuracy and the number of deep learning age estimation models developed. Furthermore, numerous datasets availability is making the method an increasingly attractive approach. However, face age databases mostly have limited ethnic subjects, only one or two ethnicities and may result in ethnic bias during age estimation, thus impeding progress in understanding face age estimation. This paper reviewed available face age databases, deep learning age estimation models, and discussed issues related to ethnicity when estimating age. The review revealed changes in deep learning architectural designs from 2015 to 2020, frequently used face databases, and the number of different ethnicities considered. Although model performance has improved, the widespread use of specific few multi-races databases, such as the MORPH and FG-NET databases, suggests that most age estimation studies are biased against non-Caucasians/non-white subjects. Two primary reasons for face age research’s failure to further discover and understand ethnic traits effects on a person’s facial aging process: lack of multi-race databases and ethnic traits exclusion. Additionally, this study presented a framework for accounting ethnic in face age estimation research and several suggestions on collecting and expanding multi-race databases. The given framework and suggestions are also applicable for other secondary factors (e.g. gender) that affect face age progression and may help further improve future face age estimation research.

Author 1: Hadi A. Dahlan

Keywords: Deep learning; face age estimation; face database; ethnicity bias

PDF

Paper 12: Comparative Analysis of Supervised Machine Learning Techniques for Sales Forecasting

Abstract: This study talks about how data mining can be used for sales forecasting in retail sales and demand prediction. Prediction of sales is a crucial task which determines the success of any organization in the long run. There are various techniques available for predicting the sales of a supermarket such as Time Series Algorithm, Regression Techniques, Association rule etc. In this paper, a comparative analysis of some of the Supervised Machine Learning Techniques have been done such as Multiple Linear Regression Algorithm, Random Forest Regression Algorithm, K-NN Algorithm, Support Vector Machine (SVM) Algorithm and Extra Tree Regression to build a prediction model and precisely estimate possible sales of 45 retail outlets of Walmart store which are at different geographical locations. Walmart is one of the foremost stores across the world and thus authors would like to predict the sales accurately. Certain events and holidays affect the sales periodically, which sometimes can also be on a daily basis. The forecast of probable sales is based on a combination of features such as previous sales data, promotional events, holiday week, temperature, fuel price, CPI i.e., Consumer Price Index and Unemployment rate in the state. The data is collected from 45 outlets of Walmart and the prediction about the sales of Walmart was done using various Supervised Machine Learning Techniques. The contribution of this paper is to help the business owners decide which approach to follow while trying to predict the sales of their Supermarket taken into account different scenarios including temperature, holidays, fuel price, etc. This will help them in deciding the promotional and marketing strategy for their products.

Author 1: Stuti Raizada
Author 2: Jatinderkumar R. Saini

Keywords: Sales forecasting; linear regression; random forest regression; KNN regression algorithm; SVM algorithm; supervised machine learning techniques

PDF

Paper 13: Machine Learning for Diagnosing Drug Users and Types of Drugs Used

Abstract: Drug use is very detrimental to the physical and psychological health of users. Drug abuse also causes addiction and is a global epidemic. Therefore it is not surprising that scientific research related to drugs has attracted attention for research. However, many factors become obstacles in the medical services of the drug user, including cost, flexibility, and a slow process. Meanwhile, electronic systems can speed up handling time, improve work efficiency, save costs and reduce inspection errors. It means that a breakthrough is needed in developing a platform that can identify drug users. Therefore, this research aims to build machine learning with expertise like an expert who can diagnose drug users and distinguish the types of drugs used by drug users. The expert system on machine learning was developed using the Forward Chaining and Certainty Factor methods. This study concludes that the expert system on machine learning developed can be used to diagnose drug users and distinguish the types of drugs used with an accuracy of up to 80%. The implications of the expert system on machine learning are an alternative method for narcotics officers and medical doctors in diagnosing drug users and the types of drugs used.

Author 1: Anthony Anggrawan
Author 2: Christofer Satria
Author 3: Che Ku Nuraini
Author 4: Lusiana
Author 5: Ni Gusti Ayu Dasriani
Author 6: Mayadi

Keywords: Machine learning; drug; expert system; forward chaining; certainty factor

PDF

Paper 14: Insights on Deep Learning based Segmentation Schemes Towards Analyzing Satellite Imageries

Abstract: Satellite imageries are essentially a complex form of an image when subjected to critical analytical operation. The analytical process applied on remotely sensed satellite imageries are utilized for generating the land cover map. With an abundance of traditional techniques evolved to date, deep learning-based schemes are progressively gaining pace for identifying and classifying a terrestrial object in satellite images. However, different variants of deep learning approaches have different operations, and so are the consequences. At the same time, there is no reported literature to highlight the issues, trends, and effectiveness much on a generalized scale concerning segmentation. Therefore, this paper reviews some of the recent segmentation approaches using deep learning to contribute towards review findings in the form of research trends, research gaps, and essential learning outcomes. The paper offers a compact and distinct picture of deep learning approaches used to boost segmentation for satellite images.

Author 1: Natya S
Author 2: Ramya K
Author 3: Seema Singh

Keywords: Deep learning; landcover; map generation; remotely sense image; satellite image; segmentation

PDF

Paper 15: DNA Profiling: An Investigation of Six Machine Learning Algorithms for Estimating the Number of Contributors in DNA Mixtures

Abstract: DNA (Deoxyribonucleic acid) profiling involves analysis of sequences of individual or mixed DNA profiles to identify persons these profiles belong to. DNA profiling is used in important applications such as for paternity tests, in forensic science for person identification on a crime scheme, etc. Finding the number of contributors in a DNA mixture is a major task in DNA profiling with challenges caused due to allele dropout, stutter, blobs, and noise. The existing methods for finding the number of unknowns in a DNA mixture suffer from issues including computational complexity and accuracy of estimating the number of unknowns. Machine learning has received attention recently in this area but with limited success. Many more efforts are needed for improving the robustness and accuracy of these methods. Our research aims to advance the state-of-the-art in this area. Specifically, in this paper, we investigate the performance of six machine learning algorithms -- Nearest Neighbors (KNN), Random Forest (RF), Support Vector Machine (SVM), Logistic Regression (LR), Stochastic Gradient Descent (SGD), and Gaussian Naïve-Bayes (GNB) -- applied to a publicly available dataset called PROVEDIt, containing mixtures with up to five contributors. We evaluate the algorithmic performance using confusion matrices and four performance metrics namely accuracy, F1-Score, Recall, and Precision. The results show that LR provides the highest Accuracy of 95% for mixtures with five contributors.

Author 1: Hamdah Alotaibi
Author 2: Fawaz Alsolami
Author 3: Rashid Mehmood

Keywords: Machine learning; DNA profiling; DNA mixtures; forensic science

PDF

Paper 16: Visualization of the Temporal Topic Model on Higher Education Preferences with Higher Education Ranking Indicators

Abstract: Private universities have devised a strategy to counteract the ongoing competition. Private universities can use the appropriate data analysis method to make higher education management decisions. The goal of this research is to find a new approach to data analysis methods in the form of visualization using the TTM (Temporal Topic Model) method to assist private university management. These findings are the two formulas used to generate time-based visualizations and the Temporal Topic Model per month to visually change news topics related to rankings so that management can decide on marketing strategies and policies that are in relation to public opinion.

Author 1: Winda Widya Ariestya
Author 2: Achmad Benny Mutiara
Author 3: I Made Wiryana
Author 4: Setia Wirawan

Keywords: Management decisions; temporal topic model; university; visualization

PDF

Paper 17: Sparse Distributed Memory Approach for Reinforcement Learning Driven Efficient Routing in Mobile Wireless Network System

Abstract: In recent years, researchers have explored the applicability of Q-learning, a model-free reinforcement learning technology towards designing QoS-aware, resource-efficiency, and reliable routing techniquesin a dynamically changing network environment. However, Q-learning is based on tabular representation to characterize learned policies that frequently encounter a dimension disaster problem when introduced to the uncertain and dynamically changing network environment. In addition, the time required for agent learning in the training phase is too long, which makes it difficult for the agent to generalize the observation state efficiently. To this end, this paper attempts to overcome the overhead memory problems encountered in Q-learning-based routing techniques. In this paper, the study presents a novel memory-efficient intelligent routing mechanism based on adaptive Kanerva coding, which minimizes the storage cost required for storing large action and a state value. Unlike existing schemes, the proposed method optimizes memory requirements. Also, it enables better generalization by storing the learnable parameters of the function approximator present in the agent in a Kanerva-coding data structure. The Kanerva-coding is a sparse memory with distributed reading and writing mechanism which enables optimal compression and state abstractions for learning with fewer parameterized components making it highly memory efficient. The design and implementation of the proposed technique are done on the Anaconda tool. Simulation results demonstrate that the proposed technique can adaptively adjust the routing policy according to the varying network environment to meet the transmission requirements of different services with low memory requirements.

Author 1: Varshini Vidyadhar
Author 2: Nagaraj R
Author 3: G Sudha

Keywords: Mobile wireless network; reinforcement learning; Q-learning; Kanerva coding; routing; memory optimization

PDF

Paper 18: A Systematic Review of Published Articles, Phases and Activities in an Online Social Networks Forensic Investigation Domain

Abstract: The purpose of this paper is to retrieve, evaluate and analyse the available published articles in five (5) relevant online databases from 2011 to 2021 and also critically identify the phases and activities involved in an Online Social Networks Forensic Investigation based on bibliometric analysis and Degree of confidence respectively in order to know the evolution in the research domain. A systematic literature review (SLR) technique was adopted by the author to search using pre-defined keywords. Only scholarly articles published between 2011 and 2021 written in English were included in the search. The total of 316 subscribed documents were collected from the five (5) online databases based on the search criteria although twenty-nine (29) are duplicates. ScienceDirect has the highest number with 189 documents and the year 2020 with the highest published articles. Six (6) phases and forty-three (43) activities were identified. According to a review of the recovered publications, no previous research has been done to statistically retrieve, evaluate and analyse the level of work that has been done in the domain of OSNFI, as well as the phases and activities involved in the forensic investigation of an online social networks crime.

Author 1: Aliyu Musa Bade
Author 2: Siti Hajar Othman

Keywords: Forensic; investigation; model; online social networks; bibliometric analysis; degree of confidence

PDF

Paper 19: Deep Learning based Neck Models for Object Detection: A Review and a Benchmarking Study

Abstract: Artificial intelligence is the science of enabling computers to act without being further programmed. Particularly, computer vision is one of its innovative fields that manages how computers acquire comprehension from videos and images. In the previous decades, computer vision has been involved in many fields such as self-driving cars, efficient information retrieval, effective surveillance, and a better understanding of human behaviour. Based on deep neural networks, object detection is actively growing for pushing the limits of detection accuracy and speed. Object Detection aims to locate each object instance and assign a class to it in an image or a video sequence. Object detectors are usually provided with a backbone network designed for feature extractors, a neck model for feature aggregation, and finally a head for prediction. Neck models, which are the purpose of study in this paper, are neural networks used to make a fusion between high-level features and low-level features and are known by their efficiency in object detection. The aim of this study to present a review of neck models together before making a benchmarking that would help researchers and scientists use it as a guideline for their works.

Author 1: Sara Bouraya
Author 2: Abdessamad Belangour

Keywords: Object detection; deep learning; computer vision; neck models; feature aggregation; feature fusion

PDF

Paper 20: Naïve Bayes Classification of High-Resolution Aerial Imagery

Abstract: In this study, the performance of Naïve Bayes classification on a high-resolution aerial image captured from a UAV-based remote sensing platform is investigated. K-means clustering of the study area is initially performed to assist in selecting the training pixels for the Naïve Bayes classification. The Naïve Bayes classification is performed using linear and quadratic discriminant analyses and by making use of training set sizes that are varied from 10 through 100 pixels. The results show that the 20 training set size gives the highest overall classification accuracy and Kappa coefficient for both discriminant analysis types. The linear discriminant analysis with 94.44% overall classification accuracy and 0.9395 Kappa coefficient is found higher than the quadratic discriminant analysis with 88.89% overall classification accuracy and 0.875 Kappa coefficient. Further investigations carried out on the producer accuracy and area size of individual classes show that the linear discriminant analysis produces a more realistic classification compared to the quadratic discriminant analysis particularly due to limited homogenous training pixels of certain objects.

Author 1: Asmala Ahmad
Author 2: Hamzah Sakidin
Author 3: Mohd Yazid Abu Sari
Author 4: Abd Rahman Mat Amin
Author 5: Suliadi Firdaus Sufahani
Author 6: Abd Wahid Rasib

Keywords: Naïve Bayes; k-means; classification accuracy; training set size; discriminant analysis

PDF

Paper 21: Secured and Provisioned Access Authentication using Subscribed user Identity in Federated Clouds

Abstract: Cloud computing has become an essential source for modern trade or market environments by abled frameworks. The exponential growth of cloud computing services in the last few years has resulted in extensive use, especially in storing and sharing the data on various cloud servers. The current trend in the cloud shows that the cloud owners use relative functions and target areas in such a way that cloud customers access or store their data either in the same servers or related servers. Simultaneously, from the security point of view, the lack of confidence about the customer's data on the cloud server is still questionable. The hour's need is to provide the cloud service in a 'single port way' by forming the joint management policy to increase customer satisfaction and profitability. In addition to this, the authentication steps also need to be improvised. This paper discusses issues on the security authentication and access provisioning of the cloud service consumers in federated clouds using subscribed user identity. This work proposes the user identity verification module (UidVM) in the cloud service consumer's authentication process to serve as a cloud broker to minimize the work overloads on the central cloud federation management system, thus enhancing the cloud security.

Author 1: Sudan Jha
Author 2: Sultan Ahmad
Author 3: Meshal Alharbi
Author 4: Bader Alouffi
Author 5: Shoney Sebastian

Keywords: Security authentication (SA); cloud federation (CF); cloud service provider (CSP); key distribution center (KDC); user identity verification module (UIdVM)

PDF

Paper 22: Local Frequency Descriptor and Hybrid Features for Classification of Brain Magnetic Resonance Images using Ensemble Classifier

Abstract: A brain tumor is an irregular development of cells in the human brain that causes problems with the brain's normal functionalities. Early detection of brain tumor is an essential process to help the patient to live longer than treatment. Hence in this paper, a hybrid ensemble model has been proposed to classify the input brain MRI images into two classes: brain MRI images having tumor and brain MRI images with no tumor. The hybrid features are extracted by analyzing the texture and statistical properties of brain MRI images. Further, the Local Frequency Descriptor (LFD) technique is employed to extract the prominent features from the brain tumor region. Finally, an ensemble classifier has been developed with the combination of Support Vector Machine (SVM), Decision Tree (DT) and K-Nearest Neighbour (KNN) technique to successfully classify the brain MRI images into brain tumor MRI images and non-tumor brain MRI images. The proposed model is tested on the Kaggle brain tumor dataset and the performance of the method is evaluated in terms of accuracy, sensitivity, specificity, precision, recall and f-measure (f1 score-harmonic mean of precision and recall). The results show that the proposed model is promising and encouraging.

Author 1: Shruthi G
Author 2: Krishna Raj P M

Keywords: Brain tumor; hybrid features; local frequency descriptor (LFD); ensemble classifier

PDF

Paper 23: Mutual Informative Brown Clustering based Multiattribute Cockroach Swarm Optimization for Reliable Data Dissemination in VANET

Abstract: A vehicular ad hoc network (VANETs) intends to obtain communication for vehicular networks and enhances road safety and effectiveness with help of wireless technology. Data dissemination is an important process in communication. In VANETs system, Data dissemination plays a significant role. A novel Mutual informative brown clustering-based multi attribute cockroach swarm optimization (MIBCMCSO) technique is introduced for improving data dissemination. In reliable data dissemination, clustering and optimization are the two major process of proposed MIBCMCSO technique. Initially, clustering procedure is performed for separating entire network towards different groups of vehicle nodes namely distance, direction, density and velocity of node. For each group, cluster head was chosen among the members to efficient data with minimum delay. Secondly, multi attribute cockroach swarm optimization technique is applied for finding optimal cluster head through multi attribute functions such as residual energy, bandwidth availability, and distance. Then source node performs data dissemination destination via optimal cluster head. Simulation of MIBCMCSO as well as existing technique is performed by various performance parameters like packet delivery ratio, end to end delay and throughput. MIBCMCSO achieves higher consistency of data dissemination as well as lesser delay than conventional methods.

Author 1: D. Radhika
Author 2: A. Bhuvaneswari

Keywords: VANET; data dissemination; mutual informated brown agglomerative clustering; multi attribute cockroach swarm optimization

PDF

Paper 24: Face Age Estimation and the Other-race Effect

Abstract: Age estimation is an automated method of predicting human age from 2-D facial feature representations. The majority of studies carried out in this research area use the FG-NET and MORPH 2 databases to train and test developed systems, which are lacking in black-face content. Most age frauds are perpetuated in the sub-Saharan African region due to the unavailability of an official database and unregistered births in the rural areas. The issues of unverified age in the region made it possible for under-age voters, under-age drivers, and the engagement of over-aged sportsmen. The other-race effect could reduce the performance of face recognition techniques, which could make techniques that work for white faces underperform when deployed for use in the predominately black face region. This study examines the other-race effect on face-based age estimation by analyzing the accuracy of an age estimation system trained with predominantly black faces against the same age estimation system trained with predominately white faces. The developed age estimation system uses a genetic algorithm-artificial neural network classifier and local binary pattern for texture and shape feature extraction. A total of 170 black faces were used for system testing. The result showed that the age estimation system trained with the predominantly black face database (GA-ANN-AES-855) outperformed the system trained with predominantly white faces (GA-ANN-AES-255) on testing with the aforementioned black face samples. The results obtained from the simulation were further subjected to inferential statistics, which established that the improvement in the correct classification rate was statistically significant. Hence, the other-race effect affects face-based age estimation systems.

Author 1: Oluwasegun Oladipo
Author 2: Elijah Olusayo Omidiora
Author 3: Victor Chukwudi Osamor

Keywords: Component; face recognition; age estimation; other-race effect

PDF

Paper 25: Polarity Detection of Dialectal Arabic using Deep Learning Models

Abstract: With the evolution of a new era of technology and social media networks, as well as an increase in Arabs sharing their point of view, it became necessary that this research be conducted. Sentiment analysis is concerned with identifying and extracting opinionated phrases from reviews or tweets. Specifically, to determine whether a given tweet is positive, negative, or neutral. Dialectical Arabic poses difficulties for sentiment analysis. In this paper, four deep learning models are presented, to be specific convolution neural networks (CNN), long short-term memory (LSTM), a hybrid of (CNN-LSTM), and Bidirectional LSTMs (BiLSTM), to determine the tweets polarities written in dialectal Arabic. The performance of the four models is validated on the used corpus with the use of word embedding and applying the (k-Fold Cross-Validation) method. The results show that CNN outperforms the others achieving an accuracy of 99.65%.

Author 1: Saleh M. Mohamed
Author 2: Ensaf Hussein Mohamed
Author 3: Mohamed A. Belal

Keywords: Sentiment analysis; word embedding; sentiment classification; dialectical arabic; deep learning

PDF

Paper 26: Challenges in Developing Virtual Reality, Augmented Reality and Mixed-Reality Applications: Case Studies on A 3D-Based Tangible Cultural Heritage Conservation

Abstract: A model that contributes in a simple, practical and effective way to develop 3D-based CH conservation applications involving the use of VR, AR and MR technologies was proposed based on the identification of challenges in developing applications. Identification was carried out by analyzing related and relevant articles selected randomly using Google and Google Scholar search engines. The model can prevent researchers from lack of planning in carrying out research in this field, and it is suitable for those just starting out with this type of research. In addition, this model can support researchers to more easily, practically and effectively implement 3D-based cultural heritage conservation by using virtual reality or augmented reality or mixed reality technology.

Author 1: Ahmad Zainul Fanani
Author 2: Khafiizh Hastuti
Author 3: Arry Maulana Syarif
Author 4: Prayanto Widyo Harsanto

Keywords: Virtual reality; augmented reality; mixed reality; tangible cultural heritage; 3D-based cultural heritage conservation

PDF

Paper 27: Trust-based Key Management Conglomerate ElGamal Encryption for Data Aggregation Framework in WSN using Blockchain Technology

Abstract: In wireless sensor networks (WSN), data aggregation is a widely used method. Security issues like data integrity and data confidentiality became a significant concern in data aggregation when the sensor network is deployed in a hostile environment. Many researches may carry out several works to tolerate these security issues. However, there were some limitations like delay, the arrival rate of packets, and so on. Hence, to overcome the existing problems, this approach offers a blockchain-dependent data aggregation scheme in WSN. The main intention of the proposed work is to generate a certificateless key generation so that the proposed system's secrecy rate is improved. The use of blockchain is employed for security purposes, and it enables the user to acquire the information stored internally in an effortless manner. Initially, deployment of sensor and base station (BS) is carried out, followed by node registration at which the public/private keys are generated. The computation of private hash values is carried by performing certificateless key generation. After that, the formation of blockchain is made using the PoW (Proof of Work) detection algorithm followed by the aggregation of data. In the data aggregation process, Elgammal based cryptographic approach is introduced to acquire member data, perform aggregation logic, and transfer the aggregated data. Finally, cluster-based routing is established with the use of Knapsack based cluster routing strategy. The performance investigation of the proposed system is estimated and the outcomes attained are compared with the existing techniques in terms of arrival rate, average delay, and the delay ratio of the packets. The investigation illustrates that the suggested approach is better than the traditional techniques.

Author 1: T. G. Babu
Author 2: V. Jayalakshmi

Keywords: Wireless sensor networks; data aggregation; PoW detection scheme; blockchain technology; cluster formation; key generation; security; delay ratio

PDF

Paper 28: Statistical Analysis for Revealing Defects in Software Projects: Systematic Literature Review

Abstract: Defect detection in software is the procedure to identify parts of software that may comprise defects. Software companies always seek to improve the performance of software projects in terms of quality and efficiency. They also seek to deliver the soft-ware projects without any defects to the communities and just in time. The early revelation of defects in software projects is also tried to avoid failure of those projects, save costs, team effort, and time. Therefore, these companies need to build an intelligent model capable of detecting software defects accurately and efficiently. The paper is organized as follows. Section 2 presents the materials and methods, PRISMA, search questions, and search strategy. Section 3 presents the results with an analysis, and discussion, visualizing analysis and analysis per topic. Section 4 presents the methodology. Finally, in Section 5, the conclusion is discussed. The search string was applied to all electronic repositories looking for papers published between 2015 and 2021, which resulted in 627 publications. The results focused on finding three important points by linking the results of manuscript analysis and linking them to the results of the bibliometric analysis. First, the results showed that the number of defects and the number of lines of code are among the most important factors used in revealing software defects. Second, neural networks and regression analysis are among the most important smart and statistical methods used for this purpose. Finally, the accuracy metric and the error rate are among the most important metrics used in comparisons between the efficiency of statistical and intelligent models.

Author 1: Alia Nabil Mahmoud
Author 2: Vítor Santos

Keywords: Defects; software projects; statistical model; linear regression; logistic regression

PDF

Paper 29: A Hybrid Deep Neural Network for Human Activity Recognition based on IoT Sensors

Abstract: Internet of things (IOT) sensors, has received a lot of interest in recent years due to the rise of application demands in domains like ubiquitous and context-aware computing, activity surveillance, ambient assistive living and more specifically in Human activity recognition. The recent development in deep learning allows to extract high-level features automatically, and eliminates the reliance on traditional machine learning techniques, which depended heavily on hand crafted features. In this paper, we introduce a network that can identify a variety of everyday human actions that can be carried out in a smart home environment, by using raw signals generated from Internet of Thing’s motion sensors. We design our architecture basing on a combination of convolutional neural network (CNN) and Gated recurrent unit (GRU) layers. The CNN is first deployed to extract local and scale-invariance features, then the GRU layers are used to extract sequential temporal dependencies. We tested our model called (CNGRU) on three public datasets. It achieves an accuracy better or comparable to existing state of the art models.

Author 1: Zakaria BENHAILI
Author 2: Youssef BALOUKI
Author 3: Lahcen MOUMOUN

Keywords: IoT; deep learning; CNN; GRU; CNGRU; human activity recognition

PDF

Paper 30: Bioinformatics Research Through Image Processing of Histopathological Response to Stonefish Venom

Abstract: The present study utilizes coastal and environmental engineering to investigate the histopathological effects of Synanceia verrucosa venom on Albino BALB/c mice. S. verrucosa is the most hazardous venomous marine fish that belong to the family Synanceiidae, generally known as the "Reef Stonefish". Crude venom was collected from venom glands of the dorsal spines of stonefish samples taken from the Jordanian coastline of the Gulf of Aqaba, Red Sea. The mice were given intramuscular injections of the venom. Consequently, the research evaluated the acute toxicity and influence on selected serum biomarker enzymes, as well as possible histological alterations of the soleus skeletal muscles. The mice 24 h LD50 was 0.107 µg toxin/kg mouse body weight. After the treatment using venom sublethal dose, the serum biomarkers, including Lactate dehydrogenase (LDH) and Alanine aminotransferase (ALT), were significantly improved (P≤0.05). In addition, Lipid Peroxidation (LPO) contents were significantly increased (P≤0.05) after venom treatment. Moreover, we combined routine medical procedures and artificial intelligence-assisted image analysis for a rapid qualitative and quantitative diagnosis of stonefish injury, based on the histophotography of mice tissue samples during the observation period (1, 2, and 3 hours respectively). The novelty of our method is that we could detect severe and mild damage with an accuracy of 93% and 91%, respectively. The most histological abnormalities in muscles were the great variety in diameters, content, and widespread among randomly distributed muscle fibres. In addition, loss of the tissue's striated appearance was noticed in toxin-treated groups compared with the control group. Consequently, our findings indicate the Stonefish's harmful influences that may endanger human life and highlight the need for appropriate measures to be considered. This, in turn, can ensure beach safety in the Gulf of Aqaba.

Author 1: Mohammad Wahsha
Author 2: Heider A. M. Wahsheh
Author 3: Wissam Hayek
Author 4: Haya Al-Tarawneh
Author 5: Maroof Khalaf
Author 6: Tariq Al-Najjar

Keywords: Synanceia verrucosa; Gulf of Aqaba; artificial intelligence; marine biotoxins

PDF

Paper 31: Real Time Distributed and Decentralized Peer-to-Peer Protocol for Swarm Robots

Abstract: This contribution proposes an approach to enhance the capability of robotic agents to join the Internet of Things (IoT) and act autonomously in extreme and hostile environment. This capability will help in the development in environments where the connectivity, availability, and responsivity of the devices are subject to variations and noises. A real time distributed and decentralized Peer-to-Peer protocol was designed to allow Autonomous Unmanned Surface Vessels (AUSV) extend their context awareness. The developed Middleware allows a real time communication and is designed to run on top of a Real Time Operating System (RTOS). Furthermore, the proposed Middleware will give researchers access to a large amount of data collected by sensors, and thus solve one of the major problems encountered while training artificial intelligence models which is the lack of sufficient data.

Author 1: Mahmoud Almostafa RABBAH
Author 2: Nabila RABBAH
Author 3: Hicham BELHADAOUI
Author 4: Mounir RIFI

Keywords: Autonomous robots; smart objects; peer-to-peer; real time communication; ROS2; ZeroMQ; middleware

PDF

Paper 32: Improving Customer Churn Classification with Ensemble Stacking Method

Abstract: Due to the high cost of acquiring new customers, accurate customer churn classification is critical in any company. The telecommunications industry has employed single classifiers to classify customer churn; however, the classification accuracy remains low. Nevertheless, combining several classifiers' decisions improves classification accuracy. This article attempts to enhance ensemble integration via stack generalisation. This paper proposed a stacking ensemble based on six different learning algorithms as the base-classifiers and tested on five different meta-model classifiers. We compared the performance of the proposed stacking ensemble model with single classifiers, bagging and boosting ensemble. The performances of the models were evaluated with accuracy, precision, recall and ROC criteria. The findings of the experiments demonstrated that the proposed stacking ensemble model resulted in the improvement of the customer churn classification. Based on the results of the experiments, it indicates that the prediction accuracy, precision, recall and ROC of the proposed stacking ensemble with MLP meta-model outperformed other single classifiers and ensemble methods for the customer churn dataset.

Author 1: Mohd Khalid Awang
Author 2: Mokhairi Makhtar
Author 3: Norlina Udin
Author 4: Nur Farraliza Mansor

Keywords: Stacking ensemble; customer churn prediction; bagging; boosting

PDF

Paper 33: Enhancing the Takhrij Al-Hadith based on Contextual Similarity using BERT Embeddings

Abstract: Muslims are required to conduct Takhrij to validate the truth of Hadith text, especially when it is obtained from online media. Typically, the traditional Takhrij processes are conducted by experts and apply to Arabic Hadith text. This study introduces a contextual similarity model based on BERT Embedding to handle Takhrij on Indonesian Hadith Text. This study examines the effectiveness of BERT Fine-Tuning on the six pre-trained models to produce embedding models. The result shows that BERT Fine-Tuning improves the embedding model average accuracy by 47.67%, with a mean of 0.956845. The most high-grade accuracy was the BERT embedding built based on the indobenchmark/indobert-large-p2 pre-trained model on 1.00. In addition, the manual evaluation achieved 91.67% accuracy.

Author 1: Emha Taufiq Luthfi
Author 2: Zeratul Izzah Mohd Yusoh
Author 3: Burhanuddin Mohd Aboobaider

Keywords: Hadith text; Takhrij; natural language processing; text-similarity; word embedding; BERT fine-tuning

PDF

Paper 34: UX Testing for Mobile Learning Applications of Deaf Children

Abstract: Many studies are focusing on deaf children mobile learning. However, they are not concentrating on user experience (UX) testing. Current UX testing is based on existing UX evaluation models that are hard to apply due to the comprehensive measurements and lack of description on how to conduct evaluation for a more specific mobile learning process. Moreover, the existing UX evaluation models are not highlighted to be applied in testing UX for deaf children’s mobile learning. Hence, this paper proposed questions for UX testing for deaf children’s mobile learning to explore UX issues in offering an enjoyable learning application. Smileyometer is used to capture the data from deaf children after using a selected mobile learning application, KoTBaM and Learning Fakih. This study involves deaf children aged 7 – 12 years old familiar with the mobile application. The survey is divided into two sections: i) demographic information and ii) 24 questions that the respondent must answer using a smileyometer. The survey included 38 deaf children from Malaysian Deaf School. The participating deaf children completed the questionnaires with the assistance of their teachers after using the mobile learning application in the classroom. Yet, various issues needed to be addressed in order to improve the deaf children's user experience. Special exercises should be developed for deaf children connected to their school syllabus to consolidate their knowledge and self-learn everywhere. Furthermore, games elements should be adapted so the deaf children are able to learn while playing.

Author 1: Normala Mohamad
Author 2: Nor Laily Hashim

Keywords: User experience; UX testing; UX dimension; UX metrics; mobile learning application; deaf children; smileyometer

PDF

Paper 35: The Regularization Effect of Pre-activation Batch Normalization on Convolutional Neural Network Performance for Face Recognition System Paper

Abstract: Face recognition is of pronounced significance to real-world applications such as video surveillance systems, human computing interaction, and security systems. This biometric authenticating system encompasses rich real human face characteristics. As such, it has been one of the important research topics in computer vision. Face recognition systems based on deep learning approaches suffer from internal covariate shift problems that cause gradients to explode or gradient disappearance, which leads to improper network training. Improper network training causes network overfitting and computational load. This reduces recognition accuracy and slows down network speed. This paper proposes a modified pre-activation batch normalization convolutional neural network by adding a batch normalization layer after each convolutional layer within each of the four convolutional units of the proposed model. The performance of the proposed model is validated with a new dataset, AS-Darmaset, which is built out of two publicly available databases. This paper compared the convergence behavior of four different CNN models: the Pre-activation Batch Normalization CNN model, the Traditional CNN without Batch Normalization, the Post-Activation Batch Normalization CNN model, and the Sparse Batch Normalization CNN Architecture. The evaluation results show that the recognition performance of Pre-activation BN CNN has training and validation accuracies of 100.00% and 99.87%, the Post activation Batch normalization has 100.00% and 99.81%, and the traditional CNN without BN has 96.50% and 98.93%. The sparse batch normalization CNN has 96.25% and 97.60% success rate, respectively. The result shows that the Pre-activation BN CNN model is more effective than the other three deep learning models.

Author 1: Abu Sanusi Darma
Author 2: Fatma Susilawati Binti Mohamad

Keywords: Face recognition; pre-active batch normalization; convolutional neural network

PDF

Paper 36: Heuristics and Think-aloud Method for Evaluating the Usability of Game-based Language Learning

Abstract: Digital learning environments are increasingly becoming popular in recent years. The rising usage of cell phones has invited researchers to design and develop learning applications and games for mobile phones. Specifically, game-based language learning is being promoted by researchers in many parts of the world. “Language Learning Serious Game (LLSG)” is based on a theoretical model constructed by the researcher that supports children learning English as a second language in a cultural context. The usability of such games is evaluated based on well-defined heuristics and other standard methods. This research aims to appraise the usability of LLSG through heuristics and think-aloud approaches while involving all essential stakeholders, including language experts, students, teachers, and game developers. The researcher proposed the heuristics in a cultural context, whereas the think-aloud review is compiled from the rigorous discussion session involving these stakeholders to evaluate the LLSG. The findings obtained from the heuristics evaluation reveal that the usability of LLSG is acceptable. On the other hand, various interesting suggestions and reviews were gathered from the discussion between experts and students. This evaluation will further improve the future versions of the game.

Author 1: Kashif Ishaq
Author 2: Fadhilah Rosdi
Author 3: Nor Azan Mat Zin
Author 4: Adnan Abid

Keywords: Engagement; game-based; heuristic evaluation; language learning; motivation; think aloud; usability

PDF

Paper 37: Towards Measuring User Experience based on Software Requirements

Abstract: User Experience (UX) provides insights into the users’ product perceptions while using or intending to use an application. Software products are known for complexity and changeability, starting from requirements engineering until the product operation. Users often evaluate software UX based on a prototype; however, UX is semantically embedded in the software requirements, a crucial indicator for project success. The problem of current UX evaluation methods is their dependence on the actual involvement of users or experts, a time-consuming process. First, this paper builds a benchmark dataset of UX based on textual software requirements crowdsourcing several UX experts. Second, the paper develops a machine learning model to measure UX based on the dataset. This research describes the dataset characteristics and reports its statistical internal consistency and reliability. Results indicate a high Cronbach Alpha and a low root mean square error of the dataset. We conclude that the new benchmark dataset could be used to estimate UX instantly without the need for subjective UX evaluation. The dataset will serve as a foundation of UX features for machine learning models.

Author 1: Issa Atoum
Author 2: Jameel Almalki
Author 3: Saeed Masoud Alshahrani
Author 4: Waleed Al Shehri

Keywords: User experience; benchmark dataset; requirements engineering; UX evaluation; software engineering

PDF

Paper 38: Machine Learning Driven Feature Sensitive Progressive Sampling Model for BigData Analytics

Abstract: BigData requires processing a huge data volume, which is an undeniable challenge for academia-industries. The classical sampling techniques are limited when addressing data-imbalance, large data-heterogeneity, multi-dimensionality etc. To alleviate it, in this paper a novel machine learning driven feature sensitive progressive sampling (ML-FSPS) that in conjunction with an improved feature selection and classification environment achieves more than 95.7% of accuracy, even with 10-14% of the original data size. The proposed ML-FSPS model was applied for IoT-device classification problem that possesses exceedingly high data-imbalance, multi-dimensionality and heterogeneity issues. Functionally, the FSPS-driven analytics model at first performed active period segmentation followed by multi-dimensional (descriptive) statistical feature extraction and Wilcoxon Rank Sum Test based feature selection. Subsequently, it executed K-Means clustering over a gigantically huge feature instances (16,00,000,000 network traces) Here, K-means algorithm clustered each feature samples into five distinct clusters. With initial sample size of 10%, FSPS model selected same amount of data elements (0.5-5% iteratively) from each cluster for each feature to perform multi-class classification using homogenous ensemble learning (HEL) model. Here HEL encompassed AdaBoost, Random Forest and Extended Tree ensemble algorithms as base classifiers. The simulation results affirmed that the proposed model achieves accuracy of almost 99% even with 10-16% of sample size.

Author 1: Nandita Bangera
Author 2: Kayarvizhy N

Keywords: Feature sensitive progressive sampling; BigData analytics; machine learning; ensemble learning; rank sum test; IoT-device classification

PDF

Paper 39: Thermal-aware Dynamic Weighted Adaptive Routing Algorithm for 3D Network-on-Chip

Abstract: 3D Network-on-Chip NoC based systems have severe thermal problems due to the stacking of dies and disproportionate cooling efficiency of different layers. While adaptive routing can help with thermal issues, current routing algorithms are either thermally imbalanced or suffer from traffic congestion. In this work a novel thermal aware dynamic weighted adaptive routing algorithm has been proposed that takes traffic and temperature information into account and prevents packets being routed across congested and thermally aggravated areas. Dynamic weighted model will consider parameters related to congestion and thermal issues and provide a balanced suitable approach according to the current scenario at each node. The efficiency of the proposed algorithm is analyzed and evaluated with state-of-the-art thermal-aware routing algorithms using a simulation environment. Results obtained from the simulation shows that the proposed algorithm has performed better in terms of global average delay with 17-33 percent improvement and better thermal profiling under various synthetic traffic conditions.

Author 1: Muhammad Kaleem
Author 2: Ismail Fauzi Bin Isnin

Keywords: Routing algorithms; thermal-aware; dynamic weighted model; 3D Network-on-Chip

PDF

Paper 40: A New Back-off Algorithm with Priority Scheduling for MQTT Protocol and IoT Protocols

Abstract: The Internet of Things (IoT) protocols have encountered great challenges as the growth of technology has led to many limitations of the performance of the IoT protocols. Message Queuing Telemetry Transport protocol (MQTT) is one of the most dominant protocols in most fields of smart applications, so it has been chosen in this research to be a use case for implementing and evaluating a new proposed Back-off algorithm that is designed to eliminate suspicious and fake messages by calculating an initial frequent rate for each publisher connected to the MQTT broker. The proposed Back-off algorithm was designed to mitigate the traffic load of the uplink traffic by applying an exponential delay factor to suspicious publishers. Another priority scheduling algorithm was proposed to classify publishers as high priority or low priority depending on the new calculated frequent rate. The two algorithms were implemented on the Mosquitto broker and evaluated using a simulation environment by measuring specified performance metrics. The simulated results proved that the Back-off algorithm eliminated network load and introduced an acceptable range of CPU and RAM consumption. The results also concluded that the priority classification algorithm managed to reduce the latency of high-priority publishers.

Author 1: Marwa O Al Enany
Author 2: Hany M. Harb
Author 3: Gamal Attiya

Keywords: Back-off algorithm; priority scheduling; MQTT protocol; average transmission frequency rate; IoT protocols

PDF

Paper 41: Examining User Experience of Moodle e-Learning System

Abstract: This research investigates the user experience (UX) of the Moodle e-learning system employed at one university in Malaysia from students’ perspectives. Comprehensive user experience (UX) criteria were suggested, which was adopted from two reliable criteria, to evaluate the user experience (UX) of the e-learning system. The suggested comprehensive user experience (UX) criteria consist of 8 categories and 29 corresponding sub-categories; these can be used to evaluate teaching and learning, usability, and hedonic aspects of an e-learning system. Semi-structured interviews and questionnaires were employed based on the suggested user experience (UX) criteria to collect qualitative and quantitative data regarding users’ experience (UX) of the tested e-learning system. The results showed that the e-learning system had positive user experience (UX) in general from the students’ perspectives. The results also showed that the students were satisfied with most of the metrics related to teaching and learning, usability and hedonic. However, the students identified some challenges they faced while interacting with the e-learning system which could be improved in order to improve their user experience (UX) and gain more benefits from a good user experience (UX) e-learning system.

Author 1: Layla Hasan

Keywords: User Experience (UX); e-learning system; Moodle; usability; learning management system

PDF

Paper 42: Query Expansion based on Word Embeddings and Ontologies for Efficient Information Retrieval

Abstract: Information retrieval has been an ever-going process for end users to fetch relevant data at one go. The problem intensifies more with unstructured data in a semantic web environment. It is also a promising area for researchers to dive in and refine it from time to time. Expanding the user query and reformulating it is one probable solution to increase the efficiency of the information retrieval system. In this paper we propose “WeOnto”, a novel two-level query expansion algorithm that utilizes the combination of web ontologies and word embeddings for similarity calculation. In the first level, the Real estate Ontology (REO) is created using Protégé and Sparql queries are passed to retrieve probable semantic words from the given ontology for each inputted user query. The first level gave significant results and improved the information retrieval by 18%. The second level of algorithm uses word embedding enhanced with the domain knowledge that helps to retrieve similar meaningful words based on cosine similarity for the same user query. Word embeddings are implemented using Word2Vec method that follows two architectures namely CBOW or Skip Gram. Most similar semantic words are retrieved using the CBOW word embeddings method in the proposed algorithm and concatenated with the semantic keywords generated from the real estate ontology to form a powerful reformulated query that gives promising relevant results. Finally, two topmost words as per their similarity index are taken to reformulate the original user query. Experimental results depict that proposed algorithm has given distinct results and has showcased significant improvement of 93% over the initial user query.

Author 1: Namrata Rastogi
Author 2: Parul Verma
Author 3: Pankaj Kumar

Keywords: CBOW; Information retrieval; ontology; query reformulation; semantic web; skip gram; word embeddings; word2vec

PDF

Paper 43: A Novel Integrated Scheme for Detection and Mitigation of Route Diversion Attack in MANET

Abstract: With the involvement of Mobile Adhoc Network (MANET) in many upcoming technologies and applications, there is an increasing concern about secure data transmission. Until the last decade, various solutions have evolved to circumvent this threat; however, the security issue is still a more significant threat. The problems studied during the review are usage of Complex Cryptographic Usage, Less Energy Efficient, Fewer studies towards Route Diversion Attack, and Less Emphasis towards Securing Beacon. An analytical method has been used to study these problems. This paper introduces a novel scheme that carries out dual operation viz. i) assessing the link legitimacy for detection of route diversion attack, and ii) cost-effective countermeasures for the same attack. The key findings of proposed study is token generation process when associated with link legitimacy offers more routing security from various ranges of threats. The broader implication of this finding is that proposed system when characterized by lightweight encryption operation, it is capable of excelling better balance between data transmission and security performance unlike existing security solutions in MANET.

Author 1: H C Ramaprasad
Author 2: S. C. Lingareddy

Keywords: Mobile Adhoc network; route diversion attack; routing attack; link legitimacy; encryption

PDF

Paper 44: Multi-level Hierarchical Controller Assisted Task Scheduling and Resource Allocation in Large Cloud Infrastructures

Abstract: The high-pace emergence in Cloud Computing technologies demands and alarmed academia-industries to attain Quality-of-Service (QoS) oriented solutions to ensure optimal network performance in terms of Service Level Agreement (SLA) provision as well as Energy-Efficiency. Majority of the at-hand solutions employ Virtual Machine Migration to perform dynamic resource allocation which fails in addressing the key problem of SLA-sensitive scheduling where it demands timely and reliable task-migration solution. Undeniably, VM consolidation may help achieve energy-efficiency along with dynamic resource allocation where the classical heuristic methods which are often criticized for its local minima and premature convergence doesn’t guarantee the optimality of the solution, especially over large cloud infrastructures. Considering these key problems as motivation, in this paper a highly robust and improved meta-heuristic model based on Ant Colony System is developed to achieve Task Scheduling and Resource Allocation. CloudSim based simulation over different PlanetLab cloud traces exhibited superior performance by the proposed task-scheduling model in terms of negligible SLA violence, minimum downtime, minimum energy-consumption and higher number of migrations over other heuristic variants, which make it suitable towards realistic Cloud Computing purposes.

Author 1: Jyothi S
Author 2: B S Shylaja

Keywords: Task-scheduling; VM-migration; improved ant colony system; SLA assurance; energy-efficient consolidation

PDF

Paper 45: A Review of a Biomimicry Swimming Robot using Smart Actuator

Abstract: Biomimicry-based robotic mobility is a newer subgenre of bio-inspired design and it's all about applying natural concepts to the development of real-world engineered systems. Previously, researchers used actuators such as motors, pumps, and intelligent materials or intelligent actuators to build many biomimicry robots. Due to the field's growing interest, this study will examine the performance of several biomimicry robots that have been built based on their different design, the type of material the robot utilizes, and the type of propulsion for the robot to swim while providing huge thrust. Robots must not only design such an animal, but its maneuverability and control tactics must also be tied to wildlife to provide the finest impersonation of biological life. Fish propulsion can be separated into two categories which are body and/or caudal fins (BCF) and median and/or paired fins (MPF). The old propeller system in underwater robot usually uses motor and pump. Many researchers have begun developing smart materials as drivers in recent years that can be grouped into four categories: shape memory alloy SMA, ionic polymer metal composite IPMC, lead zirconate titanate (PZT) and pneumatic soft actuator as replacement for pump or motor. Varied materials produce different result and can be applied for different propulsion modes. Future researchers working on biomimetic fish robots will be guided by the findings of this study.

Author 1: Muhammad Shafique Ashroff Md Nor
Author 2: Mohd Aliff
Author 3: Nor Samsiah

Keywords: Biomimicry; fish propulsion; biological life; smart actuator

PDF

Paper 46: Fuel Consumption Prediction Model using Machine Learning

Abstract: In the paper, we are enhancing the accuracy of the fuel consumption prediction model with Machine Learning to minimize Fuel Consumption. This will lead to an economic improvement for the business and satisfy the domain needs. We propose a machine learning model to predict vehicle fuel consumption. The proposed model is based on the Support Vector Machine algorithm. The Fuel Consumption estimation is given as a function of Mass Air Flow, Vehicle Speed, Revolutions Per Minute, and Throttle Position Sensor features. The proposed model is applied and tested on a vehicle’s On-Board Diagnostics Dataset. The observations were conducted on 18 features. Results achieved a higher accuracy with an R-Squared metric value of 0.97 than other related work using the same Support Vector Machine regression algorithm. We concluded that the Support Vector Machine has a great effect when used for fuel consumption prediction purposes. Our model can compete with other Machine Learning algorithms for the same purpose which will help manufacturers find more choices for successful Fuel Consumption Prediction models.

Author 1: Mohamed A. HAMED
Author 2: Mohammed H.Khafagy
Author 3: Rasha M.Badry

Keywords: Fuel consumption; machine learning; support vector machine; feature weight; feature selection; on-board diagnostic

PDF

Paper 47: Scalable and Reactive Multi Micro-Agents System Middleware for Massively Distributed Systems

Abstract: IT transformation has revolutionized the business landscape and changed most of organizations business model into digital and innovation driven firms. To fully take advantage of this digitalization and the exponential growth of data, organizations need to rely on resilient, scalable, extremely connected, highly available & very performant systems. To meet this need, this paper presents a model of middleware for multi micro-agents system based on reactive programming and designed for massively distributed systems and High-Performance Computing, especially to face big data challenges. This middleware is based on multi-agents systems (MAS) which are known as a reliable solution for High Performance Computing. This proposal framework is built on abstraction and modularity principles through a multi-layered architecture. The design choices aim to ensure cooperation between heterogeneous distributed systems by decoupling the communication model and the cognitive pattern of micro agents. To ensure high scalability and to overcome networks latency, the proposal architecture uses distribution model of data & computing, that allows an adaptation of the grid size as needed. The resilience problem is addressed by adopting the same mechanism as Hazelcast middleware, thanks to his peer-to-peer architecture with no single point of failure.

Author 1: EZZRHARI Fatima Ezzahra
Author 2: EL ABID AMRANI Noureddine
Author 3: YOUSSFI Mohamed
Author 4: BOUATTANE Omar

Keywords: Massively distributed system; multi agent system (MAS); high performance computing; reactive programming; hazelcast

PDF

Paper 48: Expert System in Enhancing Efficiency in Basic Educational Management using Data Mining Techniques

Abstract: The importance of basic education is well noted in every country. Proper planning and utilization of resources at the basic level, helps in leveraging the success of education at all other levels of education in a country. Ghana is noted to be the country that spends higher in education than its West African counterparts. In Ghana, attempts made to plan by predicting and projecting expenditure as well as the available resources to manage basic education are not accurate enough to address the challenges of education in the country. With the issue of COVID 19 pandemic, more expenditure is realized in managing educational institutions as more resources are needed in observing the protocols to curtail the pandemic. This throws a serious challenge to the effective and efficient utilization of the limited resources in the country. In this paper, the data from the Ministry of Education is analysed using data mining techniques. This has helped to identify the inaccuracies in the data. Inaccurate population projection affects the Key Performance Indicators (KPIs) in education because population is a common denominator for educational indicators. A proposed expert system is to be developed to assist in managing the situation.

Author 1: Fuseini Inusah
Author 2: Yaw Marfo Missah
Author 3: Najim Ussiph
Author 4: Frimpong Twum

Keywords: Basic education; data mining; educational management; expert system; population

PDF

Paper 49: Machine Learning for Predicting Employee Attrition

Abstract: Employee attrition has become a focus of researchers and human resources because of the effects of poor performance on organizations regardless of geography, industry, or size. In this context, the use of machine learning classification models to predict whether an employee is likely to quit could greatly increase the human resource department’s ability to intervene on time and possibly provide a remedy to the situation to prevent attrition. This study is conducted with an objective to compare the performance machine learning techniques, namely, Decision Tree (DT) classifier, Support Vector Machines (SVM) classifier, and Artificial Neural Networks (ANN) classifier, and select the best model. These machine learning techniques are compared using the IBM Human Resource Analytic Employee Attrition and Performance dataset. Preprocessing steps for the dataset used in this comparative study include data exploration, data visualization, data cleaning and reduction, data transformation, discretization, and feature selection. In this study, parameter tuning and regularization techniques to overcome overfitting issues are applied for optimization purposes. The comparative study conducted on the three classifiers found that the optimized SVM model stood as the best model that can be used to predict employee attrition with the highest accuracy percentage of 88.87% as compared to the other classification models experimented with, followed by ANN and DT.

Author 1: Norsuhada Mansor
Author 2: Nor Samsiah Sani
Author 3: Mohd Aliff

Keywords: Artificial neural networks; decision tree; employee attrition; machine learning; support vector machines

PDF

Paper 50: Finding Good Binary Linear Block Codes based on Hadamard Matrix and Existing Popular Codes

Abstract: Because of their algebraic structure and simple hardware implementation, linear codes as class of error-correcting codes, are used in a multitude of situations such as Compact disk, backland bar code, satellite and wireless communication, storage systems, ISBN numbers and so more. Nevertheless, the design of linear codes with high minimum Hamming distance to a given dimension and length of the code, remains an open challenge in coding theory. In this work, we propose a code construction method for constructing good binary linear codes from popular ones, while using the Hadamard matrix. The proposed method takes advantage of the MacWilliams identity for computing the weight distribution, to overcome the problem of computing the minimum Hamming distance for larger dimensions.

Author 1: Driss Khebbou
Author 2: Reda Benkhouya
Author 3: Idriss Chana
Author 4: Hussain Ben-azza

Keywords: Binary linear codes; code construction; minimum hamming distance; error-correcting codes; weight distribution; coding theory; hadamard matrix

PDF

Paper 51: EFPT-OIDS: Evaluation Framework for a Pre-processing Techniques of Automatic Optho-Imaging Diagnosis and Detection System

Abstract: The modalities of FUNDUS images and the availability of public domain data sets provides a starting point in designing an ecosystem for developing an automatic detection of degenerative early-stage Glaucoma and Diabetic Retinopathy, and other eye-related diseases. The existing techniques for these operations lack flexibility and robustness in their design implementation and are limited to only certain preprocessing requirements. However, the existing methods are useful but provide lower performance when the FUNDUS image quality degrades due to misalignment of lens opening in camera and poor functioning of visual sensors. This paper presents a unified framework that mechanizes different preprocessing techniques to benefit the Optho-imaging diagnosis and disease detection process. The proposed framework facilitates on-demand data treatment operations that include image interpolation, brightness adjustment, illumination correction, and noise reduction. The proposed techniques for FUNDUS image enhancement provide better PSNR and SSIM-performance metrics for image quality than existing popular image enhancement techniques when tested on two standard publicly available datasets. The contribution of the proposed framework is that it offers flexible and effective mechanisms that meet dynamic preprocessing operations on an on-demand basis to prepare better data representation for building machine learning models. The framework can also be used in real-time for eye disease diagnosis by an ophthalmologist.

Author 1: Sobia Naz
Author 2: Radha Krishna Rao K. A
Author 3: Shreekanth T

Keywords: Pre processing; FUNDUS image; glaucoma; diabetic retinopathy; interpolation; image enhancement

PDF

Paper 52: A Delay-tolerant MAC Protocol for Emergency Care in WBAN Considering Preemptive and Non-preemptive Methods

Abstract: For facilitating pilgrims with no delay, quick and real-time emergency medical services at ritual sites a delay-tolerant Medium Access Control (MAC) protocol for the IEEE 802.15.6 standard based Wireless Body Area Networks (WBANs) has been proposed. Since MAC protocol is application-specific hence any particular MAC technique may not be appropriate for diverse applications. In this research work, we consider dealing with medical emergency traffics which is random, independent of each other and can be generated at any time. Moreover, emergency traffics must be transmitted ahead of normal medical data or emergency traffic with a lower severity level; because any delay in emergency data transmission may endanger patients’ life. The proposed MAC protocol is compared with both preemptive and non-preemptive methods. Where, a modified MAC superframe (SF) structure, minimum backoff period and minimum Contention Window (CWmin) for quick data access to the IEEE 802.15.6 standard based EAP channel are also considered. The proposed delay-tolerant MAC protocol has been experimented with and simulated by the Castalia simulator which is based on the OMNeT++ platform. The experimental results show that data transmission using the preemptive method works faster with reduced delay than that of the non-preemptive method. Furthermore, the delay metric of the proposed delay-tolerant MAC protocol is analyzed, calculated and compared with the current Traffic-aware TA-MAC protocol. Results demonstrate that delay is relatively low during emergency data transmission using the proposed MAC in WBANs environment.

Author 1: Shah Murtaza Rashid Al Masud
Author 2: Aloke Kumar Saha

Keywords: WBAN; MAC; preemptive; non-preemptive; delay; emergency traffic

PDF

Paper 53: Transformer based Contextual Model for Sentiment Analysis of Customer Reviews: A Fine-tuned BERT

Abstract: The Bidirectional Encoder Representations from Transformers (BERT) is a state-of-the-art language model used for multiple natural language processing tasks and sequential modeling applications. The accuracy of predictions from context-based sentiment and analysis of customer review data from various social media platforms are challenging and time-consuming tasks due to the high volumes of unstructured data. In recent years, more research has been conducted based on the recurrent neural network algorithm, Long Short-Term Memory (LSTM), Bidirectional LSTM (BiLSTM) as well as hybrid, neutral, and traditional text classification algorithms. This paper presents our experimental research work to overcome these known challenges of the sentiment analysis models, such as its performance, accuracy, and context-based predictions. We’ve proposed a fine-tuned BERT model to predict customer sentiments through the utilization of customer reviews from Twitter, IMDB Movie Reviews, Yelp, Amazon. In addition, we compared the results of the proposed model with our custom Linear Support Vector Machine (LSVM), fastText, BiLSTM and hybrid fastText-BiLSTM models, as well as presented a comparative analysis dashboard report. This experiment result shows that the proposed model performs better than other models with respect to various performance measures.

Author 1: Ashok Kumar Durairaj
Author 2: Anandan Chinnalagu

Keywords: Transformers model; BERT; sequential model; deep learning; RNN; LSVM; LSTM; BiLSTM; fastText

PDF

Paper 54: A New Energy-efficient Multi-hop Routing Protocol for Heterogeneous Wireless Sensor Networks

Abstract: Energy use of sensor nodes efficiently and extending the lifetime of heterogeneous wireless sensor networks (HWSNs) is a main goal of HWSNs routing optimization methods, and therefore building an energy-efficient routing protocol becomes critical for HWSN performance improvement. They present an energy-efficient routing protocol based on the grey wolf optimizer (GWO) and the Tabu search algorithm (TSA) in this paper. Proposed routing system with primary objectives include clustering and the selection of cluster heads (CH) utilizing GWO with a fitness function based on the residual energy of sensor nodes and the average distance between CH and sink nodes base station (BS ) due to the mobility of sensors, the quality of service (QoS) parameters such as reliability and energy consumption can be improved by discovering multiple optimized paths for data transmission from CH to BS and by TSA selecting the optimal route from CH to BS based on the forwarding of reliable route packets (FRRPs). The experimental results indicate that the proposed grey wolf optimizer with tabu Search Algorithm (GWO-TSA) can reduce HWSNs energy consumption by 10% and 20%, increase lifetime by 13% and 18%, and finally increase throughput by 6% and 14% when compared to the genetic algorithm with tabu search algorithm (GA-TSA) and grey wolf optimizer with crow search algorithm (GWO-CSA). When compared to GA-TSA & GWO-CSA, simulation reveals that the proposed GWO-TSA protocol improves HWSNs performance by minimizing energy consumption, maximizing network lifetime, and boosting network throughput.

Author 1: Rowayda A. Sadek
Author 2: Doha M. Abd-alazeem
Author 3: Mohamed M. Abbassy

Keywords: Heterogeneous wireless sensor networks (HWSNs); forwarding of reliable route packets (FRRPs); grey wolf optimizer (GWO); routing optimization; tabu search algorithm (TSA); quality of service (QoS)

PDF

Paper 55: Improved GRASP Technique based Resource Allocation in the Cloud

Abstract: In the era of cloud computing, everyone is somehow using cloud resources. However, the resources are limited in the Cloud. Cloud vendors look for enhanced returns on investments. Promising return on investment is possible only when the cloud resources are scheduled efficiently to execute jobs within the stipulated time. However, brute force methods require exponential time to produce a schedule. Heuristic and meta-heuristic algorithms have been proposed in the literature to allocate resources to the jobs. These algorithms still suffer from slow convergence. To overcome this problem, researchers clubbed various heuristics and meta-heuristic to form a new hybrid algorithm. With the same motive, this paper explores the limitations of greedy random adaptive search and shows that learning through a fixed set search enhances efficiency. Based on the results, it can be concluded that the proposed algorithm is on par with existing hybrid meta-heuristic algorithms.

Author 1: Madhukar E
Author 2: Ragunathan T

Keywords: Cloud computing; task scheduling; meta-heuristics; fixed set search; GRASP; resource allocation

PDF

Paper 56: Ada-IDS: AdaBoost Intrusion Detection System for ICMPv6 based Attacks in Internet of Things

Abstract: The magical buzzword Internet of Things (IoT) connects any objects which are diverse in nature. The restricted capacity, heterogeneity and large scale implementation of the IoT technology tend to have lot of security threats to the IoT networks. RPL is the routing protocol for the constraint devices like IoT nodes. ICMPv6 protocol plays a major role in constructing the tree-like topology called DODAG. It is vulnerable to several security attacks. Version Number Attack, DIS flooding attack and DAO attack are the ICMPv6 based attacks discussed in this paper. The network traffic is collected from the simulated environment in the normal and attacker settings. An AdaBoost ensemble model termed Ada-IDS is developed in this research to detect these three ICMPv6 based security attacks in RPL based Internet of Things. The proposed model detects the attacks with 99.6% accuracy and there is no false alarm rate. The Ada-IDS ensemble model is deployed in the Border Router of the IoT network to safeguard the IoT nodes and network.

Author 1: A. Arul Anitha
Author 2: L. Arockiam

Keywords: IoT; ICMPv6; version number attack; DIS attack; DAO attack; Ada-IDS

PDF

Paper 57: Analytical Framework for Binarized Response for Enhancing Knowledge Delivery System

Abstract: The student feedback offer effective insight into their experience of knowledge transfer, routinely collected in academic institutions. However, the existing research literature lacks reporting whether the comments in education system are helpful or non-useful. Most of the existing works are limited to sentiment polarity computation only, and teacher evaluation is carried out without considering the aspects of the teaching. This study analyzes student comments and classifies comments as applicable and non-useful for the teacher scoring system. In the proposed research, the data considered is the student feedback collected from the teacher rating website. The study performed phase-by-phase text data modeling. First, exploratory analysis is carried out on the student feedback dataset to understand text data characteristics and features. Based on the exploratory analysis, appropriate steps are determined to perform preprocessing operations for data cleaning. Using natural language processing context, the study only focuses on removing stop words and common words that belong to both useful and non-useful contexts. BoW model is considered for features extraction, and two probabilistic supervised machine learning models are used for comment classification. The study outcome exhibits that Gaussian Naïve Bayes outperforms Multinominal Naïve Bayes in accuracy, precision, recall rate, and F1-score.

Author 1: Chethan G S
Author 2: Vinay S

Keywords: Education; knowledge transfer; machine learning; natural language processing; student feedback

PDF

Paper 58: A Linguistic Analysis Metric in Detecting Ransomware Cyber-attacks

Abstract: Originating and striking from anywhere, cyber-attacks have become ever more sophisticated in our modern society and users are forced to adopt increasingly good and vigilant practices to protect from them. Among these, ransomware remains a major cyber-attack whose major threat to end users (disrupted operations, restricted files, scrambled sensitive data, financial demands, etc.) does not particularly lie in number but in severity. In this study we explore the possibility of real-time detection of ransomware source through a linguistic analysis that examines machine translation relative to the Levenshtein Distance and may thereby provide important indications as to attacker’s language of origin. Specifically, the aim of our research is to advance a metric to assist in determining whether an external ransom text is an indicator of either a human- or a machine-generated cyber-attack. Our proposed method works its argument on a set of Eastern European languages but is applicable to a large(r) range of languages and/or probabilistic patterns, being characterized by usage of limited resources and scalability properties.

Author 1: Diana Florea
Author 2: Wayne Patterson

Keywords: Cyber-security; cyber-attacks; machine translation; language; Levenshtein distance

PDF

Paper 59: Design and Implementation of Dynamic Packet Scheduling with Waiting Time Aware: DPSW2A

Abstract: One of the principal goals of 5G is to enhance performance in connection with speed and delay curtailment. To accomplish this, IETF has proposed Multipath TCP to utilize the accessible interface for communication. The demand for mobile communication is escalating day by day. The predominant communication option for people is mobile. For giving better service for users’, nodes are fitted out with multiple interfaces. Multiple interfaces are as well one of the benefits of 5G. Multi path protocols are used to load balancing and resilience to failure. When communicating with asymmetric interfaces, latency is an imperative factor. To attain low latency is hard when asymmetric interfaces are used. When communication happens using multiple interfaces, the scheduler plays a central role since it decides which interface needs to be used for the packet. In this article we spotlight on scheduling algorithms, how this schedule will play a vital role to transfer data to receiver nodes with low latency. In this paper, we emphasize on the Scheduler named DPSWWA with the objective of minimizing delay and effective usage of interfaces.

Author 1: K Raghavendra Rao
Author 2: B N Jagadesh

Keywords: MPTCP; scheduler; latency; delay; transport protocol; waiting time

PDF

Paper 60: Evaluating Domain Knowledge and Time Series Features for Automated Detection of Schizophrenia from EEG Signals

Abstract: Over the recent years, Schizophrenia has become a serious mental disorder that is affecting almost 21 million people globally. There are different symptoms to recognize schizophrenia from healthy people. It can affect the thinking pattern of the brain. Delusions, hallucinations, and disorganized speech are the common symptoms of Schizophrenia. In this study, we have used electroencephalography (EEG) signals to analyze and diagnose Schizophrenia using machine learning algorithms and found that temporal features performed well as compared to statistical features. EEG signals are the best way to analyze this disorder as they are intimately linked with human thinking patterns and provide information about brain activities. The present work proposes a Machine Learning (ML) model based on Logistic Regression (LR) along with two feature extraction libraries Time Series Feature Extraction Library (TSFEL) and MNE Python toolkit to diagnose Schizophrenia from EEG signals. The results are analyzed based on 5 different sampling techniques. The dataset was cross-validated using leave one subject out cross-validation (LOSOCV) using Scikit learn and achieve greater accuracy, sensitivity, specificity, macro average recall, and macro f1 score on temporal features respectively.

Author 1: Saqib Hussain
Author 2: Nasrullah Pirzada
Author 3: Erum Saba
Author 4: Muhammad Aamir Panhwar
Author 5: Tanveer Ahmed

Keywords: Artificial intelligence (AI); Logistic Regression (LR); smote-class weight (S-CW); borderline smote-class weight (BS-CW); electroencephalography; Time Series Feature Extraction Library (TSFEL)

PDF

Paper 61: Long Term Solar Power Generation Prediction using Adaboost as a Hybrid of Linear and Non-linear Machine Learning Model

Abstract: The usage of renewable energy sources has increased manifolds in terms of electric utilities. From other non-conventional sources, solar energy has been seen as a promising and convenient source used around the globe. In terms of meeting the daily requirements, solar energy has huge potential to fulfil the world’s demand. However, firstly the characteristic of solar energy is very unpredictable and intermittent due to variation of weather. Secondly, the optimization and the planning of smart grid effect the operation of PV system. Thus, prediction on the long horizon is needed to address this problem. Nevertheless, long term forecasting of solar power generation is deliberated as a challenging problem. Therefore, this paper proposes a 10 day ahead solar power forecasting using combination of linear and non-linear machine learning models. At first, the outputs are generated from Recurrent Neural Network (RNN), Support Vector Machine (SVM) and Autoregressive with exogenous variable (ARX). Later on, these three outputs are combined and are made as a strong classifier with the Adaptive boost (Adaboost) algorithm. The simulations were conducted on the data obtained from real PV plant. By the experimental results and discussion, it was endogenously concluded that the combination of all techniques with the Adaboost have increased the performances and showing the high accuracy as compare to the individual machine learning models. The hybrid Adaboost shows %MAPE 8.88, which proven high accuracy. While on the other hand, for the individual technique, RNN shows 10.88, SVM reveals 11.78 and ARX gives 13.00 of percentage MAPE. The improvement proves that combination of techniques performs better than individual models and proclaims the high accuracy.

Author 1: Sana Mohsin Babbar
Author 2: Chee Yong Lau
Author 3: Ka Fei Thang

Keywords: Recurrent neural network (RNN); support vector machine (SVM); autoregressive with exogenous variable (ARX); adaptive boosting (Adaboost) and photovoltaic system (PV)

PDF

Paper 62: HORAM: Hybrid Oblivious Random Access Memory Scheme for Secure Path Hiding in Distributed Environment

Abstract: Now-a-days in most of the sectors digitization has taken place to store data and process it easily with enhanced techniques. Online transactions produce very huge data daily in various sectors like health care, military, government office. To store huge data many firms, take the help of third-party organizations and store data on machines provided by them which creates new security issues. While performing operations on the data or accessing data metadata leakage may happen due to untrustworthy systems. This paper proposed hybrid oblivious random-access memory (HORAM) offers users to access their data from untrusted storage devices without sharing any information about their access patterns or techniques. Here random data block shuffling approach is used which helps in hiding storage policies about the user data blocks placement and preserving privacy of data. HORAM techniques perform pull-push operations on data in a parallel manner which in turn minimizes network overhead and reduces the execution time of operation. An extensive experimental analysis of the proposed system produces better results than weak and strong Federated oblivious random access memory (FEDORAM) respectively. The method is faster than weak FEDORAM and strong FEDORAM as it takes 0.96 seconds for communication with 5 servers whereas weak and strong FEDORAM takes 1.5 and 2 seconds respectively for reading and writing data.

Author 1: Snehalata Funde
Author 2: Gandharba Swain

Keywords: HORAM; metadata; data blocks; privacy; block shuffling

PDF

Paper 63: Blockchain-based Secure Data Transmission for UAV Swarm using Modified Particle Swarm Optimization Path Planning Algorithm

Abstract: With the rapid development of unmanned aerial vehicles assisted applications enabled with a communication system, they are open to various malicious attacks. As a new form of flying things, they can access the network for better communication via the aerial base station. Most of the Unmanned aerial vehicles assisted flying objects' optimal path selection schemes does not consider the path deviation. In path deviation attacks, secure data transmission are not addressed in existing works. The secure communication process between Unmanned aerial vehicles and base station are exploited through security-based attacks. Moreover, path loss issue leads to multicast packet loss and unsecured broadcast. The existing network architecture setup does not fulfill the secure data communication and privacy issues. In this paper, Blockchain is utilized to investigate the secure communication between Unmanned aerial vehicles to Wireless Unmanned aerial vehicles Base stations. Since the destination information is dynamic under an uncertain environment, it will cause a delay in data communication. Unmanned aerial vehicles are more vulnerable to security attacks. The proposed blockchain-based architecture supports secure data communication in Unmanned aerial vehicles uncertain environments. To improve network security, this paper designs a modified particle swarm optimization method for better path selection. Through these experimental results, a blockchain-based data communication scheme is outer performed concerning network security.

Author 1: M. Kayalvizhi
Author 2: S. Ramamoorthy

Keywords: Unmanned aerial vehicles; path planning; swarm optimization; denial of service attack; blockchain; security and privacy; data communication

PDF

Paper 64: Resource Allocation in Spectrum Deployment for Cognitive Third-party Users

Abstract: Spectrum scarcity is a major challenge in wireless communication for next generation applications. The spectrum sharing and utilization of other user available spectrum is an optimal solution, which has outcome with a new mode of communication called cognitive network. Cognitive devices are capable of requesting and sharing free spectrum among each other in a communication range. The spectrums are shared among primary and secondary user (PU, SU). To extend the range of spectrum utilization, users from other clusters are requested in sharing of spectrum which is called as third party user (TSU). In detection of free spectrum a jamming based approach is proposed in recent work. Jamming approach generates a periodic jamming signal for TSU in sensing of free spectrum. The repetitive requesting of spectrum availability result in large jamming probability for engaged TSU users resulting in large delay. In minimizing the delay observed, in this paper, a new monitored jamming approach is proposed. Proposed Monitored jamming approach is develop in recent to the engagement of each TSU for communication and governing the jamming signal based on the spectrum engagement. The proposed approach minimizes the delay due to jamming signaling in existing system. The Experimental results obtained shows an enhancement to the system throughput, fairness factor and minimizes the delay metric for proposed approach.

Author 1: Arikatla Jaya Lakshmi
Author 2: G. N. Swamy
Author 3: M. N. Giri Prasad

Keywords: Slot monitoring; spectrum sensing; primary user (PU); secondary user (SU); third-party spectrum utilization (TSU)

PDF

Paper 65: The Use of the Relational Concept in the Arabic Morphological Analysis

Abstract: The Arabic language differs from other natural languages in its structures and compositions. In this article we have developed an Arabic morphological analyzer. For this, we have used the relational concept in the database to build our Arabic morphological analyzer. This analyzer uses a set of tables which are linked together by relationships. These relations model certain numbers of compatibility rules between different affixes. Our morphological analysis have been trained and tested on the same databases. The tests of our new approach have given good results and the numbers obtained are very close to those of existing analyzer.

Author 1: Said Iazzi
Author 2: Abderrazak Iazzi
Author 3: Saida Laaroussi
Author 4: Abdellah Yousfi

Keywords: Arabic language rules; morphology; morphological analyzers; database; relational concept

PDF

Paper 66: Development of Wearable System to Help Preventing the Spread of Covid-19 in Public Indoor Area

Abstract: Smart wearables as a part of the Internet of Things nowadays gaining confidence in our daily lives because of its accessibility and simplicity. Today, with the outbreak of Coronavirus around the world, a smart wearable device can become another solution to help slowing the virus spreading by ensuring public health and social measures. In this paper, a system consisting of distance detector and touchless door access is proposed to help the personal, physical and social distancing measures practice in a public indoor area. A BLE positioning method based on RSSI localization is used to ensure the physical distancing around the user. WPA2 and MAC address-based authentication for the touchless door access is used to restrict and trace the visitor of the indoor area. The system is implemented in ESP microcontroller. A proof of concept is conducted to see if the functionality of the system already satisfied the public health and social measures practice. The results show that only registered devices can give a signal to open the door and the device can guarantee the physical distance around the user with 4.51% error in indoor area.

Author 1: Annisa Istiqomah Arrahmah
Author 2: Surya Ramadhan

Keywords: Wearable device; healthcare system; COVID-19; door-lock system; radio signal strength indicator (RSSI)

PDF

Paper 67: A Comparative Study of Segmentation Method for Computer-aided Diagnosis (CAD) Leukemia AML Subtype M0, M1, and M2

Abstract: A computer-based diagnosis model for Acute Myelogenous Leukemia (AML) is carried out using white blood cell image processing. The stages in computer-aided diagnosis (CAD) are included pre-processing, segmentation, feature extraction, and classification. The segmentation method has many approaches, namely, clustering, region growing, and thresholding. The number of approaches that can be used requires proper selection because it will have an impact on CAD performance. This study aims to conduct a comparative study of the performance of the WBC segmentation method on the AML M0, M1, and M2 subtype leukemia CAD system. The segmentation algorithm used is k-means, fuzzy c-means, SOM, watershed, chan vese (active contour), otsu thresholding, and histogram. The feature extraction method uses GLCM, while the classification algorithms tested are SVM, Random-forest, decision tree, naive Bayesian, and k-NN. The test results show that the histogram segmentation method is able to provide the best average performance when using SVM, namely 90.3% accuracy, 85.9% sensitivity, and 92.7% specificity.

Author 1: Wiharto
Author 2: Wisnu Widiarto
Author 3: Esti Suryani
Author 4: Nurmajid Hidayatullah

Keywords: Acute myelogenous leukemia; leukemia; segmentation; feature extraction; classification

PDF

Paper 68: EC-Elastic an Explicit Congestion Control Mechanism for Named Data Networking

Abstract: In recent years, Named Data Networking (NDN) has attracted researchers’ attention as a new internet architecture. NDN changes the internet communication paradigm from a host-to-host IP model to a name-based model. Thus, NDN permits the retrieval of requested content by name, from different sources and via multiple paths, and the use of caching in intermediate routers. These features transform the transport control model from sender to receiver and make traditional end-to-end congestion control mechanisms incompatible with NDN architecture. To deal with this problem, a reliable congestion control mechanism becomes necessary for a successful deployment of NDN. This paper presents a new hybrid congestion control mechanism for NDN, EC-Elastic (Explicit Congestion Elastic), which adopts the basic concept of Elastic-TCP to control the sending rates of the interest packets at the consumer nodes. In the intermediate nodes, a queue has been associated with the Controlled Delay-Active Queue Management CoDel-AQM to measure the packet sojourn time and notify the consumer to decrease its interest packet sending rate when it receives an explicit congestion signal. EC-Elastic was implemented in ndnSIM and evaluated with Agile-SD, CUBIC, and STCP in different scenarios. Simulation results show that EC-Elastic provides a significant improvement in bandwidth utilization while maintaining lower delay and packet loss rates.

Author 1: Asmaa EL-BAKKOUCHI
Author 2: Mohammed EL GHAZI
Author 3: Anas BOUAYAD
Author 4: Mohammed FATTAH
Author 5: Moulhime EL BEKKALI

Keywords: NDN; named data networking; congestion control; explicit congestion control; TCP-elastic

PDF

Paper 69: 1D-CNN based Model for Classification and Analysis of Network Attacks

Abstract: With the advancement in technology and upsurge in network devices, more and more devices are getting connected to the network leading to more data and information on the network which emphasizes the security of the network to be of paramount importance. Malicious traffic must be detected in networks and machine learning or more precisely deep learning (DL), which is an upcoming approach, should be used for better detection. In this paper, Detection of attacks through a classification of traffic into normal and attack data is done using 1D-CNN, a special variant of convolutional neural network (CNN). For this, the CICIDS2017 dataset consisting of 14 attack types spread across 8 different files, is considered for evaluating model performance and various indicators like recall, precision, F1-score have been utilized. Separate 1D-CNN based DL models were built on individual sub-datasets as well as on combined datasets. Also, an evaluation of the model is done by comparing it with an artificial neural network (ANN) model. Experimental results have demonstrated that the proposed model has performed better and shown great capability in detecting network attacks as the majority of the class labels had achieved excellent scores in each of the evaluation indicators used.

Author 1: Kuljeet Singh
Author 2: Amit Mahajan
Author 3: Vibhakar Mansotra

Keywords: 1D-CNN; CICIDS2017; network attacks; deep learning

PDF

Paper 70: Applying Grey Systems and Inverse Distance Weighted Method to Assess Water Quality from a River

Abstract: The Cañete River basin, in Peru, has suffered an increase in pollution due to various causes, among the main ones being the lack of knowledge, culture of individuals and municipal authorities, economic activities, among others. We analyze this degree of pollution reflected in the upper part of the Cañete river basin through the Grey Clustering method, based on grey systems, which is presented as a good alternative to evaluate water quality in a comprehensive way, making use of Historical data from the monitoring program for the years 2014 and 2015 with nine monitoring points carried out by the National Water Authority (ANA), 6 parameters were defined to evaluate: Hydrogen potential (pH), Biochemical oxygen demand, Chemical demand of Oxygen, Total Suspended Solids, Total Manganese and Total Iron based on the PRATI index. For the spatial distribution, interpolation surfaces of the clustering coefficients were created, using the Spatial Analyst extension of the ArcGIS software, which provides tools to create, analyze and map data in raster format or surfaces. The interpolation method used is Inverse Distance Weighted (IDW). The results of the evaluation showed that in 2014 the monitoring points determined, through the Grey Clustering method, a level of contamination "Uncontaminated" at each point except for point P7 which gives us an "Acceptable" level according to the PRATI indices, while for the year 2015 points P1 and P2 indicate a level of contamination "Moderately contaminated", point 3 an "Acceptable" level, after points P4 to P9 they present a level of "Not contaminated". Finally, the Grey Clustering analysis method will determine the water quality in the 9 monitoring points of the upper-middle basin of the Cañete River in the years 2014 and 2015. Allowing to observe the reduction of water quality in points P1 and P2 for the period of the years 2014 and 2015 respectively, being crucial to achieve water resource management among local governments that can insert awareness and sustainable development policies.

Author 1: Alexi Delgado
Author 2: Anthonny Fernandez
Author 3: Eduardo Lozano
Author 4: Dennis Miguel
Author 5: Félix León
Author 6: Jhosep Arteta
Author 7: Ch. Carbajal

Keywords: Grey systems; inverse distance weighted (IDW); water quality

PDF

Paper 71: Unsupervised Machine Learning Approach for Identifying Biomechanical Influences on Protein-Ligand Binding Affinity

Abstract: Drug discovery is incredibly time-consuming and expensive, averaging over 10 years and $985 million per drug. Calculating the binding affinity between a target protein and a ligand through Virtual Screening is critical for discovering viable drugs. Although supervised machine learning (ML) can predict binding affinity accurately, models experience severe overfitting due to an inability to identify informative properties of protein-ligand complexes. This study used unsupervised ML to reveal underlying protein-ligand characteristics that strongly influence binding affinity. Protein-ligand 3D models were collected from the PDBBind database and vectorized into 2422 features per complex. Principal Component Analysis (PCA), t-Distributed Stochastic Neighbor Embedding (t-SNE), K-Means Clustering, and heatmaps were used to identify groups of complexes and the features responsible for the separation. ML benchmarking was used to determine the features’ effect on ML performance. The PCA heatmap revealed groups of complexes with binding affinity of pKd<6 and pKd>8 and identified the number of CCCH and CCCCCH fragments in the ligand as the most responsible features. A high correlation of 0.8337, their ability to explain 18% of the binding affinity’s variance, and an error increase of 0.09 in Decision Trees when trained without the two features suggests that the fragments exist within a larger ligand substructure that significantly influences binding affinity. This discovery is a baseline for informative ligand representations to be generated so that ML models overfit less and can more reliably identify novel drug candidates. Future work will focus on validating the ligand substructure’s presence and discovering more informative intra-ligand relationships.

Author 1: Arjun Singh

Keywords: Drug discovery; unsupervised machine learning; feature engineering; protein-ligand binding affinity; virtual screening

PDF

Paper 72: Statistical Analysis of Cybersecurity Awareness Issues in Higher Education Institutes

Abstract: Now-a-days, computers and the Internet are becoming increasingly indispensable tools in several aspects of our lives, including academic study, professional work, entertainment, and communication. Despite the significant advantages of information technology, particularly in information accessibility and internet applications, cyber security has risen to become a national concern in Saudi Arabia, and cyber security threats now need to be taken more seriously. Therefore, computer and network security are a concern not only for traditional security awareness organisations, for example, military, bank, or financial institutions, but also for every individual and government official who use computers. Besides, nowadays, more and more organisations’ valuable assets are stored in the computerised information system; security has become an essential and urgent issue. However, it is remarkable that most systems today are designed with little attention to security concerns. This study aims to examine and analyse cyber security issues, including cyber risk, cyber security, cyber security awareness, and cyber trust, among higher education students in Saudi Arabia. Based on an analysis of the collected data using SPSS, the findings of this study highlight a lack of awareness of basic information related to cyber security among Saudi students. In addition, the number of students attending training programs was very low. Considering other security issues, this study reveals that while Saudi students are aware of cyber risk, they are not aware of cyber security. In addition, Saudi students are not aware of and do not have cyber trust.

Author 1: Latifa Alzahrani

Keywords: Cyber security; cyber security awareness; educational institutes; cyber risk; higher education; cyber trust

PDF

Paper 73: Ontology-based Daily Menu Recommendation System for Complementary Food According to Nutritional Needs using Naïve Bayes and TOPSIS

Abstract: Babies begin to be given complementary feeding at the age of 6 to 24 months. Complementary foods given to babies need to meet nutritional needs according to their ages. Since, at these ages, babies are just learning to eat, it is necessary to plan a complementary food menu referring to the nutritional needs and the baby and mother's preferences. It is certainly not an easy thing for a mother. Therefore, a recommendation system is needed to determine the baby's daily menu according to those all. This research proposes a complementary food menu recommendation system that considers the balanced composition of three significant nutrients (carbohydrates, protein, and fat) in the diet. It also takes into account the baby and mother's preferences. The ontology contains Knowledge-based about food and its nutritional content and the nutritional needs of babies according to their ages. Naive Bayes is used to prepare menu options according to user preferences. TOPSIS method is used in this study to provide optimal recommendations regarding nutritional balance and user preferences. Several mothers who have had babies aged 6-24 months and mothers of babies aged 6-24 months were asked to test the recommendation system. The results of the usability testing of the system using SUS showed a good level of user satisfaction.

Author 1: Mujahidah Showafah
Author 2: Sari Widya Sihwi
Author 3: Winarno

Keywords: Calorie; complementary food; babies; Naïve Bayes; nutrition needs; ontology; recommendation system; SUS; TOPSIS

PDF

Paper 74: Emerging Requirement Engineering Models: Identifying Challenges is Important and Providing Solutions is Even Better

Abstract: Requirement Engineering is one of the most crucial tasks because it serves as the foundation for any software. Four pillars of requirement engineering procedures underpin the entire software. The bricks that make up the software edifice are functional and non-functional needs. Finally, design, implementation, and testing add stories to the foundation, allowing a full software tower to be built on top of it. As a result, the foundation must be strong enough to support the remainder of the software tower. Requirement engineers have various hurdles to design successful software for this purpose. Requirement Engineering (RE) is emerging as an increasingly important discipline to promote the development of web applications, as these are designed to meet various stakeholder requirements, additional functional, information, multimedia, and usability requirements compared to traditional software applications. The requirements of software systems are a very important area in software engineering. The success of software systems depends on how it effectively meets the requirements of users. In this paper, the review of current state of requirements engineering in which requirements from users are checked analyzed with their consistency and correctness is presented, and then identifies the emerging models of requirement engineering. Firstly, the paper highlight the current activities that enable the understanding of goals and objectives for developing proposed software systems, then the focus is on the techniques for improving the precision, accuracy, and variety of requirements. Next, identification of the challenges of emerging requirement engineering models is explained. The challenges like security and global trend that posed by emerging models of the future. Finally, we are trying to suggest some solutions for the mentioned challenges.

Author 1: Hina Noor
Author 2: Maheen Tariq
Author 3: Anam Yousaf
Author 4: Hafiz Wajid Ali
Author 5: Arqam Abdul Moqeet
Author 6: Abu Bakar Hamid
Author 7: Mahek Hanif
Author 8: Huma Naz
Author 9: Nabeel Tariq
Author 10: Ijaz Amin
Author 11: Osama Naseer

Keywords: Requirement engineering; current requirement engineering methods; emerging models for requirement engineering; challenges; pros and cons

PDF

Paper 75: An Advanced Ontology based Deep Learning for Computer-aided Interpretation of Mammography Images

Abstract: This paper has been removed.

Author 1: Hamida Samiha Rahli
Author 2: Nacéra Benamrane

Keywords: Computer assisted detection (CAD); breast cancer; the digital database for screening mammography (DDSM); ontology; deep learning; breast imaging-reporting and data system (BIRADS)

PDF

Paper 76: Electronic Commerce Product Recommendation using Enhanced Conjoint Analysis

Abstract: While finding any product, there are many identical products sold in the marketplace, so buyers usually compare the items according to the desired preferences, for example, price, seller reputation, product reviews, and shipping cost. From each preference, buyers count subjectively to make a final decision on which product is should be bought. With hundreds of thousands of products to be compared, the buyer may not get the product that meets his preferences. To that end, we proposed the Enhanced Conjoint Analysis method. Conjoint Analysis is a common method to draw marketing strategy from a product or analyze important factors of a product. From its feature, this method also can be used to analyze important factors from a product in the marketplace based on price. We convert importance factor percentage as a coefficient to calculate weight from every attributes and summarize it. To evaluate this method, we compared the ECA method to another prediction algorithm: generalized linear model (GLM), decision tree (DT), random forest (RF), gradient boosted trees (GBT), and support vector machine (SVM). Our experimental results, ECA running time is 6.146s, GLM (5.537s), DT (1s), RF (10,119s), GBT (45.881s), and SVM (11.583s). With this result, our proposed method can be used to create recommendations besides the neural network or machine learning approach.

Author 1: Andrew Brian Osmond
Author 2: Fadhil Hidayat
Author 3: Suhono Harso Supangkat

Keywords: Enhanced conjoint analysis; marketplace; ecommerce

PDF

Paper 77: Normalisation of Indonesian-English Code-Mixed Text and its Effect on Emotion Classification

Abstract: Usage of code-mixed text has increased in re-cent years among Indonesian internet users, who often mix Indonesian-language with English-language text. Normalisation of this code-mixed text into Indonesian needs to be performed to capture the meaning of English parts of the text and process them effectively. We improve a state-of-the-art code-mixed Indonesian-English normalisation system by modifying its pipeline modules. We further analyse the effect of code-mixed normalisation on emotion classification tasks. Our approach significantly improved on a state-of-the-art Indonesian-English code-mixed text normal-isation system in both the individual pipeline modules and the overall system. The new feature set in the language identification module showed an improvement of 4.26% in terms of F1 score. The combination of machine translation and ruleset in the lexical normalisation module improved BLEU score by 25.22% and lowered WER by 62.49%. The use of context in the translation module improved BLEU score by 2.5% and lowered WER by 8.84%. The effectiveness of the overall pipeline normalisation system increased by 32.11% and 33.82%, in terms of BLEU score and WER, respectively. Code-mixed normalisation also improved the accuracy of emotion classification by up to 37.74% in terms of F1 score.

Author 1: Evi Yulianti
Author 2: Ajmal Kurnia
Author 3: Mirna Adriani
Author 4: Yoppy Setyo Duto

Keywords: Code-mixed normalisation; Indonesian; English; emotion classification

PDF

Paper 78: Sign Language Gloss Translation using Deep Learning Models

Abstract: Converting sign language to a form of natural language is one of the recent areas of the machine learning domain. Many research efforts have focused on categorizing sign language into gesture or facial recognition. However, these efforts ignore the linguistic structure and the context of natural sentences. Traditional translation methods have low translation quality, poor scalability of their underlying models, and are time-consuming. The contribution of this paper is twofold. First, it proposes a deep learning approach for bidirectional translation using GRU and LSTM. In each of the proposed models, Bahdanau and Luong’s attention mechanisms are used. Second, the paper experiments proposed models on two sign languages corpora: namely, ASLG-PC12 and Phoenix-2014T. The experiment con-ducted on 16 models reveals that the proposed model outperforms the other previous work on the same corpus. The results on the ASLG-12 corpus, when translating from text to gloss, reveal that the GRU model with Bahdanau attention gives the best result with ROUGE (Recall-Oriented Understudy for Gisting Evaluation) score 94.37% and BLEU (Bilingual Evaluation Understudy)-4 score 83.98%. When translating from gloss to text, the results also show that the GRU model with Bahdanau attention achieves the best result with ROUGE score 87.31% and BLEU-4 66.59%. On Phoenix-2014T corpus, the results of text to gloss translation show that the GRU model with Bahdanau attention gives the best result in ROUGE with a score of 42.96%, while the GRU model with Luong attention gives the best result in BLEU-4 with 10.53%. When translating from gloss to text, the results report that the GRU model with Luong attention achieves the best result in ROUGE with a score of 45.69% and BLEU-4 with a score of 19.56%.

Author 1: Mohamed Amin
Author 2: Hesahm Hefny
Author 3: Ammar Mohammed

Keywords: Sequence to sequence model; neural machine trans-lation; sign language; deep learning; LSTM; GRU

PDF

Paper 79: Swapping-based Data Sanitization Method for Hiding Sensitive Frequent Itemset in Transaction Database

Abstract: Sharing a transaction database with other parties for exploring valuable information becomes more recognized by business institutions, i.e., retails and supermarkets. It offers various benefits for the institutions, such as finding customer shopping behavior and frequently bought items, known as fre-quent itemsets. Due to the importance of such information, some institutions may consider certain frequent itemsets as sensitive information that should be kept private. Therefore, prior to handling a database, the institutions should consider privacy preserving data mining (PPDM) techniques for preventing sen-sitive information breaches. Presently, several PPDM methods, such as item suppression-based methods and item insertion-based methods have been developed. Unfortunately, the methods result in significant changes to the database and induce some side effects such as hiding failure, significant data dissimilarity, misses cost, and artificial frequent itemset occurrence. In this paper, we propose a swapping-based data sanitization method that can hide the sensitive frequent itemset while at the same time it can minimize the side effects of the data sanitization process. Experimental results indicate that the proposed method outperforms existing methods in terms of minimizing the side effects.

Author 1: Dedi Gunawan
Author 2: Yusuf Sulistyo Nugroho
Author 3: Maryam

Keywords: Transaction database; data sanitization; data min-ing; sensitive frequent itemset; swapping-based method

PDF

Paper 80: Software Security Static Analysis False Alerts Handling Approaches

Abstract: False Positive Alerts (FPA), generated by Static Analyzers Tools (SAT), reduce the effectiveness of the automatic code review, letting them be underused in practice. Researchers conduct a lot of tests to improve SAT accuracy while keeping FPA at a lower rate. They use different simulated and production datasets to validate their proposed methods. This paper surveys recent approaches dealing with FPA filtering; it compares them and discusses their usefulness. It also studies the used datasets to validate the identified methods and show their effectiveness to cover most program defects. This study focuses mainly on the security bugs covered by the datasets and handled by the existing methods.

Author 1: Aymen Akremi

Keywords: Software security; static analysis; false alert reduc-tion; source code dataset; security bugs

PDF

Paper 81: CDRA: A Community Detection based Routing Algorithm for Link Failure Recovery in Software Defined Networks

Abstract: The increase in size and complexity of the Internet has led to the introduction of Software Defined Networking (SDN). SDN is a new networking paradigm that breaks the limitations of traditional IP networks and upgrades the current network infrastructures. However, like traditional IP networks, network failures may also occur in SDN. Multiple research studies have discussed this problem by using a variety of techniques. Among them is the use of the community detection method is one of the failure recovery technique for SDN. However, this technique have not considered the specific problem of multiple link multi-community failure and inter-community link failure scenarios. This paper presents a community detection-based routing algorithm (CDRA) for link failure recovery in SDN. The proposed CDRA scheme is efficient to deal with single link intra-community failure scenarios and multiple link multi-community failure scenarios and is also able to handle the inter-community link failure scenarios in SDN. Extensive simulations are performed to evaluate the performance of the proposed CDRA scheme. The simulation results depicts that the proposed CDRA scheme have better simulations results and reduce average round trip time by 35.73%, avg data packet loss by 1.26% and average end to end delay 49.3% than the Dijkstra based general recovery algorithm and also can be used on a large scale network platform.

Author 1: Muhammad Yunis Daha
Author 2: Mohd Soperi Mohd Zahid
Author 3: Babangida Isyaku
Author 4: Abdussalam Ahmed Alashhab

Keywords: Software Defined Network (SDN); community detection methods; CDRA; link failure

PDF

Paper 82: A Reliable Lightweight Trust Evaluation Scheme for IoT Security

Abstract: The rapid development of smart devices and the consequent demand their reliability have posed many challenges limiting their versatility. One of the most significant challenges is safeguarding the widespread network of sensors and devices within harsh remote environments. Numerous trust schemes have been proposed to overcome related IoT security concerns. How-ever, most of these schemes are not lightweight and consequently are not energy-efficient. This paper proposes a reliable lightweight trust evaluation scheme (RTE) to mitigate the malicious behavior of the nodes within IoT networks. The nodes are grouped into a set of clusters each having a cluster head while cluster members are categorized by evaluating their associated residual energy. Nodes with residual energy lower than the threshold (which is determined by the base station) are suspended until they recover and regain their activity. The computations are handled by the CH which is elected by an algorithm according to its energy and coverage degree in order to optimize the energy consumption in the network. For validation and performance evaluation, the proposed RTE scheme was compared to three of the recent schemes in its category. The obtained results have revealed that the proposed RTE scheme outperforms all of them in terms of detection rate, trust evaluation time, and energy efficiency.

Author 1: Hamad Aldawsari
Author 2: Abdel Monim Artoli

Keywords: IoT security; clustering; trust; energy efficient algorithm

PDF

Paper 83: Multi-Criteria Decision-Making Approach for Selection of Requirements Elicitation Techniques based on the Best-Worst Method

Abstract: During requirements elicitation stage, requirements engineers gather system requirements and drive stakeholders to convey needs and desired software functionality. The elicitation techniques used to acquire software requirements significantly impact the quality of elicited requirements. Several elicitation techniques have been proposed for the Requirement Engineering (RE) process; however, these techniques are rarely used in reality due to a lack of empirical and relative appraisals to assist software team members in deciding on the most appropriate technique. Re-quirement engineers encounter difficulty in deciding the suitable elicitation technique to adopt for a certain software project. This difficulty is due to a lack of knowledge regarding the available elicitation techniques, their efficacy, and how appropriate they are for a certain project. According to the literature, requirements engineering processes benefit from the use of Multi-Criteria Decision-Making (MCDM) approaches within particular contexts. An optimal structure is constitutionally presented within the area of the requirements engineering process; hence, the demonstra-tion of a robust decision-making method in the requirements engineering process should motivate a higher level of satisfaction with software projects developed in this way. This study proposes an approach for using the MCDM method in the requirements engineering process. The study contains a model for investigating the selection of an appropriate elicitation technique based on a decision-making method, namely, the Best-Worst Method (BWM). The findings of the proposed model demonstrate the BWM’s power in solving complex decision problems involving several criteria and alternatives.

Author 1: Abdulmajeed Aljuhani

Keywords: Requirements elicitation; elicitation techniques; de-cision support methods; Best-Worst Method; BWM

PDF

Paper 84: e-Business Model to Optimise Sales through Digital Marketing in a Peruvian Company

Abstract: The COVID-19 pandemic has affected the Peruvian market, generating a great loss in sales in Peruvian companies. The objective of the research is to develop a model to optimize sales with the use of digital marketing in a Peruvian company, the chosen methodology is DesingScrum, which is a hybrid of Scrum and Desing Thinking, with 10 phases (empathize, define. It has 10 phases (empathise, define, ideate, planning meeting, sprint backlog, daily meeting, sprint review, sprint retrospective sprint, prototype and testing) and the MarvelApp tool was used to create the prototype. The results are obtained after the completion of the review of each sprint, showing in detail how it was progressing in each sprint, and through the retrospective evaluated the development of the project for the realization of continuous improvement in the next product. Further prototype was made, with the application MarvelApp, which shows the model of e-business. Then testing was done through a survey that customers gave their opinions about the prototype and finally the digital marketing proposal was made by a model, which explains the interconnected tools to attract new customers. The conclusion is the construction of the digital marketing model according to the needs of the context to improve the sales of the company through e-business.

Author 1: Misael Lazo-Amado
Author 2: Leoncio Cueva-Ruiz
Author 3: Laberiano Andrade-Arenas

Keywords: Design thinking; desingScrum methodology; digital marketing; e-Business; scrum

PDF

Paper 85: A Fast and Efficient Algorithm for Outlier Detection Over Data Streams

Abstract: Outlier detection over data streams is an important task in data mining. It has various applications such as fraud detection, public health, and computer network security. Many approaches have been proposed for outlier detection over data streams such as distance-,clustering-, density-, and learning-based approaches. In this paper, we are interested in the density-based outlier detection over data streams. Specifically, we propose an improvement of DILOF, a recent density-based algorithm. We observed that the main disadvantage of DILOF is that its summarization method has many drawbacks such as it takes a lot of time and the algorithm accuracy is significant degradation. Our new algorithm is called DILOFC that utilizing an efficient summarization method. Our performance study shows that DILOF􀀀 outperforms DILOF in terms of total response time and outlier detection accuracy.

Author 1: Mosab Hassaan
Author 2: Hend Maher
Author 3: Karam Gouda

Keywords: Data mining; outlier detection; data streams; density-based approach; clustering-based approach

PDF

Paper 87: Near-ground Measurement and Modeling for Archaeological Park of Pisac in Cusco for LoRa Technology

Abstract: This research work presents the details of a near-ground received power measurement and modeling based on it, inside the archaeological park of Pisac in the city of Cusco. The measurements were performed at a working frequency of 920 MHz in the industrial, scientific, and medical (ISM) band, using transceiver devices with LoRa technology. The power of the received signal is obtained while the transmitter moves at a constant speed of 0.4 m/s, to characterize the fading that occurred within this archaeological park, making appropriate use of the moving average filter separates the fading to large and small scale, then on the filtered signal is used the algorithm of linear regression, to obtain a model that characterizes the exponent of propagation loss and the shading factor. In addition, the small-scale fading is characterized according to its proba-bility distribution, the statistical parameters of the distributions are obtained based on the small-scale measurements. We also measured variations of receiving antenna height, ranging from 20cm, 50cm, 80cm, and 120cm above the ground, finding that the height has a strong influence on the propagation loss exponent, while the shadow variation shows a smaller variation. The model obtained is validated by the coefficient of determination and the root mean square error (RMSE) value.

Author 1: Yhon D. Lezama
Author 2: Jinmi Lezama
Author 3: Jorge Arizaca-Cusicuna

Keywords: Archaeological park; LoRa; propagation model; statistics; near-ground

PDF

Paper 88: Development of Predictions through Machine Learning for Sars-Cov-2 Forecasting in Peru

Abstract: The SARS-COV-2 virus of the coronavirus family was identified in 2019. This is a type of virus that infects humans and some animals, in Peru it has seriously affected everyone, causing so many deaths, which has resulted in that people be tested to rule out contagion, using laboratory methods recommended by the government of the country. Therefore, the data science methodology was used with this research, where its objective is to predict what types of people are contaminated during SARS-COV-2 by the regions of Peru, identified through laboratory methods, therefore, the ”data bank” was taken by PNDA, the CSV file was used for that study, apart from the fact that it comes from the INS and the CDC of the MINSA. In which, machine learning was developed with the decision tree algorithm and then began coding, in such a way that the distribution called Anaconda was used where it is encoded in Python language, together with that distribution, Jupyter Notebook was used which is a client-server application. The results generated by this research prove that it was possible to identify the types of individuals by SARS-COV-2. These results can help prevention entities against SARS-COV-2 to apply the corresponding preventive measures in a more focused way.

Author 1: Shal´om Adonai Huaraz Morales
Author 2: Marissel Fabiola Mio Antayhua
Author 3: Laberiano Andrade-Arenas

Keywords: Forecast; laboratory methods; machine learning; Python; SARS-COV-2

PDF

Paper 89: Inclusive Education: Implementation of a Mobile Application for Blind Students

Abstract: Currently, the world is going through an era of changes in the education sector, but most Latin American countries are lagging, especially Peru, which does not have the technological tools that allow it to advance to an adequate level of inclusion of disabled students, especially blind students who are 60% of students who drop out of school for lack of education that have be suitable for their need. Consequently, the present research work is originated which aims to develop a mobile application oriented to the benefit of inclusive education of blind students. Therefore, the agile scrum methodology was used for the development of this project, executed in 5 phases, the requirements identified for the development of the mobile application were obtained through a questionnaire to 25 parents of visually impaired students, allowing the development of a mobile application that meets quality of inclusive education that can be applied in the education sector. Finally, as a result of the research work, another satisfaction survey was conducted with 50 parents where the application was evaluated, obtaining 90%of acceptance and satisfaction.

Author 1: Alejandro Boza-Chua
Author 2: Karen Gabriel-Gonzales
Author 3: Laberiano Andrade-Arenas

Keywords: Blind; disability; educational inclusion; mobile ap-plication; scrum methodology

PDF

Paper 90: Design and Implementation of HSQL: A SQL-like language for Data Analysis in Distributed Systems

Abstract: In today’s modern world, we’re experiencing a substantial increase in the use of data in various fields, and this has necessitated the use of distributed systems to consume and process Big Data. Machine learning tends to benefit from the usage of Big Data, and the models generated from such techniques tend to be more effective. However, there is a steep learning curve to getting used to handling Big Data, as traditional data management tools fail to perform well. Distributed systems have become popular, where the task of data processing is split amongst various nodes in clusters. SQL, is a popular database management language popular to data scientists. It is often given second class support, where SQL can be embedded into a primary language of use (e.g. SQL in Scala for Spark), which allows for using SQL but one still needs to know the primary language of the platform (Scala, as per the example, or ECL in HPCC Systems). It may also be present as a supported language. In either case, using useful tooling such as Visualizing data and creating and using machine learning models become difficult, as the user needs to fall back to the primary language of the system. In the proposed work, a new SQL-like language, HSQL, an open source distributed systems solution, was developed for allowing new users to get used to its distributed architecture and the ECL language, with which it primarily operates with (which was chosen as a target). Additionally, a program that could translate HSQL-based programs to ECL for use was made. HSQL was made to be completely inter-compatible with ECL programs, and it was able to provide a compact and easy to comprehend SQL-like syntax for performing general data analysis, creation of Machine learning models and visualizations while allowing a modular structure to such programs.

Author 1: Anurag Singh Bhadauria
Author 2: Atreya Bain
Author 3: Jyoti Shetty
Author 4: Shobha G
Author 5: Arjuna Chala
Author 6: Jeremy Clements

Keywords: ANTLR4; big data; context free grammar; dis-tributed systems; HPCC; Javascript; language; machine learning; Parser; SQL; transpiler

PDF

Paper 91: Factors Influencing the Adoption of Cyber Security Standards Among Public Listed Companies in Malaysia

Abstract: Employee’s failure to adhere to their organization’s cyber security policies contributes most of the cyber incidents. To secure information security systems, companies need to communicate behavioral and technical solutions to their employees, due to the fragility of the human factor since it plays a critically significant role in securing cyber systems. The necessity to safeguard information systems have speed up the evolution of the present method of cyber security, which should be based on adequately adopting cyber security standards to secure business enterprise’s assets and users in cyberspace. This paper studies factors influencing the adoption of cyber security standards among public listed companies in Malaysia. Through online survey that was distributed among 275 Public listed companies. The findings indicated that expected related benefits and perceived ease of use had significant impact on the adoption of cyber security standards. On the other hand, perceived security had played important moderating influence on the relationship between organizational factors and the adoption of cyber security standards.

Author 1: Mohamed Abdalla
Author 2: Muath Jarrah
Author 3: Ahmed Abu-Khadrah
Author 4: Yusri bin Arshad

Keywords: Cyber security; human factor; cyberspace; cyber security standards

PDF

Paper 92: Novel Algorithm Utilizing Deep Learning for Enhanced Arabic Lip Reading Recognition

Abstract: Computerized lip reading is the science of translating visemes and oral lip reading into written text, where visemes are lip movement without sound. Video processing is applied for the recognition of those visemes. Previous research developed automated systems to for computerized lip reading recognition to be hearing-impaired aid. Many challenges face such an automation process, including insufficient training datasets. Also, speaker-dependency is one of the challenges that are faced. Real-time applications which respond within a specific time period are also widely required. Real-time human computer interaction are systems which require real time response. Response time for human computer interaction is measured by number of elapsed video frames. Video processing of lip reading necessitates real-time implementation. There are applications for viseme recognition, as an aid for deaf people, video games with human computer interaction, and surveillance systems. In this paper, a real-time viseme recognition system is introduced. In order to enhance existing methods to overcome these pitfalls, this paper proposed a computerized lip reading technique based on feature extraction. We utilized blocks arrangement techniques to reach a near-optimal appearance feature extraction technique. A Deep Neural Network is utilized to enhance recognition. The benchmark dataset SAVE, for Arabic visemes, is employed in this research, and high viseme recognition accuracies are achieved. The described computerized lip reading recognition technique is advantageous for the hearing impaired and for other speakers in noisy environments.

Author 1: Doaa Sami Khafaga
Author 2: Hanan A. Hosni Mahmoud
Author 3: Norah S. Alghamdi
Author 4: Amani A. Albraikan

Keywords: Lip reading; deep CNN; deep learning; recognition; viseme

PDF

The Science and Information (SAI) Organization
BACK TO TOP

Computer Science Journal

  • About the Journal
  • Call for Papers
  • Submit Paper
  • Indexing

Our Conferences

  • Computing Conference
  • Intelligent Systems Conference
  • Future Technologies Conference
  • Communication Conference

Help & Support

  • Contact Us
  • About Us
  • Terms and Conditions
  • Privacy Policy

© The Science and Information (SAI) Organization Limited. All rights reserved. Registered in England and Wales. Company Number 8933205. thesai.org