Z. Hassani; M. Alambardar Meybodi
Abstract
A major pitfall in the standard version of Particle Swarm Optimization (PSO) is that it might get stuck in the local optima. To escape this issue, a novel hybrid model based on the combination of PSO and AntLion Optimization (ALO) is proposed in this study. The proposed method, called H-PSO-ALO, uses ...
Read More
A major pitfall in the standard version of Particle Swarm Optimization (PSO) is that it might get stuck in the local optima. To escape this issue, a novel hybrid model based on the combination of PSO and AntLion Optimization (ALO) is proposed in this study. The proposed method, called H-PSO-ALO, uses a local search strategy by employing the Ant-Lion algorithm to select the less correlated and salient feature subset. The objective is to improve the prediction accuracy and adaptability of the model in various datasets by balancing the exploration and exploitation processes. The performance of our method has been evaluated on 30 benchmark classification problems, CEC 2017 benchmark problems, and some well-known datasets. To verify the performance, four algorithms, including FDR-PSO, CLPSO, HFPSO, MPSO, are elected to be compared with the efficiency of H-PSO-ALO. Considering the experimental results, the proposed method outperforms the others in many cases, so it seems it is a desirable candidate for optimization problems on real-world datasets.
H.3.2.6. Games and infotainment
Shaqayeq Saffari; Morteza Dorrigiv; Farzin Yaghmaee
Abstract
Procedural Content Generation (PCG) through automated and algorithmic content generation is an active research field in the gaming industry. Recently, Machine Learning (ML) approaches have played a pivotal role in advancing this area. While recent studies have primarily focused on examining one or a ...
Read More
Procedural Content Generation (PCG) through automated and algorithmic content generation is an active research field in the gaming industry. Recently, Machine Learning (ML) approaches have played a pivotal role in advancing this area. While recent studies have primarily focused on examining one or a few specific approaches in PCG, this paper provides a more comprehensive perspective by exploring a wider range of approaches, their applications, advantages, and disadvantages. Furthermore, the current challenges and potential future trends in this field are discussed. Although this paper does not aim to provide an exhaustive review of all existing research due to the rapid and expansive growth of this domain, it is based on the analysis of selected articles published between 2020 and 2024.
H.3. Artificial Intelligence
Mahdi Rasouli; Vahid Kiani
Abstract
The identification of emotions in short texts of low-resource languages poses a significant challenge, requiring specialized frameworks and computational intelligence techniques. This paper presents a comprehensive exploration of shallow and deep learning methods for emotion detection in short Persian ...
Read More
The identification of emotions in short texts of low-resource languages poses a significant challenge, requiring specialized frameworks and computational intelligence techniques. This paper presents a comprehensive exploration of shallow and deep learning methods for emotion detection in short Persian texts. Shallow learning methods employ feature extraction and dimension reduction to enhance classification accuracy. On the other hand, deep learning methods utilize transfer learning and word embedding, particularly BERT, to achieve high classification accuracy. A Persian dataset called "ShortPersianEmo" is introduced to evaluate the proposed methods, comprising 5472 diverse short Persian texts labeled in five main emotion classes. The evaluation results demonstrate that transfer learning and BERT-based text embedding perform better in accurately classifying short Persian texts than alternative approaches. The dataset of this study ShortPersianEmo will be publicly available online at https://github.com/vkiani/ShortPersianEmo.
G.3.7. Database Machines
Abdul Aziz Danaa Abukari; Mohammed Daabo Ibrahim; Alhassan Abdul-Barik
Abstract
Hidden Markov Models (HMMs) are machine learning models that has been applied to a range of real-life applications including intrusion detection, pattern recognition, thermodynamics, statistical mechanics among others. A multi-layered HMMs for real-time fraud detection and prevention whilst reducing ...
Read More
Hidden Markov Models (HMMs) are machine learning models that has been applied to a range of real-life applications including intrusion detection, pattern recognition, thermodynamics, statistical mechanics among others. A multi-layered HMMs for real-time fraud detection and prevention whilst reducing drastically the number of false positives and negatives is proposed and implemented in this study. The study also focused on reducing the parameter optimization and detection times of the proposed models using a hybrid algorithm comprising the Baum-Welch, Genetic and Particle-Swarm Optimization algorithms. Simulation results revealed that, in terms of Precision, Recall and F1-scores, our proposed model performed better when compared to other approaches proposed in literature.
B.3. Communication/Networking and Information Technology
S. Mojtaba Matinkhah; Roya Morshedi; Akbar Mostafavi
Abstract
The Internet of Things (IoT) has emerged as a rapidly growing technology that enables seamless connectivity between a wide variety of devices. However, with this increased connectivity comes an increased risk of cyber-attacks. In recent years, the development of intrusion detection systems (IDS) has ...
Read More
The Internet of Things (IoT) has emerged as a rapidly growing technology that enables seamless connectivity between a wide variety of devices. However, with this increased connectivity comes an increased risk of cyber-attacks. In recent years, the development of intrusion detection systems (IDS) has become critical for ensuring the security and privacy of IoT networks. This article presents a study that evaluates the accuracy of an intrusion detection system (IDS) for detecting network attacks in the Internet of Things (IoT) network. The proposed IDS uses the Decision Tree Classifier and is tested on four benchmark datasets: NSL-KDD, BOT-IoT, CICIDS2017, and MQTT-IoT. The impact of noise on the training and test datasets on classification accuracy is analyzed. The results indicate that clean data has the highest accuracy, while noisy datasets significantly reduce accuracy. Furthermore, the study finds that when both training and test datasets are noisy, the accuracy of classification decreases further. The findings of this study demonstrate the importance of using clean data for training and testing an IDS in IoT networks to achieve accurate classification. This research provides valuable insights for the development of a robust and accurate IDS for IoT networks.
H.3. Artificial Intelligence
Seyed Alireza Bashiri Mosavi; Omid Khalaf Beigi
Abstract
A speedy and accurate transient stability assessment (TSA) is gained by employing efficient machine learning- and statistics-based (MLST) algorithms on transient nonlinear time series space. In the MLST’s world, the feature selection process by forming compacted optimal transient feature space ...
Read More
A speedy and accurate transient stability assessment (TSA) is gained by employing efficient machine learning- and statistics-based (MLST) algorithms on transient nonlinear time series space. In the MLST’s world, the feature selection process by forming compacted optimal transient feature space (COTFS) from raw high dimensional transient data can pave the way for high-performance TSA. Hence, designing a comprehensive feature selection scheme (FSS) that populates COTFS with the relevant-discriminative transient features (RDTFs) is an urgent need. This work aims to introduce twin hybrid FSS (THFSS) to select RDTFs from transient 28-variate time series data. Each fold of THFSS comprises filter-wrapper mechanisms. The conditional relevancy rate (CRR) is based on mutual information (MI) and entropy calculations are considered as the filter method, and incremental wrapper subset selection (IWSS) and IWSS with replacement (IWSSr) formed by kernelized support vector machine (SVM) and twin SVM (TWSVM) are used as wrapper ones. After exerting THFSS on transient univariates, RDTFs are entered into the cross-validation-based train-test procedure for evaluating their efficiency in TSA. The results manifested that THFSS-based RDTFs have a prediction accuracy of 98.87 % and a processing time of 102.653 milliseconds for TSA.
H.6. Pattern Recognition
Sadegh Rahmani Rahmani-Boldaji; Mehdi Bateni; Mahmood Mortazavi Dehkordi
Abstract
Efficient regular-frequent pattern mining from sensors-produced data has become a challenge. The large volume of data leads to prolonged runtime, thus delaying vital predictions and decision makings which need an immediate response. So, using big data platforms and parallel algorithms is an appropriate ...
Read More
Efficient regular-frequent pattern mining from sensors-produced data has become a challenge. The large volume of data leads to prolonged runtime, thus delaying vital predictions and decision makings which need an immediate response. So, using big data platforms and parallel algorithms is an appropriate solution. Additionally, an incremental technique is more suitable to mine patterns from big data streams than static methods. This study presents an incremental parallel approach and compact tree structure for extracting regular-frequent patterns from the data of wireless sensor networks. Furthermore, fewer database scans have been performed in an effort to reduce the mining runtime. This study was performed on Intel 5-day and 10-day datasets with 6, 4, and 2 nodes clusters. The findings show the runtime was improved in all 3 cluster modes by 14, 18, and 34% for the 5-day dataset and by 22, 55, and 85% for the 10-day dataset, respectively.
R. Azizi; A. M. Latif
Abstract
In this work, we show that an image reconstruction from a burst of individually demosaicked RAW captures propagates demosaicking artifacts throughout the image processing pipeline. Hence, we propose a joint regularization scheme for burst denoising and demosaicking. We model the burst alignment functions ...
Read More
In this work, we show that an image reconstruction from a burst of individually demosaicked RAW captures propagates demosaicking artifacts throughout the image processing pipeline. Hence, we propose a joint regularization scheme for burst denoising and demosaicking. We model the burst alignment functions and the color filter array sampling functions into one linear operator. Then, we formulate the individual burst reconstruction and the demosaicking problems into a three-color-channel optimization problem. We introduce a crosschannel prior to the solution of this optimization problem and develop a numerical solver via alternating direction method of multipliers. Moreover, our proposed method avoids the complexity of alignment estimation as a preprocessing step for burst reconstruction. It relies on a phase correlation approach in the Fourier’s domain to efficiently find the relative translation, rotation, and scale among the burst captures and to perform warping accordingly. As a result of these steps, the proposed joint burst denoising and demosaicking solution improves the quality of reconstructed images by a considerable margin compared to existing image model-based methods.