F. Baratzadeh; Seyed M. H. Hasheminejad
Abstract
With the advancement of technology, the daily use of bank credit cards has been increasing exponentially. Therefore, the fraudulent use of credit cards by others as one of the new crimes is also growing fast. For this reason, detecting and preventing these attacks has become an active area of study. ...
Read More
With the advancement of technology, the daily use of bank credit cards has been increasing exponentially. Therefore, the fraudulent use of credit cards by others as one of the new crimes is also growing fast. For this reason, detecting and preventing these attacks has become an active area of study. This article discusses the challenges of detecting fraudulent banking transactions and presents solutions based on deep learning. Transactions are examined and compared with other traditional models in fraud detection. According to the results obtained, optimal performance is related to the combined model of deep convolutional networks and short-term memory, which is trained using the aggregated data received from the generative adversarial network. This paper intends to produce sensible data to address the unequal class distribution problem, which is far more effective than traditional methods. Also, it uses the strengths of the two approaches by combining deep convolutional network and Long Short Term Memory network to improve performance. Due to the inefficiency of evaluation criteria such as accuracy in this application, the measure of distance score and the equal error rate has been used to evaluate models more transparent and more precise. Traditional methods were compared to the proposed approach to evaluate the efficiency of the experiment.
M. M. Jaziriyan; F. Ghaderi
Abstract
Most of the existing neural machine translation (NMT) methods translate sentences without considering the context. It is shown that exploiting inter and intra-sentential context can improve the NMT models and yield to better overall translation quality. However, providing document-level data is costly, ...
Read More
Most of the existing neural machine translation (NMT) methods translate sentences without considering the context. It is shown that exploiting inter and intra-sentential context can improve the NMT models and yield to better overall translation quality. However, providing document-level data is costly, so properly exploiting contextual data from monolingual corpora would help translation quality. In this paper, we proposed a new method for context-aware neural machine translation (CA-NMT) using a combination of hierarchical attention networks (HAN) and automatic post-editing (APE) techniques to fix discourse phenomena when there is lack of context. HAN is used when we have a few document-level data, and APE can be trained on vast monolingual document-level data to improve results further. Experimental results show that combining HAN and APE can complement each other to mitigate contextual translation errors and further improve CA-NMT by achieving reasonable improvement over HAN (i.e., BLEU score of 22.91 on En-De news-commentary dataset).
H.3.15.3. Evolutionary computing and genetic algorithms
Mahdieh Maazalahi; Soodeh Hosseini
Abstract
Detecting and preventing malware infections in systems is become a critical necessity. This paper presents a hybrid method for malware detection, utilizing data mining algorithms such as simulated annealing (SA), support vector machine (SVM), genetic algorithm (GA), and K-means. The proposed method combines ...
Read More
Detecting and preventing malware infections in systems is become a critical necessity. This paper presents a hybrid method for malware detection, utilizing data mining algorithms such as simulated annealing (SA), support vector machine (SVM), genetic algorithm (GA), and K-means. The proposed method combines these algorithms to achieve effective malware detection. Initially, the SA-SVM method is employed for feature selection, where the SVM algorithm identifies the best features, and the SA algorithm calculates the SVM parameters. Subsequently, the GA-K-means method is utilized to identify attacks. The GA algorithm selects the best chromosome for cluster centers, and the K-means algorithm has applied to identify malware. To evaluate the performance of the proposed method, two datasets, Andro-Autopsy and CICMalDroid 2020, have been utilized. The evaluation results demonstrate that the proposed method achieves high true positive rates (0.964, 0.985), true negative rates (0.985, 0.989), low false negative rates (0.036, 0.015), and false positive rates (0.022, 0.043). This indicates that the method effectively detects malware while reasonably minimizing false identifications.
H.3.2.2. Computer vision
Razieh Rastgoo
Abstract
Sign language (SL) is the primary mode of communication within the Deaf community. Recent advances in deep learning have led to the development of various applications and technologies aimed at facilitating bidirectional communication between the Deaf and hearing communities. However, challenges remain ...
Read More
Sign language (SL) is the primary mode of communication within the Deaf community. Recent advances in deep learning have led to the development of various applications and technologies aimed at facilitating bidirectional communication between the Deaf and hearing communities. However, challenges remain in the availability of suitable datasets for deep learning-based models. Only a few public large-scale annotated datasets are available for sign sentences, and none exist for Persian Sign Language sentences. To address this gap, we have collected a large-scale dataset comprising 10,000 sign sentence videos corresponding to 100 Persian sign sentences. This dataset includes comprehensive annotations such as the bounding box of the detected hand, class labels, hand pose parameters, and heatmaps. A notable feature of the proposed dataset is that it contains isolated signs corresponding to the sign sentences within the dataset. To analyze the complexity of the proposed dataset, we present extensive experiments and discuss the results. More concretely, the results of the models in key sub-domains relevant to Sign Language Recognition (SLR), including hand detection, pose estimation, real-time tracking, and gesture recognition, have been included and analyzed. Moreover, the results of seven deep learning-based models on the proposed datasets have been discussed. Finally, the results of Sign Language Production (SLP) using deep generative models have been presented. We report the experimental results of these models from these sub-areas, showcasing their performance on the proposed dataset.
A. Omondi; I. Lukandu; G. Wanyembi
Abstract
Variable environmental conditions and runtime phenomena require developers of complex business information systems to expose configuration parameters to system administrators. This allows system administrators to intervene by tuning the bottleneck configuration parameters in response to current changes ...
Read More
Variable environmental conditions and runtime phenomena require developers of complex business information systems to expose configuration parameters to system administrators. This allows system administrators to intervene by tuning the bottleneck configuration parameters in response to current changes or in anticipation of future changes in order to maintain the system’s performance at an optimum level. However, these manual performance tuning interventions are prone to error and lack of standards due to fatigue, varying levels of expertise and over-reliance on inaccurate predictions of future states of a business information system. As a result, the purpose of this research is to investigate on how the capacity of probabilistic reasoning to handle uncertainty can be combined with the capacity of Markov chains to map stochastic environmental phenomena to ideal self-optimization actions. This was done using a comparative experimental research design that involved quantitative data collection through simulations of different algorithm variants. This provided compelling results that indicate that applying the algorithm in a distributed database system improves performance of tuning decisions under uncertainty. The improvement was quantitatively measured by a response-time latency that was 27% lower than average and a transaction throughput that was 17% higher than average.
H.R. Koosha; Z. Ghorbani; R. Nikfetrat
Abstract
In the last decade, online shopping has played a vital role in customers' approach to purchasing different products, providing convenience to shop and many benefits for the economy. E-commerce is widely used for digital media products such as movies, images, and software. So, recommendation systems are ...
Read More
In the last decade, online shopping has played a vital role in customers' approach to purchasing different products, providing convenience to shop and many benefits for the economy. E-commerce is widely used for digital media products such as movies, images, and software. So, recommendation systems are of great importance, especially in today's hectic world, which search for content that would be interesting to an individual. In this research, a new two-steps recommender system is proposed based on demographic data and user ratings on the public MovieLens datasets. In the first step, clustering on the training dataset is performed based on demographic data, grouping customers in homogeneous clusters. The clustering includes a hybrid Firefly Algorithm (FA) and K-means approach. Due to the FA's ability to avoid trapping into local optima, which resolves K-means' main pitfall, the combination of these two techniques leads to much better performance. In the next step, for each cluster, two recommender systems are proposed based on K-Nearest Neighbor (KNN) and Naïve Bayesian Classification. The results are evaluated based on many internal and external measures like the Davies-Bouldin index, precision, accuracy, recall, and F-measure. The results showed the effectiveness of the K-means/FA/KNN compared with other extant models.
H.3. Artificial Intelligence
M. Taghian; A. Asadi; R. Safabakhsh
Abstract
The quality of the extracted features from a long-term sequence of raw prices of the instruments greatly affects the performance of the trading rules learned by machine learning models. Employing a neural encoder-decoder structure to extract informative features from complex input time-series has proved ...
Read More
The quality of the extracted features from a long-term sequence of raw prices of the instruments greatly affects the performance of the trading rules learned by machine learning models. Employing a neural encoder-decoder structure to extract informative features from complex input time-series has proved very effective in other popular tasks like neural machine translation and video captioning. In this paper, a novel end-to-end model based on the neural encoder-decoder framework combined with deep reinforcement learning is proposed to learn single instrument trading strategies from a long sequence of raw prices of the instrument. In addition, the effects of different structures for the encoder and various forms of the input sequences on the performance of the learned strategies are investigated. Experimental results showed that the proposed model outperforms other state-of-the-art models in highly dynamic environments.
H.5. Image Processing and Computer Vision
Sekine Asadi Amiri; Mahda Nasrolahzadeh; Zeynab Mohammadpoory; AbdolAli Movahedinia; Amirhossein Zare
Abstract
Improving the quality of food industries and the safety and health of the people’s nutrition system is one of the important goals of governments. Fish is an excellent source of protein. Freshness is one of the most important quality criteria for fish that should be selected for consumption. It ...
Read More
Improving the quality of food industries and the safety and health of the people’s nutrition system is one of the important goals of governments. Fish is an excellent source of protein. Freshness is one of the most important quality criteria for fish that should be selected for consumption. It has been shown that due to improper storage conditions of fish, bacteria, and toxins may cause diseases for human health. The conventional methods of detecting spoilage and disease in fish, i.e. analyzing fish samples in the laboratory, are laborious and time-consuming. In this paper, an automatic method for identifying spoiled fish from fresh fish is proposed. In the proposed method, images of fish eyes are used. Fresh fish are identified by shiny eyes, and poor and stale fish are identified by gray color changes in the eye. In the proposed method, Inception-ResNet-v2 convolutional neural network is used to extract features. To increase the accuracy of the model and prevent overfitting, only some useful features are selected using the mRMR feature selection method. The mRMR reduces the dimensionality of the data and improves the classification accuracy. Then, since the number of samples is low, the k-fold cross-validation method is used. Finally, for classifying the samples, Naïve bayes and Random forest classifiers are used. The proposed method has reached an accuracy of 97% on the fish eye dataset, which is better than previous references.
H.3.8. Natural Language Processing
Ali Reza Ghasemi; Javad Salimi Sartakhti
Abstract
This paper evaluates the performance of various fine-tuning methods in Persian natural language processing (NLP) tasks. In low-resource languages like Persian, which suffer from a lack of rich and sufficient data for training large models, it is crucial to select appropriate fine-tuning ...
Read More
This paper evaluates the performance of various fine-tuning methods in Persian natural language processing (NLP) tasks. In low-resource languages like Persian, which suffer from a lack of rich and sufficient data for training large models, it is crucial to select appropriate fine-tuning techniques that mitigate overfitting and prevent the model from learning weak or surface-level patterns. The main goal of this research is to compare the effectiveness of fine-tuning approaches such as Full-Finetune, LoRA, AdaLoRA, and DoRA on model learning and task performance. We apply these techniques to three different Persian NLP tasks: sentiment analysis, named entity recognition (NER), and span question answering (QA). For this purpose, we conduct experiments on three Transformer-based multilingual models with different architectures and parameter scales: BERT-base multilingual (~168M parameters) with Encoder only structure, mT5-small (~300M parameters) with Encoder-Decoder structure, and mGPT (~1.4B parameters) with Decoder only structure. Each of these models supports the Persian language but varies in structure and computational requirements, influencing the effectiveness of different fine-tuning approaches. Results indicate that fully fine-tuned BERT-base multilingual consistently outperforms other models across all tasks in basic metrics, particularly given the unique challenges of these embedding-based tasks. Additionally, lightweight fine-tuning methods like LoRA and DoRA offer very competitive performance while significantly reducing computational overhead and outperform other models in Performance-Efficiency Score introduced in the paper. This study contributes to a better understanding of fine-tuning methods, especially for Persian NLP, and offers practical guidance for applying Large Language Models (LLMs) to downstream tasks in low-resource languages.
M. Zarbazoo Siahkali; A.A. Ghaderi; Abdol H. Bahrpeyma; M. Rashki; N. Safaeian Hamzehkolaei
Abstract
Scouring, occurring when the water flow erodes the bed materials around the bridge pier structure, is a serious safety assessment problem for which there are many equations and models in the literature to estimate the approximate scour depth. This research is aimed to study how surrogate models estimate ...
Read More
Scouring, occurring when the water flow erodes the bed materials around the bridge pier structure, is a serious safety assessment problem for which there are many equations and models in the literature to estimate the approximate scour depth. This research is aimed to study how surrogate models estimate the scour depth around circular piers and compare the results with those of the empirical formulations. To this end, the pier scour depth was estimated in non-cohesive soils based on a subcritical flow and live bed conditions using the artificial neural networks (ANN), group method of data handling (GMDH), multivariate adaptive regression splines (MARS) and Gaussian process models (Kriging). A database containing 246 lab data gathered from various studies was formed and the data were divided into three random parts: 1) training, 2) validation and 3) testing to build the surrogate models. The statistical error criteria such as the coefficient of determination (R2), root mean squared error (RMSE), mean absolute percentage error (MAPE) and absolute maximum percentage error (MPE) of the surrogate models were then found and compared with those of the popular empirical formulations. Results revealed that the surrogate models’ test data estimations were more accurate than those of the empirical equations; Kriging has had better estimations than other models. In addition, sensitivity analyses of all surrogate models showed that the pier width’s dimensionless expression (b/y) had a greater effect on estimating the normalized scour depth (Ds/y).
H.3.7. Learning
Mohammad Rezaei; Mohsen Rezvani; Morteza Zahedi
Abstract
With the increasing interconnectedness of communications and social networks, graph-based learning techniques offer valuable information extraction from data. Traditional centralized learning methods faced challenges, including data privacy violations and costly maintenance in a centralized environment. ...
Read More
With the increasing interconnectedness of communications and social networks, graph-based learning techniques offer valuable information extraction from data. Traditional centralized learning methods faced challenges, including data privacy violations and costly maintenance in a centralized environment. To address these, decentralized learning approaches like Federated Learning have emerged. This study explores the significant attention Federated Learning has gained in graph classification and investigates how Model Agnostic Meta-Learning (MAML) can improve its performance, especially concerning non-IID (Non-Independent Identically Distributed) data distributions.In real-world scenarios, deploying Federated Learning poses challenges, particularly in tuning client parameters and structures due to data isolation and diversity. To address this issue, this study proposes an innovative approach using Genetic Algorithms (GA) for automatic tuning of structures and parameters. By integrating GA with MAML-based clients in Federated Learning, various aspects, such as graph classification structure, learning rate, and optimization function type, can be automatically adjusted. This novel approach yields improved accuracy in decentralized learning at both the client and server levels.
M. Khanzadi; H. Veisi; R. Alinaghizade; Z. Soleymani
Abstract
One of the main problems in children with learning difficulties is the weakness of phonological awareness (PA) skills. In this regard, PA tests are used to evaluate this skill. Currently, this assessment is paper-based for the Persian language. To accelerate the process of the assessments and make it ...
Read More
One of the main problems in children with learning difficulties is the weakness of phonological awareness (PA) skills. In this regard, PA tests are used to evaluate this skill. Currently, this assessment is paper-based for the Persian language. To accelerate the process of the assessments and make it engaging for children, we propose a computer-based solution that is a comprehensive Persian phonological awareness assessment system implementing expressive and pointing tasks. For the expressive tasks, the solution is powered by recurrent neural network-based speech recognition systems. To this end, various recognition modules are implemented, including a phoneme recognition system for the phoneme segmentation task, a syllable recognition system for the syllable segmentation task, and a sub-word recognition system for three types of phoneme deletion tasks, including initial, middle, and final phoneme deletion. The recognition systems use bidirectional long short-term memory neural networks to construct acoustic models. To implement the recognition systems, we designed and collected Persian Kid’s Speech Corpus that is the largest in Persian for children’s speech. The accuracy rate for phoneme recognition was 85.5%, and for syllable recognition was 89.4%. The accuracy rates of the initial, middle, and final phoneme deletion were 96.76%, 98.21%, and 95.9%, respectively.
P. Abdzadeh; H. Veisi
Abstract
Automatic Speaker Verification (ASV) systems have proven to bevulnerable to various types of presentation attacks, among whichLogical Access attacks are manufactured using voiceconversion and text-to-speech methods. In recent years, there has beenloads of work concentrating on synthetic speech detection, ...
Read More
Automatic Speaker Verification (ASV) systems have proven to bevulnerable to various types of presentation attacks, among whichLogical Access attacks are manufactured using voiceconversion and text-to-speech methods. In recent years, there has beenloads of work concentrating on synthetic speech detection, and with the arrival of deep learning-based methods and their success in various computer science fields, they have been a prevailing tool for this very task too. Most of the deep neural network-based techniques forsynthetic speech detection have employed the acoustic features basedon Short-Term Fourier Transform (STFT), which are extracted from theraw audio signal. However, lately, it has been discovered that the usageof Constant Q Transform's (CQT) spectrogram can be a beneficialasset both for performance improvement and processing power andtime reduction of a deep learning-based synthetic speech detection. In this work, we compare the usage of the CQT spectrogram and some most utilized STFT-based acoustic features. As lateral objectives, we consider improving the model's performance as much as we can using methods such as self-attention and one-class learning. Also, short-duration synthetic speech detection has been one of the lateral goals too. Finally, we see that the CQT spectrogram-based model not only outperforms the STFT-based acoustic feature extraction methods but also reduces the processing time and resources for detecting genuine speech from fake. Also, the CQT spectrogram-based model places wellamong the best works done on the LA subset of the ASVspoof 2019 dataset, especially in terms of Equal Error Rate.
H.5. Image Processing and Computer Vision
Khosro Rezaee
Abstract
Bipolar disorder (BD) remains a pervasive mental health challenge, demanding innovative diagnostic approaches beyond traditional, subjective assessments. This study pioneers a non-invasive handwriting-based diagnostic framework, leveraging the unique interplay between psychological states and motor expressions ...
Read More
Bipolar disorder (BD) remains a pervasive mental health challenge, demanding innovative diagnostic approaches beyond traditional, subjective assessments. This study pioneers a non-invasive handwriting-based diagnostic framework, leveraging the unique interplay between psychological states and motor expressions in writing. Our hybrid deep learning model, combining ResNet for intricate feature extraction and external attention mechanisms for global pattern analysis, achieves a remarkably high accuracy 99%, validated through Leave-One-Subject-Out (LOSO) cross-validation. Augmented with advanced data preprocessing and augmentation techniques, the framework adeptly addresses dataset imbalances and handwriting variability. For the first time, Persian handwriting serves as a medium, bridging cultural gaps in BD diagnostics. This work not only establishes handwriting as a transformative tool for mental health diagnostics but also sets the stage for accessible, scalable, and culturally adaptive solutions in global mental healthcare.
R. Ghotboddini; H. Toossian Shandiz
Abstract
Lighting continuity is one of the preferences of citizens. Public lighting management from the viewpoint of city residents improves social welfare. The quality of lamps and duration of lighting defect correction is important in lighting continuity. In this regard, reward and penalty mechanism plays an ...
Read More
Lighting continuity is one of the preferences of citizens. Public lighting management from the viewpoint of city residents improves social welfare. The quality of lamps and duration of lighting defect correction is important in lighting continuity. In this regard, reward and penalty mechanism plays an important role in contract. Selecting labor and lamps has a significant impact on risk reduction during the contract period. This research improves strategies for public lighting asset management. The lifespan of lamp that announced by manufacturers is used to calculate maintenance cost in order to provide a possibility to estimate the actual cost of high-pressure sodium luminaire in public lighting system. Guarantee period of lamps and maximum permissible lighting defect detection and correction time is used for reward and penalty mechanism. The result shows that the natural life guarantee and permissible correction time have a considerable effect in maintenance cost and city resident’s satisfaction.
H.5. Image Processing and Computer Vision
Pouria Maleki; Abbas Ramazani; Hassan Khotanlou; Sina Ojaghi
Abstract
Providing a dataset with a suitable volume and high accuracy for training deep neural networks is considered to be one of the basic requirements in that a suitable dataset in terms of the number and quality of images and labeling accuracy can have a great impact on the output accuracy of the trained ...
Read More
Providing a dataset with a suitable volume and high accuracy for training deep neural networks is considered to be one of the basic requirements in that a suitable dataset in terms of the number and quality of images and labeling accuracy can have a great impact on the output accuracy of the trained network. The dataset presented in this article contains 3000 images downloaded from online Iranian car sales companies, including Divar and Bama sites, which are manually labeled in three classes: car, truck, and bus. The labels are in the form of 5765 bounding boxes, which characterize the vehicles in the image with high accuracy, ultimately resulting in a unique dataset that is made available for public use.The YOLOv8s algorithm, trained on this dataset, achieves an impressive final precision of 91.7% for validation images. The Mean Average Precision (mAP) at a 50% threshold is recorded at 92.6%. This precision is considered suitable for city vehicle detection networks. Notably, when comparing the YOLOv8s algorithm trained with this dataset to YOLOv8s trained with the COCO dataset, there is a remarkable 10% increase in mAP at 50% and an approximately 22% improvement in the mAP range of 50% to 95%.
A. Fakhari; K. Kiani
Abstract
Image restoration and its different variations are important topics in low-level image processing. One of the main challenges in image restoration is dependency of current methods to the corruption characteristics. In this paper, we have proposed an image restoration architecture that enables us to address ...
Read More
Image restoration and its different variations are important topics in low-level image processing. One of the main challenges in image restoration is dependency of current methods to the corruption characteristics. In this paper, we have proposed an image restoration architecture that enables us to address different types of corruption, regardless of type, amount and location. The main intuition behind our approach is restoring original images from abstracted perceptual features. Using an encoder-decoder architecture, image restoration can be defined as an image transformation task. Abstraction of perceptual features is done in the encoder part of the model and determines the sampling point within original images' Probability Density Function (PDF). The PDF of original images is learned in the decoder section by using a Generative Adversarial Network (GAN) that receives the sampling point from the encoder part. Concretely, sampling from the learned PDF restores original image from its corrupted version. Pretrained network extracts perceptual features and Restricted Boltzmann Machine (RBM) makes the abstraction over them in the encoder section. By developing a new algorithm for training the RBM, the features of the corrupted images have been refined. In the decoder, the Generator network restores original images from abstracted perceptual features while Discriminator determines how good the restoration result is. The proposed approach has been compared with both traditional approaches like BM3D and with modern deep models like IRCNN and NCSR. We have also considered three different categories of corruption including denoising, inpainting and deblurring. Experimental results confirm performance of the model.
Seyed Mahdi Sadatrasoul; Omid Mahdi Ebadati; Amir Amirzadeh Irani
Abstract
Companies have different considerations for using smoothing in their financial statements, including annual general meeting, auditing, Regulatory and Supervisory institutions and shareholders requirements. Smoothing is done based on the various possible and feasible choices in identifying company’s ...
Read More
Companies have different considerations for using smoothing in their financial statements, including annual general meeting, auditing, Regulatory and Supervisory institutions and shareholders requirements. Smoothing is done based on the various possible and feasible choices in identifying company’s incomes, costs, expenses, assets and liabilities. Smoothing can affect credit scoring models reliability, it can cause to providing/not providing facilities to a non-worthy/worthy organization orderly, which are both known as decision errors and are reported as “type I” and “type II” errors, which are very important for Banks Loan portfolio. This paper investigates this issue for the first time in credit scoring studies on the authors knowledge and searches. The data of companies associated with a major Asian Bank are first applied using logistic regression. Different smoothing scenarios are tested, using wilcoxon statistic indicated that traditional credit scoring models have significant errors when smoothing procedures have more than 20% change in adjusting company’s financial statements and balance sheets parameters.
B.3. Communication/Networking and Information Technology
Ali Abdi Seyedkolaei
Abstract
Deploying multiple sinks instead of a single sink is one possible solution to improve the lifetime and durability of wireless sensor networks. Using multiple sinks leads to the definition of a problem known as the sink placement problem. In this context, the goal is to determine the optimal locations ...
Read More
Deploying multiple sinks instead of a single sink is one possible solution to improve the lifetime and durability of wireless sensor networks. Using multiple sinks leads to the definition of a problem known as the sink placement problem. In this context, the goal is to determine the optimal locations and number of sink nodes in the network to maximize the network's lifetime. In this paper, we propose a dynamic sensor assignment algorithm to address the sink placement problem and evaluate its performance against existing solution methods on a diverse set of instances. We conducted experiments in two stages. In the first stage, based on random instances and compared to the exact computational method using the CPLEX solver, and in the second stage, based on real-world instances compared to MC-JMSP (Model-Based Clustering- Joint Multiple Sink Placement) method. The results obtained in the first stage of the experiments indicate the superiority of the dynamic sensor assignment algorithm in runtime for all instances. Furthermore, the solution obtained by the dynamic sensor assignment algorithm is very close to the solution obtained by the CPLEX solver. In particular, the percentage error of the solution found by the proposed method compared to CPLEX in all experimented instances is less than 0.15%, indicating the effectiveness of the proposed method in finding the appropriate solution for assigning sensors to sinks. Also, the results of the second stage experiments show the superiority of the proposed method in both execution time and energy efficiency compared to the MC-JMSP method.
H.3. Artificial Intelligence
Afrooz Moradbeiky; Farzin Yaghmaee
Abstract
Knowledge graphs are widely used tools in the field of reasoning, where reasoning is facilitated through link prediction within the knowledge graph. However, traditional methods have limitations, such as high complexity or an inability to effectively capture the structural features of the graph. The ...
Read More
Knowledge graphs are widely used tools in the field of reasoning, where reasoning is facilitated through link prediction within the knowledge graph. However, traditional methods have limitations, such as high complexity or an inability to effectively capture the structural features of the graph. The main challenge lies in simultaneously handling both the structural and similarity features of the graph. In this study, we employ a constraint satisfaction approach, where each proposed link must satisfy both structural and similarity constraints. For this purpose, each constraint is considered from a specific perspective, referred to as a view. Each view computes a probability score using a GRU-RNN, which satisfies its own predefined constraint. In the first constraint, the proposed node must have a probability of over 0.5 with frontier nodes. The second constraint computes the Bayesian graph, and the proposed node must have a link in the Bayesian graph. The last constraint requires that a proposed node must fall within an acceptable fault. This allows for N-N relationships to be accurately determined, while also addressing the limitations of embedding. The results of the experiments showed that the proposed method improved performance on two standard datasets.
A. Lakizadeh; E. Moradizadeh
Abstract
Text sentiment classification in aspect level is one of the hottest research topics in the field of natural language processing. The purpose of the aspect-level sentiment analysis is to determine the polarity of the text according to a particular aspect. Recently, various methods have been developed ...
Read More
Text sentiment classification in aspect level is one of the hottest research topics in the field of natural language processing. The purpose of the aspect-level sentiment analysis is to determine the polarity of the text according to a particular aspect. Recently, various methods have been developed to determine sentiment polarity of the text at the aspect level, however, these studies have not yet been able to model well complementary effects of the context and aspect in the polarization detection process. Here, we present ACTSC, a method for determining the sentiment polarity of the text based on separate embedding of aspects and context. In the first step, ACTSC deals with separate modelling of the aspects and context to extract new representation vectors. Next, by combining generative representations of aspect and context, it determines the corresponding polarity to each particular aspect using a short-term memory network and a self-attention mechanism. Experimental results in the SemEval2014 dataset in both restaurant and laptop categories show that ACTSC has been able to improve the accuracy of aspect-based sentiment classification compared to the latest proposed methods.
H. Sadr; Mir M. Pedram; M. Teshnehlab
Abstract
With the rapid development of textual information on the web, sentiment analysis is changing to an essential analytic tool rather than an academic endeavor and numerous studies have been carried out in recent years to address this issue. By the emergence of deep learning, deep neural networks have attracted ...
Read More
With the rapid development of textual information on the web, sentiment analysis is changing to an essential analytic tool rather than an academic endeavor and numerous studies have been carried out in recent years to address this issue. By the emergence of deep learning, deep neural networks have attracted a lot of attention and become mainstream in this field. Despite the remarkable success of deep learning models for sentiment analysis of text, they are in the early steps of development and their potential is yet to be fully explored. Convolutional neural network is one of the deep learning methods that has been surpassed for sentiment analysis but is confronted with some limitations. Firstly, convolutional neural network requires a large number of training data. Secondly, it assumes that all words in a sentence have an equal contribution to the polarity of a sentence. To fill these lacunas, a convolutional neural network equipped with the attention mechanism is proposed in this paper which not only takes advantage of the attention mechanism but also utilizes transfer learning to boost the performance of sentiment analysis. According to the empirical results, our proposed model achieved comparable or even better classification accuracy than the state-of-the-art methods.
H.3. Artificial Intelligence
Habib Khodadadi; Vali Derhami
Abstract
The exploration-exploitation trade-off poses a significant challenge in reinforcement learning. For this reason, action selection methods such as ε-greedy and Soft-Max approaches are used instead of the greedy method. These methods use random numbers to select an action that balances exploration ...
Read More
The exploration-exploitation trade-off poses a significant challenge in reinforcement learning. For this reason, action selection methods such as ε-greedy and Soft-Max approaches are used instead of the greedy method. These methods use random numbers to select an action that balances exploration and exploitation. Chaos is commonly utilized across various scientific disciplines because of its features, including non-periodicity, unpredictability, ergodicity and pseudorandom behavior. In this paper, we employ numbers generated by different chaotic systems to select action and identify better maps in diverse states and quantities of actions. Based on our experiments on various environments such as the Multi-Armed Bandit (MAB), taxi-domain, and cliff-walking, we found that many of the chaotic methods increase the speed of learning and achieve higher rewards.
H. Morshedlou; A.R. Tajari
Abstract
Edge computing is an evolving approach for the growing computing and networking demands from end devices and smart things. Edge computing lets the computation to be offloaded from the cloud data centers to the network edge for lower latency, security, and privacy preservation. Although energy efficiency ...
Read More
Edge computing is an evolving approach for the growing computing and networking demands from end devices and smart things. Edge computing lets the computation to be offloaded from the cloud data centers to the network edge for lower latency, security, and privacy preservation. Although energy efficiency in cloud data centers has been widely studied, energy efficiency in edge computing has been left uninvestigated. In this paper, a new adaptive and decentralized approach is proposed for more energy efficiency in edge environments. In the proposed approach, edge servers collaborate with each other to achieve an efficient plan. The proposed approach is adaptive, and consider workload status in local, neighboring and global areas. The results of the conducted experiments show that the proposed approach can improve energy efficiency at network edges. e.g. by task completion rate of 100%, the proposed approach decreases energy consumption of edge servers from 1053 Kwh to 902 Kwh.
H.3.9. Problem Solving, Control Methods, and Search
Zahra Jahan; Abbas Dideban; Farzaneh Tatari
Abstract
This paper introduces an adaptive optimal distributed algorithm based on event-triggered control to solve multi-agent discrete-time zero-sum graphical games for unknown nonlinear constrained-input systems with external disturbances. Based on the value iteration heuristic dynamic programming, the proposed ...
Read More
This paper introduces an adaptive optimal distributed algorithm based on event-triggered control to solve multi-agent discrete-time zero-sum graphical games for unknown nonlinear constrained-input systems with external disturbances. Based on the value iteration heuristic dynamic programming, the proposed algorithm solves the event-triggered coupled Hamilton-Jacobi-Isaacs equations assuming unknown dynamics to develop distributed optimal controllers and satisfy leader-follower consensus for agents interacting on a communication graph. The algorithm is implemented using the actor-critic neural network, and unknown system dynamics are approximated using the identifier network. Introducing and solving nonlinear zero-sum discrete-time graphical games in the presence of unknown dynamics, control input constraints and external disturbances, differentiate this paper from the previously published works. Also, the control input, external disturbance, and the neural network's weights are updated aperiodic and only at the triggering instants to simplify the computational process. The closed-loop system stability and convergence to the Nash equilibrium are proven. Finally, simulation results are presented to confirm theoretical findings.