School of Electrical Engineering and Technology (SEET)

Permanent URI for this communityhttp://197.211.34.35:4000/handle/123456789/24

School of Electrical Engineering and Technology (SEET)

Browse

Search Results

Now showing 1 - 3 of 3
  • Item
    Comparative Analysis of Machine Learning Algorithms for Eccentricity Fault Classification in Salient Pole Synchronous Machine
    (IEEE, 2024-03-22) Shejwalkar, Ashwin; Yusuf, Latifa; Ilamparithi, Thirumarai Chelvan
    The paper performs a comparative study of Artificial Neural Networks (ANNs) and Convolutional Neural Networks (CNNs) for the classification of Static Eccentricity (SE) and Dynamic Eccentricity (DE) faults in a Salient Pole Synchronous Machine (SPSM). The SPSM was subjected to varying SE and DE severities, unbalanced source voltages, and load conditions. Stator and field current data were measured, and various time-domain and frequency-domain features were extracted from the above-mentioned data. Both networks were fed these features and compared based on classification accuracy, robustness, and computational complexity.
  • Item
    A Comparison of Strategies for Missing Values in Data on Machine Learning Classification Algorithms
    (IEEE, 2019) Makaba, T.; Dogo, E.
    Dealing with missing values in data is an important feature engineering task in data science to prevent negative impacts on machine learning classification models in terms of accurate prediction. However, it is often unclear what the underlying cause of the missing values in real-life data is or rather the missing data mechanism that is causing the missingness. Thus, it becomes necessary to evaluate several missing data approaches for a given dataset. In this paper, we perform a comparative study of several approaches for handling missing values in data, namely listwise deletion, mean, mode, k-nearest neighbors, expectation-maximization, and multiple imputations by chained equations. The comparison is performed on two real-world datasets, using the following evaluation metrics: Accuracy, root mean squared error, receiver operating characteristics, and the F1 score. Most classifiers performed well across the missing data strategies. However, based on the result obtained, the support vector classifier method overall performed marginally better for the numerical data and naïve Bayes classifier for the categorical data when compared to the other evaluated missing value methods.
  • Item
    Accessing Imbalance Learning Using Dynamic Selection Approach in Water Quality Anomaly Detection
    (MDPI, 2021) Dogo, E. M.; Nwulu, N. I.; Twala, B.; Aigbavboa, C.
    Automatic anomaly detection monitoring plays a vital role in water utilities’ distribution systems to reduce the risk posed by unclean water to consumers. One of the major problems with anomaly detection is imbalanced datasets. Dynamic selection techniques combined with ensemble models have proven to be effective for imbalanced datasets classification tasks. In this paper, water quality anomaly detection is formulated as a classification problem in the presences of class imbalance. To tackle this problem, considering the asymmetry dataset distribution between the majority and minority classes, the performance of sixteen previously proposed single and static ensemble classification methods embedded with resampling strategies are first optimised and compared. After that, six dynamic selection techniques, namely, Modified Class Rank (Rank), Local Class Accuracy (LCA), Overall-Local Accuracy (OLA), K-Nearest Oracles Eliminate (KNORA-E), K-Nearest Oracles Union (KNORA-U) and Meta-Learning for Dynamic Ensemble Selection (META-DES) in combination with homogeneous and heterogeneous ensemble models and three SMOTE-based resampling algorithms (SMOTE, SMOTE+ENN and SMOTE+Tomek Links), and one missing data method (missForest) are proposed and evaluated. A binary real-world drinking-water quality anomaly detection dataset is utilised to evaluate the models. The experimental results obtained reveal all the models benefitting from the combined optimisation of both the classifiers and resampling methods. Considering the three performance measures (balanced accuracy, F-score and G-mean), the result also shows that the dynamic classifier selection (DCS) techniques, in particular, the missForest+SMOTE+RANK and missForest+SMOTE+OLA models based on homogeneous ensemble-bagging with decision tree as the base classifier, exhibited better performances in terms of balanced accuracy and G-mean, while the Bg+mF+SMENN+LCA model based on homogeneous ensemble-bagging with random forest has a better overall F1-measure in comparison to the other models.