An efficient approach to Data clustering using the K-Means algorithm in Big data analytics

Autor

  • Jaganath S MGR Educational And Research Institute Autor
  • Logesh K Madha Engineering College Autor
  • Barath S Madha Engineering College Autor
  • Athiraja A Saveetha Engineering College Autor

Słowa kluczowe:

Big Data Analytics, K-Means Clustering, Optimization, Parallel Computing, Data Mining

Abstrakt

With the exponential growth of data, efficient clustering techniques are essential for extracting meaningful patterns in Big Data Analytics. The K-Means algorithm is widely used due to its simplicity and scalability. However, its performance is often hindered by high-dimensional data, initialization sensitivity, and computational complexity. This study proposes an optimized K-Means clustering approach that integrates an improved centroid initialization method and parallel processing to enhance efficiency in Big Data environments. The proposed method was evaluated using real-world datasets such as KDD Cup and UCI Machine Learning Repository, with data sizes ranging from 10GB to 100GB. Experimental results demonstrate a 30% reduction in execution time and a 15% improvement in clustering accuracy compared to traditional K-Means. The optimized approach also shows a 20% lower convergence time, making it suitable for large-scale applications. In conclusion, the enhanced K-Means algorithm significantly improves clustering performance in Big Data settings. The combination of advanced initialization and parallel computing ensures better scalability and accuracy, making it a viable solution for real-time analytics. Future work will focus on extending this approach to handle streaming data and non-Euclidean spaces.

Opublikowane

2025-06-23

Numer

Dział

Articles

Jak cytować

An efficient approach to Data clustering using the K-Means algorithm in Big data analytics. (2025). Frontiers in Science and Technology, 2(1). https://journal.dharapublishers.com/index.php/FST/article/view/9

Inne teksty tego samego autora

1 2 > >>