Submitted:
23 July 2025
Posted:
24 July 2025
You are already at the latest version
Abstract
Keywords:
1. Introduction
2. Related Works
3. Methodology
3.1. Instruments:
3.2. Performance
3.2.1. First Step
3.2.2. Second Step
3.2.3. Third Step
3.2.4. Forth Step
3.2.5. Fifth Step
4. Results
5. Discussion
6. Conclusions
Acknowledgments
Appendix A
| Cluster Metod | Quantity Of Cluster | Seeding | With Out Membership | Poor Cluster | Kruskal Wallis | DSCF | |
| K-Mean | 2 | - | no | no | Significant For Courtesy, Perceived Value, Aesthetics | Significant Courtesy-Perceived Value- Aesthetics | |
| K-Mean Fast | 2 | - | no | no | Significant For Courtesy, Aesthetics, Perceived value |
Significant For Courtesy, Aesthetics, Perceived value |
|
| K-Mean H2O | 2 | - | no | no | Unsignificant For at attitude | Significant For Access Aesthetics, Perceived Value, Courtesy, Loyalty |
|
| K-Mean Kernel | 2 | - | no | no | Unsignificant For Access, | Unsignificant For Access, | |
| K-Medoid | 4 | - | no | no | Significant All Except Loyalty | Significant For at attitude | |
| Fuzzy C-Mean | 2 | - | no | no | Significant For Aesthetics, Courtesy, Access, at attitude, Perceived value | Significant For Aesthetics, Courtesy, Access, at attitude, Perceived value - | |
| Canopy | 2 | 5 | no | no | Significant For Access, Courtesy, Perceived value, Quality | Significant For Access, Courtesy, Perceived value, Quality | |
| Cascade k-Mean | 2 | - | no | no | Significant For Aesthetics, Courtesy, Perceived value | Significant For Aesthetics, Courtesy, Perceived value | |
| EM | 2 | - | no | no | Significant For Access, Courtesy, Perceived value | Significant For Access, Courtesy, Perceived value | |
| Farthest First | 2 | 2 | no | no | Significant All Index |
Significant All Index |
|
| Gen clust plus Plus | 2 | 10 | no | no | Significant For Courtesy, at attitude | Significant For Courtesy, at attitude | |
| LVQ | 2 | - | no | no | Significant Except Loyalty | Significant Except Loyalty | |
| Make Density Based | 2 | - | no | no | Significant Aesthetics, Perceived value, Courtesy |
Significant Aesthetics, Perceived value, Courtesy |
|
| sIB | 2 | 1 | no | no | Significant Except Loyalty | Significant Except Loyalty | |
| X-Mean | 2 | 10 | no | no | Significant For Aesthetics, Courtesy, Perceived value | Significant For Aesthetics, Courtesy, Perceived value |
| Fit Measures | ||||||||
| χ² | df | p-value | CFI | TLI | SRMR | RMSEA | Lower | Upper |
| 784.82 | 329 | < .001 | 0.92282 | 0.91132 | 0.054881 | 0.049958 | 0.059824 | |
| Cluster Method | Cluster Quantity | Quality | Aesthetics | Perceived value | Attitude | Access | Courtesy | Loyalty | Total Effect Size |
Silhouette Index |
| K-Mean E-S | 2 | 0.00 | 0.04 | 0.33 | 0.00 | 0.00 | 0.57 | 0.00 | 0.94 | 0.209 |
| P Value | unsignificant | significant | significant | unsignificant | unsignificant | significant | unsignificant | |||
| K-Mean Fast E-S | 2 | 0.00 | 0.04 | 0.33 | 0.00 | 0.00 | 0.57 | 0.00 | 0.94 | 0.210 |
| P Value | unsignificant | significant | significant | unsignificant | unsignificant | significant | unsignificant | |||
| X-Mean E-S | 2 | 0.00 | 0.05 | 0.34 | 0.00 | 0.00 | 0.54 | 0.00 | 0.94 | 0.210 |
| P Value | unsignificant | significant | significant | unsignificant | unsignificant | significant | unsignificant | |||
|
Cascade K-Mean E-S |
2 | 0.00 | 0.06 | 0.34 | 0.00 | 0.00 | 0.54 | 0.00 | 0.94 | 0.209 |
| P Value | unsignificant | significant | significant | unsignificant | unsignificant | significant | unsignificant | |||
| Make-Density Based E-S | 2 | 0.00 | 0.04 | 0.29 | 0.00 | 0.00 | 0.60 | 0.00 | 0.93 | 0.209 |
| P Value | unsignificant | significant | significant | unsignificant | unsignificant | significant | unsignificant | |||
|
K-Mean-H2o E-S |
2 | 0.01 | 0.04 | 0.33 | 0.00 | 0.02 | 0.55 | 0.01 | 0.93 | 0.207 |
| P Value | unsignificant | significant | significant | unsignificant | significant | significant | significant | |||
|
Fuzzy C-Mean E-S |
2 | 0.003 | 0.012 | 0.390 | 0.044 | 0.009 | 0.462 | 0.004 | 0.918 | 0.201 |
| P Value | unsignificant | significant | significant | significant | significant | significant | unsignificant | |||
| sIB E-S | 2 | 0.01 | 0.02 | 0.27 | 0.01 | 0.02 | 0.54 | 0.00 | 0.87 | 0.198 |
| P Value | significant | significant | significant | significant | significant | significant | unsignificant | |||
| Em E-S | 2 | 0.00 | 0.01 | 0.28 | 0.00 | 0.01 | 0.57 | 0.00 | 0.86 | 0.200 |
| P Value | unsignificant | unsignificant | significant | unsignificant | significant | significant | unsignificant | |||
| Canopy E-S | 2 | 0.01 | 0.00 | 0.43 | 0.00 | 0.04 | 0.35 | 0.00 | 0.85 | 0.196 |
| P Value | significant | unsignificant | significant | unsignificant | significant | significant | unsignificant | |||
| LVQ E-S | 2 | 0.08 | 0.01 | 0.49 | 0.04 | 0.06 | 0.02 | 0.00 | 0.71 | 0.166 |
| P Value | significant | significant | significant | significant | significant | significant | unsignificant | |||
| Gen clust Plus Plus E-S | 2 | 0.00 | 0.00 | 0.01 | 0.68 | 0.00 | 0.01 | 0.00 | 0.69 | 0.157 |
| P Value | unsignificant | unsignificant | unsignificant | significant | unsignificant | significant | unsignificant | |||
|
Farthest First E-S |
2 | 0.11 | 0.04 | 0.05 | 0.04 | 0.01 | 0.05 | 0.02 | 0.32 | 0.175 |
| P Value | significant | significant | significant | significant | significant | significant | significant | |||
|
K-Mean-Kernel E-S |
2 | 0.03 | 0.02 | 0.01 | 0.12 | 0.00 | 0.05 | 0.06 | 0.28 | 0.047 |
| P Value | significant | significant | significant | significant | unsignificant | significant | significant | |||
| Kmedoid E-S | 4 | 0.16 | 0.03 | 0.13 | 0.24 | 0.05 | 0.39 | 0.04 | 0.24 | 0.167 |
| P Value | unsignificant | unsignificant | unsignificant | significant | unsignificant | unsignificant | unsignificant | |||
| Random, Clope, SOM, Cobweb, DBSCAN | Unsignificant all | |||||||||
References
- K. Rubarth, M. Pauly, and F. Konietschke, “Ranking procedures for repeated measures designs with missing data: Estimation, testing and asymptotic theory,” Stat Methods Med Res, vol. 31, no. 1, pp. 105–118, Jan. 2022. [CrossRef]
- P. D’Urso, M. Disegna, R. Massari, and G. Prayag, “Bagged fuzzy clustering for fuzzy data: An application to a tourism market,” Knowledge-Based Systems, vol. 73, pp. 335–346, Jan. 2015, https://eprints.bournemouth.ac.uk/23278/.
- J. A. Saunders, “Cluster Analysis for Market Segmentation,” European Journal of Marketing, vol. 14, no. 7, pp. 422–435, Jul. 1980. [CrossRef]
- P. Kotler,Part7: Customer Value–Driven Marketing Strategy , Principles of marketing, Global, Seventeen edition. Harlow, England: Pearson, 2018. (p213-220).
- B. G. Pitts and D. K. Stotlar, Chapter 4 “Market Segmentation, Targeting, and Positioning”, Fundamentals of sport marketing, 4. ed. in Sport management library. Morgantown, W.Va: Fitness Information Technology, 2013. (p69-83). [Online]. Available: https://fitpublishing.com/books/fundamentals-sport-business-marketing.
- H. M. Zangana and A. M. Abdulazeez, “Developed Clustering Algorithms for Engineering Applications: A Review,” Int. J. Inform. Inf. Sys. and Comp. Eng., vol. 4, no. 2, Art. no. 2, Dec. 2023. [CrossRef]
- M. Deldadehasl, H. H. Karahroodi, and P. Haddadian Nekah, “Customer Clustering and Marketing Optimization in Hospitality: A Hybrid Data Mining and Decision-Making Approach from an Emerging Economy,” Tourism and Hospitality, vol. 6, no. 2, Art. no. 2, Jun. 2025. [CrossRef]
- K. R. Clarke, P. J. Somerfield, and R. N. Gorley, “Clustering in non-parametric multivariate analyses,” J. Exp. Mar. Biol. Ecol., vol. 483, pp. 147–155, Oct. 2016. [CrossRef]
- M. Z. Rodriguez et al., “Clustering algorithms: A comparative approach,” PLOS ONE, vol. 14, no. 1, p. e0210236, Jan. 2019. [CrossRef]
- C. Hennig, “An empirical comparison and characterisation of nine popular clustering methods,” Adv. Data Anal. Classif., vol. 16, no. 1, pp. 201–229, Mar. 2022. [CrossRef]
- M.-F. Kaya and M. Schoop, “Analytical Comparison of Clustering Techniques for the Recognition of Communication Patterns,” Group Decis. Negot., vol. 31, no. 3, pp. 555–589, Jun. 2022. [CrossRef]
- E. Costa, I. Papatsouma, and A. Markos, “Benchmarking distance-based partitioning methods for mixed-type data,” Adv. Data Anal. Classif., vol. 17, no. 3, pp. 701–724, Sep. 2023. [CrossRef]
- P. Sepin, J. Kemnitz, S. R. Lakani, and D. Schall, “Comparison of Clustering Algorithms for Statistical Features of Vibration Data Sets,” in Data Science—Analytics and Applications, Cham: Springer Nature Switzerland, 2024, pp. 3–11. [CrossRef]
- F. Faul, E. Erdfelder, A.-G. Lang, and A. Buchner, “G*Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences,” Behavior Research Methods, vol. 39, no. 2, pp. 175–191, May 2007. [CrossRef]
- F. Faul, E. Erdfelder, A. Buchner, and A.-G. Lang, “Statistical power analyses using G*Power 3.1: Tests for correlation and regression analyses,” Behavior Research Methods, vol. 41, no. 4, pp. 1149–1160, Nov. 2009. [CrossRef]
- M. F. Diallo and A. M. Seck, “How store service quality affects attitude toward store brands in emerging countries: Effects of brand cues and the cultural context,” Journal of Business Research, vol. 86, pp. 311–320, May 2018. [CrossRef]
- R. Sánchez-Fernández, M. Á. Iniesta-Bonillo, and M. B. Holbrook, “The Conceptualisation and Measurement of Consumer Value in Services,” International Journal of Market Research, vol. 51, no. 1, pp. 1–17, Jan. 2009. [CrossRef]
- D. Soares-Silva, G. H. S. M. De Moraes, A. Cappellozza, and C. Morini, “Explaining library user loyalty through perceived service quality: What is wrong?,” Asso for Info Science & Tech, vol. 71, no. 8, pp. 954–967, Aug. 2020. [CrossRef]
- J. F. Hair, G. T. M. Hult, C. M. Ringle, M. Sarstedt, N. P. Danks, and S. Ray, “Evaluation of Reflective Measurement Models,” in Partial Least Squares Structural Equation Modeling (PLS-SEM) Using R, in Classroom Companion: Business., Cham: Springer International Publishing, 2021, pp. 75–90. [CrossRef]
- M. G. Thorpe, C. M. Milte, D. Crawford, and S. A. McNaughton, “A comparison of the dietary patterns derived by principal component analysis and cluster analysis in older Australians,” Int J Behav Nutr Phys Act, vol. 13, no. 1, p. 30, Dec. 2016. [CrossRef]
- J. D. Spurrier, “Additional Tables for Steel–Dwass–Critchlow–Fligner Distribution-Free Multiple Comparisons of Three Treatments,” Communications in Statistics - Simulation and Computation, vol. 35, no. 2, pp. 441–446, Jul. 2006. [CrossRef]
- P. J. Rousseeuw, “Silhouettes: A graphical aid to the interpretation and validation of cluster analysis,” Journal of Computational and Applied Mathematics, vol. 20, pp. 53–65, Nov. 1987. [CrossRef]
- R. Bakeman, “Recommended effect size statistics for repeated measures designs,” Behavior Research Methods, vol. 37, no. 3, pp. 379–384, Aug. 2005. [CrossRef]
- C. O. Fritz, P. E. Morris, and J. J. Richler, “Effect size estimates: Current use, calculations, and interpretation,” Journal of Experimental Psychology: General, vol. 141, no. 1, pp. 2–18, 2012. [CrossRef]
- B. W. Yap and C. H. Sim, “Comparisons of various types of normality tests,” Journal of Statistical Computation and Simulation, vol. 81, no. 12, pp. 2141–2155, Dec. 2011. [CrossRef]
- M. Saculinggan and E. A. Balase, “Empirical Power Comparison Of Goodness of Fit Tests for Normality In The Presence of Outliers,” J. Phys.: Conf. Ser., vol. 435, p. 012041, Apr. 2013. [CrossRef]
- M. Sarstedt and E. Mooi, “Principal Component and Factor Analysis,” in A Concise Guide to Market Research, Springer, Berlin, Heidelberg, 2019, pp. 257–299. [CrossRef]
- M. Neuhäuser and F. Bretz, “Nonparametric All-Pairs Multiple Comparisons,” Biometrical Journal, vol. 43, no. 5, pp. 571–580, 2001. [CrossRef]
- E. Brunner and U. Munzel, “The Nonparametric Behrens-Fisher Problem: Asymptotic Theory and a Small-Sample Approximation,” Biometrical Journal, vol. 42, no. 1, pp. 17–25, 2000. [CrossRef]
- M. W. Fagerland, L. Sandvik, and P. Mowinckel, “Parametric methods outperformed non-parametric methods in comparisons of discrete numerical variables,” BMC Med Res Methodol, vol. 11, no. 1, p. 44, Apr. 2011. [CrossRef]
- J. D. Spurrier, “Additional Tables for Steel–Dwass–Critchlow–Fligner Distribution-Free Multiple Comparisons of Three Treatments,” Communications in Statistics - Simulation and Computation, vol. 35, no. 2, pp. 441–446, July 2006. [CrossRef]
- M. Hollander, D. A. Wolfe, and E. Chicken, Nonparametric statistical methods, Third edition. in Wiley series in probability and statistics. Hoboken, NJ: Wiley, 2014. (p293-300) [Online]. Available: https://onlinelibrary.wiley.com/. [CrossRef]
| 1 | Dwass-Steel-Critchlow-Fligner |


| Authors | Sample Size | Subject Used for Clustering | Compare Method | Compared Algorithm | Result |
|
Rodriguez et al. 2019 |
400 | Artificial datasets | distances between classes and correlations between features and average of the best accuracies | hierarchical, Clara, k-means, spectral, hc model, subspace, optics, DBSCAN, EM | Spectral algorithm consistently provided best performance for datasets with 10+ feature |
|
Hening 2021 |
35043 sample | Different categories (images, Texts, digits) there is large variation between the data sets | Just compare output of software | K-means, (Clara), (m-clust), (em-skew), (teigen), Single linkage hierarchical clustering, Average linkage hierarchical clustering, Complete linkage hierarchical clustering, Spectral clustering | “ The Gaussian mixture is the best for the largest amount of data. Differences between the other methods are not that pronounced, and all of them did best in some data sets.” |
|
Kaya And Schoop 2022 |
Ten negotiation experiments of several hundred participants with a total of 7026 exchanged negotiation interactions | Negotiation Support Systems (text) |
Principal Component Analysis | K-means, X-means, DBSCAN agglomerative |
With internal index k-Means and with external index k-Means or DBSCAN, performance is better. |
|
Costa et al 2023 |
n = 100 low, 600 moderates and 1000 |
Normal Mixed-type data | ANOVA-ARI-AMI-Effect size η2 | KAMILA, FAMD/K-Means, K-Prototypes, M-S K-Means, Mixed RKM, HL/PAM, Mixed K-Means, Gower/PAM | KAMILA, K-Prototypes and sequential Factor Analysis and K-Means clustering typically performed better than other methods |
|
Sepin et al 2024 |
data1:512 data2: ambiguous data3: ambiguous |
vibration data |
1 Arithmetic mean of absolute values (Abs Mean). 2Median of absolute values (Abs Median). 3 Standard deviation (Std). 4 Interquartile range (IQR). 5 Skewness of absolute values (Abs Skew). 6 Kurtosis of absolute values (Abs Kurt) | K-means, Gaussian Mixture Model, OPTICS | K-means outperformed GMM slightly, whereas OPTICS performed significantly worse. |
| Data Mining Platform | Rapid Miner | Weka |
| Algorithms Compared | K-Mean, K-Mean (H2O), K-Mean Kernal, K-Mean fast, X-Mean, K-Medoids, Fuzzy C-mean, Agglomerative Clustering, DBSCAN, Top-Down Clustering, Flatten Clustering, G-Means, Random | BANG-File, Canopy, Cascade Simple K-Mean, Cobweb, farthest first, Gen-Clust plus plus, Hierarchical Clustering, LVQ, EM, make density-based Clustering, OPTICS, SOM, sIB, , X-means |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
