狠狠撸

狠狠撸Share a Scribd company logo
Benchmarking Automated
Machine Learning for Clustering
Biagio Licari
Candidate
Prof. Sylvio Barbon Junior
Supervisor
A.Y 22/23
? Introduction
? Why benchmarking
? Benchmark Design
? Conclusion
? Results
Index
Automated Machine Learning
Introduction
? AutoML automates complex and time-consuming tasks in Machine
Learning pipeline
? Enhancing accessibility for individuals with diverse expertise levels
? AutoML tasks include feature engineering, algorithm selection and
hyperparameter optimization or both combined (CASH).
? AutoML frameworks mainly tackle supervised problems
? Validation of results for unsupervised problems is challenging
? Clustering: A foundational technique with broad applications in
pattern recognition, image segmentation, and anomaly detection
? Data clustering demands expertise in handling complex and diverse
datasets
? Automating the generation of unsupervised clustering solutions poses
a significant challenge
? Unsupervised clustering lacks clear targets
? No single metric adequately describes every dataset
? Meta-learning approaches hold potential in tackling the CASH
problem
AutoML for Clustering
Introduction
? The lack of standardized benchmarking criteria for AutoML in
clustering is the primary motivation for this research
? Evaluating existing AutoML frameworks for unsupervised learning
remains an open issue
? Benchmarking is essential for users to navigate the AutoML
landscape effectively
? Standardized benchmarks promote transparency, replicability, and
future research
? The proposed benchmarking criteria only consider approaches that
support a working repository of code for reproducibility
? AutoML4Clust, cSmartML, Autocluster, and ML2DAC
AutoML for Clustering
Why benchmarking
? General Comparison
? Compares and contrasts design principles, parallelization
capabilities, and clustering algorithm diversity
? Highlights unique features of each AutoML framework
? Clustering Quality
? Uses specific Cluster Validity Indices (CVI) to measure clustering
quality and similarity
? CVI metrics categorized into internal and external
? Scalability & Consistency
? Evaluates AutoML scalability and consistency across datasets
with varying dimensions, instances, and noise
Benchmark Design
Evaluation Criteria
? Experiments were conducted on a
high-end workstation
? Isolated environment for each
framework
? Employed Neptune.AI for collecting
meta-data
? Set a time limit of 5 hours for each
experiment
? RAM constraint settled up to 16GiB
? User-intent settled to maximize the
quality of the result
HW Configuration
CPU Intel Core i9-10980XE
GPU NVIDIA Quadro T1000 Mobile
RAM 62.5 GiB
Storage SK Hynix PC801 NVMe 1TB
OS Ubuntu 22.04.2 LTS
Benchmark Design
HW & Framework Configurations
Feature Range
Dimension 2-99
Cluster 2-21
Samples 200-4000
Overlapness 1e-6 - 1e-5
Aspect Ref. 1-5
Imbalance Ratio 1-3
Benchmark Design
Datasets
? 100 synthetic datasets generated using Repliclust
AutoML4CLust Autocluster cSmartML ML2DAC
Release Year 2021 2019 2021 2023
Stars on Github 2 50 0 0
Last Commit Oct. 13, 2020 Jan. 9, 2023 Dec. 27, 2021 Sept. 18, 2023
Optimization
BO, RO
Hyp, BOHB
BO, RO Evo BO
Task CASH CASH
AS->HPO
Meta-CVI
Cash
Meta-CVI
Metalearning No Yes Yes Yes
Clustering
Algorithms
4 10 8 9
Results
General Comparison
? Parallelization features:
? AutoML4Clust
? Supports concurrent optimization tasks with BOHB and
HyperBand
? cSmartML
? Excels with internal parallelization for hyperparameter
optimization
? Autocluster
? Lacks parallelization during execution
? ML2DAC
? Employs Dask, an open-source Python library for parallel
computing.
? Not implemented using BO
Results
General Comparison
Results
General Comparison
? The adaptability of AutoML frameworks significantly impacts
clustering effectiveness
? Search Space features:
? ML2DAC and Autocluster offer a wider search space range
for AS
? AutoML4Clust’s search space comprises solely K-centered
clustering algorithms
? cSmartML prioritizes hierarchy models with some density and
graph theory adoption.
Results
Clustering Quality
? Silhouette - Internal CVI
? Evaluates cohesion and separation of clusters.
? Davies-Bouldin - Internal CVI
? Measures the average similarity between each cluster and its most
similar cluster.
? Adjusted Rand Index - External CVI
? Measures the similarity between the clustering result and the
ground truth
(c) CD Plot considering ARI as reference metric.
(b) CD Plot considering DBS as reference metric.
? Performed the non-parametric Friedman test
? Conducted the Nemenyi post-hoc test to identify significant
differences between pairs
? Critical Difference (CD) displays the average rank of each
framework
? statistically indistinguishable frameworks are connected
(a) CD Plot considering SIL as reference metric.
Results
Statistical Analysis
? Performed the Bayesian Bradley-Terry model (BBT) to compare
frameworks performance.
? BBT model assigns each framework a "merit number" determining
its performance.
? The Bayesian approach enables the determination of probabilities
associated with framework rankings.
? Provides confidence beyond simple ranking, enhancing
decision-making.
Results
Bayesian Bradley-Terry Tree Analysis
? BBT procedure results contrast previous analysis
? Silhouette
? ML2DAC superior to both Autocluster and AutoML4Clust
? Autocluster outperforms AutoML4Clust
? Davies-Boudlin
? ML2DAC and Autocluster superior to other frameworks
? No claim can be made between ML2DAC and Autocluster
? Adjustet Rand Index
? ML2DAC superior to both Autocluster and AutoML4Clust
? cSmartML is confirmed as inferior to other analyzed frameworks
Results
Bayesian Bradley-Terry Tree
Results
Scalability & Consistency
? cSmartML exhibits an exponential growth across dataset instances
? AutoML4Clust exhibits a linear growth
? Autocluster 2x faster than cSmartML
? ML2DAC 9x faster than cSmartML
Results
Scalability & Consistency
? Autocluster is the most resource-efficient option
? ML2DAC requires approximately 40.51% more memory
? with higher performance in clustering quality
? cSmartML requires approximately 50.4% more memory
? with lower performance in clustering quality
? ML2DAC emerged slightly ahead
? ML2DAC excelled in multiple tasks but was not consistently the
top performer
? Rapid development within the AutoML community
? Significant progress in unsupervised learning, especially in
clustering
? Room for enhancement
? Meta-learning approaches may offer solutions to improve
performance.
? Balancing automation with transparency is crucial in order to
maintain accessibility to clustering algorithm principles.
Conclusion
Thank you for your attention !
? Z?ller, M., & Huber, M.F. (2019). Benchmark and Survey of
Automated Machine Learning Frameworks. J. Artif. Intell. Res., 70,
409-472.
? Gijsbers, P., Bueno, M.L., Coors, S., LeDell, E., Poirier, S., Thomas,
J., Bischl, B., & Vanschoren, J. (2022). AMLB: an AutoML
Benchmark. ArXiv, abs/2207.12560.
? Luxburg, U.V., Williamson, R.C., & Guyon, I. (2009). Clustering:
Science or Art? ICML Unsupervised and Transfer Learning.
? Wainer, J. (2022). A Bayesian Bradley-Terry model to compare
multiple ML algorithms on multiple data sets. ArXiv, abs/2208.04935.
References
? Tschechlov, D., Fritz, M., & Schwarz, H. (2021). AutoML4Clust:
Efficient AutoML for Clustering Analyses. International Conference on
Extending Database Technology.
? Shawi, R.E., & Sakr, S. (2022). cSmartML-Glassbox: Increasing
Transparency and Controllability in Automated Clustering. 2022 IEEE
International Conference on Data Mining Workshops (ICDMW), 47-
54.
? Treder-Tschechlov, D., Fritz, M., Schwarz, H., & Mitschang, B. (2023).
ML2DAC: Meta-Learning to Democratize AutoML for Clustering
Analysis. Proceedings of the ACM on Management of Data, 1, 1 - 26.
? Wong, W. Y. (2019). Autocluster: AutoML for clustering models in
sklearn.
References

More Related Content

Similar to Benchmarking Automated Machine Learning For Clustering (20)

PPTX
Data Quality for Machine Learning Tasks
Hima Patel
?
PPTX
AI hype or reality
Awantik Das
?
PDF
Iod session 3423 analytics patterns of expertise, the fast path to amazing ...
Rachel Bland
?
PPTX
Moving from BI to AI : For decision makers
zekeLabs Technologies
?
PDF
analytic hierarchy_process
FEG
?
PDF
5 analytic hierarchy_process
FEG
?
PDF
Kaggle Higgs Boson Machine Learning Challenge
Bernard Ong
?
PDF
Advanced Optimization for the Enterprise Webinar
SigOpt
?
PDF
The Data Science Process - Do we need it and how to apply?
Ivo Andreev
?
PDF
“A Practical Guide to Implementing ML on Embedded Devices,” a Presentation fr...
Edge AI and Vision Alliance
?
PDF
Can we induce change with what we measure?
Michaela Greiler
?
PDF
Measuring the Validity of Clustering Validation Datasets
michaelaupetit1
?
PPTX
Everything you need to know about AutoML
Arpitha Gurumurthy
?
PDF
Pr dc 2015 sql server is cheaper than open source
Terry Bunio
?
PDF
“Introduction to Optimizing ML Models for the Edge,” a Presentation from Cisc...
Edge AI and Vision Alliance
?
PPTX
Machine Learning With ML.NET
Dev Raj Gautam
?
PDF
Goal Decomposition and Abductive Reasoning for Policy Analysis and Refinement
Emil Lupu
?
PPTX
ExplainableAI.pptx
Andrea Morichetta
?
PPTX
1 st review pothole srm bi1 st review pothole srm bi1 st review pothole srm bi
sathiyasowmi
?
PPTX
Aditya Bhattacharya - Enterprise DL - Accelerating Deep Learning Solutions to...
Aditya Bhattacharya
?
Data Quality for Machine Learning Tasks
Hima Patel
?
AI hype or reality
Awantik Das
?
Iod session 3423 analytics patterns of expertise, the fast path to amazing ...
Rachel Bland
?
Moving from BI to AI : For decision makers
zekeLabs Technologies
?
analytic hierarchy_process
FEG
?
5 analytic hierarchy_process
FEG
?
Kaggle Higgs Boson Machine Learning Challenge
Bernard Ong
?
Advanced Optimization for the Enterprise Webinar
SigOpt
?
The Data Science Process - Do we need it and how to apply?
Ivo Andreev
?
“A Practical Guide to Implementing ML on Embedded Devices,” a Presentation fr...
Edge AI and Vision Alliance
?
Can we induce change with what we measure?
Michaela Greiler
?
Measuring the Validity of Clustering Validation Datasets
michaelaupetit1
?
Everything you need to know about AutoML
Arpitha Gurumurthy
?
Pr dc 2015 sql server is cheaper than open source
Terry Bunio
?
“Introduction to Optimizing ML Models for the Edge,” a Presentation from Cisc...
Edge AI and Vision Alliance
?
Machine Learning With ML.NET
Dev Raj Gautam
?
Goal Decomposition and Abductive Reasoning for Policy Analysis and Refinement
Emil Lupu
?
ExplainableAI.pptx
Andrea Morichetta
?
1 st review pothole srm bi1 st review pothole srm bi1 st review pothole srm bi
sathiyasowmi
?
Aditya Bhattacharya - Enterprise DL - Accelerating Deep Learning Solutions to...
Aditya Bhattacharya
?

Recently uploaded (20)

DOCX
ACCOMPLISHMENT AS OF MAY 15 RCT ACCOMPLISHMENT AS OF MAY 15 RCT ACCOMPLISHMEN...
JoemarAgbayani1
?
PPTX
Presentation abdominal distension (1).pptx
ChZiaullah
?
PDF
顿补迟à补补补补补补补补补别苍驳颈苍别别别别别别别别别别别别别别别别别别别别别别别
juadsr96
?
PDF
5- Global Demography Concepts _ Population Pyramids .pdf
pkhadka824
?
PPTX
在线购买英国本科毕业证苏格兰皇家音乐学院水印成绩单搁厂础惭顿学费发票
Taqyea
?
PPTX
Monitoring Improvement ( Pomalaa Branch).pptx
fajarkunee
?
DOCX
? 1. Solvent R-WPS Office work scientific
NohaSalah45
?
PPTX
Krezentios memories in college data.pptx
notknown9
?
PDF
Orchestrating Data Workloads With Airflow.pdf
ssuserae5511
?
PPTX
办理学历认证滨苍蹿辞谤尘补迟颈肠蝉尝别迟迟别谤新加坡英华美学院毕业证书,滨苍蹿辞谤尘补迟颈肠蝉成绩单
Taqyea
?
PPTX
Data anlytics Hospitals Research India.pptx
SayantanChakravorty2
?
PDF
IT GOVERNANCE 4-2 - Information System Security (1).pdf
mdirfanuddin1322
?
PDF
Unlocking Insights: Introducing i-Metrics Asia-Pacific Corporation and Strate...
Janette Toral
?
PPTX
Artificial intelligence Presentation1.pptx
SaritaMahajan5
?
PDF
TCU EVALUATION FACULTY TCU Taguig City 1st Semester 2017-2018
MELJUN CORTES
?
DOCX
COT Feb 19, 2025 DLLgvbbnnjjjjjj_Digestive System and its Functions_PISA_CBA....
kayemorales1105
?
PPTX
Discrete Logarithm Problem in Cryptography (1).pptx
meshablinx38
?
PPTX
Module-2_3-1eentzyssssssssssssssssssssss.pptx
ShahidHussain66691
?
PPT
Reliability Monitoring of Aircrfat commerce
Rizk2
?
PPTX
Project_Update_Summary.for the use from PM
Odysseas Lekatsas
?
ACCOMPLISHMENT AS OF MAY 15 RCT ACCOMPLISHMENT AS OF MAY 15 RCT ACCOMPLISHMEN...
JoemarAgbayani1
?
Presentation abdominal distension (1).pptx
ChZiaullah
?
顿补迟à补补补补补补补补补别苍驳颈苍别别别别别别别别别别别别别别别别别别别别别别别
juadsr96
?
5- Global Demography Concepts _ Population Pyramids .pdf
pkhadka824
?
在线购买英国本科毕业证苏格兰皇家音乐学院水印成绩单搁厂础惭顿学费发票
Taqyea
?
Monitoring Improvement ( Pomalaa Branch).pptx
fajarkunee
?
? 1. Solvent R-WPS Office work scientific
NohaSalah45
?
Krezentios memories in college data.pptx
notknown9
?
Orchestrating Data Workloads With Airflow.pdf
ssuserae5511
?
办理学历认证滨苍蹿辞谤尘补迟颈肠蝉尝别迟迟别谤新加坡英华美学院毕业证书,滨苍蹿辞谤尘补迟颈肠蝉成绩单
Taqyea
?
Data anlytics Hospitals Research India.pptx
SayantanChakravorty2
?
IT GOVERNANCE 4-2 - Information System Security (1).pdf
mdirfanuddin1322
?
Unlocking Insights: Introducing i-Metrics Asia-Pacific Corporation and Strate...
Janette Toral
?
Artificial intelligence Presentation1.pptx
SaritaMahajan5
?
TCU EVALUATION FACULTY TCU Taguig City 1st Semester 2017-2018
MELJUN CORTES
?
COT Feb 19, 2025 DLLgvbbnnjjjjjj_Digestive System and its Functions_PISA_CBA....
kayemorales1105
?
Discrete Logarithm Problem in Cryptography (1).pptx
meshablinx38
?
Module-2_3-1eentzyssssssssssssssssssssss.pptx
ShahidHussain66691
?
Reliability Monitoring of Aircrfat commerce
Rizk2
?
Project_Update_Summary.for the use from PM
Odysseas Lekatsas
?
Ad

Benchmarking Automated Machine Learning For Clustering

  • 1. Benchmarking Automated Machine Learning for Clustering Biagio Licari Candidate Prof. Sylvio Barbon Junior Supervisor A.Y 22/23
  • 2. ? Introduction ? Why benchmarking ? Benchmark Design ? Conclusion ? Results Index
  • 3. Automated Machine Learning Introduction ? AutoML automates complex and time-consuming tasks in Machine Learning pipeline ? Enhancing accessibility for individuals with diverse expertise levels ? AutoML tasks include feature engineering, algorithm selection and hyperparameter optimization or both combined (CASH). ? AutoML frameworks mainly tackle supervised problems ? Validation of results for unsupervised problems is challenging
  • 4. ? Clustering: A foundational technique with broad applications in pattern recognition, image segmentation, and anomaly detection ? Data clustering demands expertise in handling complex and diverse datasets ? Automating the generation of unsupervised clustering solutions poses a significant challenge ? Unsupervised clustering lacks clear targets ? No single metric adequately describes every dataset ? Meta-learning approaches hold potential in tackling the CASH problem AutoML for Clustering Introduction
  • 5. ? The lack of standardized benchmarking criteria for AutoML in clustering is the primary motivation for this research ? Evaluating existing AutoML frameworks for unsupervised learning remains an open issue ? Benchmarking is essential for users to navigate the AutoML landscape effectively ? Standardized benchmarks promote transparency, replicability, and future research ? The proposed benchmarking criteria only consider approaches that support a working repository of code for reproducibility ? AutoML4Clust, cSmartML, Autocluster, and ML2DAC AutoML for Clustering Why benchmarking
  • 6. ? General Comparison ? Compares and contrasts design principles, parallelization capabilities, and clustering algorithm diversity ? Highlights unique features of each AutoML framework ? Clustering Quality ? Uses specific Cluster Validity Indices (CVI) to measure clustering quality and similarity ? CVI metrics categorized into internal and external ? Scalability & Consistency ? Evaluates AutoML scalability and consistency across datasets with varying dimensions, instances, and noise Benchmark Design Evaluation Criteria
  • 7. ? Experiments were conducted on a high-end workstation ? Isolated environment for each framework ? Employed Neptune.AI for collecting meta-data ? Set a time limit of 5 hours for each experiment ? RAM constraint settled up to 16GiB ? User-intent settled to maximize the quality of the result HW Configuration CPU Intel Core i9-10980XE GPU NVIDIA Quadro T1000 Mobile RAM 62.5 GiB Storage SK Hynix PC801 NVMe 1TB OS Ubuntu 22.04.2 LTS Benchmark Design HW & Framework Configurations
  • 8. Feature Range Dimension 2-99 Cluster 2-21 Samples 200-4000 Overlapness 1e-6 - 1e-5 Aspect Ref. 1-5 Imbalance Ratio 1-3 Benchmark Design Datasets ? 100 synthetic datasets generated using Repliclust
  • 9. AutoML4CLust Autocluster cSmartML ML2DAC Release Year 2021 2019 2021 2023 Stars on Github 2 50 0 0 Last Commit Oct. 13, 2020 Jan. 9, 2023 Dec. 27, 2021 Sept. 18, 2023 Optimization BO, RO Hyp, BOHB BO, RO Evo BO Task CASH CASH AS->HPO Meta-CVI Cash Meta-CVI Metalearning No Yes Yes Yes Clustering Algorithms 4 10 8 9 Results General Comparison
  • 10. ? Parallelization features: ? AutoML4Clust ? Supports concurrent optimization tasks with BOHB and HyperBand ? cSmartML ? Excels with internal parallelization for hyperparameter optimization ? Autocluster ? Lacks parallelization during execution ? ML2DAC ? Employs Dask, an open-source Python library for parallel computing. ? Not implemented using BO Results General Comparison
  • 11. Results General Comparison ? The adaptability of AutoML frameworks significantly impacts clustering effectiveness ? Search Space features: ? ML2DAC and Autocluster offer a wider search space range for AS ? AutoML4Clust’s search space comprises solely K-centered clustering algorithms ? cSmartML prioritizes hierarchy models with some density and graph theory adoption.
  • 12. Results Clustering Quality ? Silhouette - Internal CVI ? Evaluates cohesion and separation of clusters. ? Davies-Bouldin - Internal CVI ? Measures the average similarity between each cluster and its most similar cluster. ? Adjusted Rand Index - External CVI ? Measures the similarity between the clustering result and the ground truth
  • 13. (c) CD Plot considering ARI as reference metric. (b) CD Plot considering DBS as reference metric. ? Performed the non-parametric Friedman test ? Conducted the Nemenyi post-hoc test to identify significant differences between pairs ? Critical Difference (CD) displays the average rank of each framework ? statistically indistinguishable frameworks are connected (a) CD Plot considering SIL as reference metric. Results Statistical Analysis
  • 14. ? Performed the Bayesian Bradley-Terry model (BBT) to compare frameworks performance. ? BBT model assigns each framework a "merit number" determining its performance. ? The Bayesian approach enables the determination of probabilities associated with framework rankings. ? Provides confidence beyond simple ranking, enhancing decision-making. Results Bayesian Bradley-Terry Tree Analysis
  • 15. ? BBT procedure results contrast previous analysis ? Silhouette ? ML2DAC superior to both Autocluster and AutoML4Clust ? Autocluster outperforms AutoML4Clust ? Davies-Boudlin ? ML2DAC and Autocluster superior to other frameworks ? No claim can be made between ML2DAC and Autocluster ? Adjustet Rand Index ? ML2DAC superior to both Autocluster and AutoML4Clust ? cSmartML is confirmed as inferior to other analyzed frameworks Results Bayesian Bradley-Terry Tree
  • 16. Results Scalability & Consistency ? cSmartML exhibits an exponential growth across dataset instances ? AutoML4Clust exhibits a linear growth ? Autocluster 2x faster than cSmartML ? ML2DAC 9x faster than cSmartML
  • 17. Results Scalability & Consistency ? Autocluster is the most resource-efficient option ? ML2DAC requires approximately 40.51% more memory ? with higher performance in clustering quality ? cSmartML requires approximately 50.4% more memory ? with lower performance in clustering quality
  • 18. ? ML2DAC emerged slightly ahead ? ML2DAC excelled in multiple tasks but was not consistently the top performer ? Rapid development within the AutoML community ? Significant progress in unsupervised learning, especially in clustering ? Room for enhancement ? Meta-learning approaches may offer solutions to improve performance. ? Balancing automation with transparency is crucial in order to maintain accessibility to clustering algorithm principles. Conclusion
  • 19. Thank you for your attention !
  • 20. ? Z?ller, M., & Huber, M.F. (2019). Benchmark and Survey of Automated Machine Learning Frameworks. J. Artif. Intell. Res., 70, 409-472. ? Gijsbers, P., Bueno, M.L., Coors, S., LeDell, E., Poirier, S., Thomas, J., Bischl, B., & Vanschoren, J. (2022). AMLB: an AutoML Benchmark. ArXiv, abs/2207.12560. ? Luxburg, U.V., Williamson, R.C., & Guyon, I. (2009). Clustering: Science or Art? ICML Unsupervised and Transfer Learning. ? Wainer, J. (2022). A Bayesian Bradley-Terry model to compare multiple ML algorithms on multiple data sets. ArXiv, abs/2208.04935. References
  • 21. ? Tschechlov, D., Fritz, M., & Schwarz, H. (2021). AutoML4Clust: Efficient AutoML for Clustering Analyses. International Conference on Extending Database Technology. ? Shawi, R.E., & Sakr, S. (2022). cSmartML-Glassbox: Increasing Transparency and Controllability in Automated Clustering. 2022 IEEE International Conference on Data Mining Workshops (ICDMW), 47- 54. ? Treder-Tschechlov, D., Fritz, M., Schwarz, H., & Mitschang, B. (2023). ML2DAC: Meta-Learning to Democratize AutoML for Clustering Analysis. Proceedings of the ACM on Management of Data, 1, 1 - 26. ? Wong, W. Y. (2019). Autocluster: AutoML for clustering models in sklearn. References