TY - GEN
T1 - Quantifying Nondeterminism and Inconsistency in Self-organizing Map Implementations
AU - Rahaman, Sydur
AU - Samuel, Raina
AU - Neamtiu, Iulian
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021/8
Y1 - 2021/8
N2 - Self-organizing maps (SOMs) are a popular approach for neural network-based unsupervised learning. However the reliability of self-organizing map implementations has not been investigated. Using internal and external metrics, we define and check two basic SOM properties. First, determinism: A given SOM implementation should produce the same SOM when run repeatedly on the same training dataset. Second, consistency: Two SOM implementations should produce similar SOMs when presented with the same training dataset. We check these properties in four popular SOM implementations. We ran our approach on 381 popular datasets used in health, medicine, and other critical domains. We found that implementations violate these basic properties. For example, 375 out of 381 datasets have nondeterministic outcomes; for 51-92% of datasets, toolkits yield significantly different SOM clusterings; and clustering accuracy might be so inconsistent as to vary by a factor of four between toolkits. This undermines SOM reliability, and the reliability of results obtained via SOMs. Our study shines a light on what to expect, in practice, when running actual SOM implementations. Our findings suggest that for critical applications, SOM users should not take reliability for granted; rather, multiple runs and different toolkits should be considered and compared.
AB - Self-organizing maps (SOMs) are a popular approach for neural network-based unsupervised learning. However the reliability of self-organizing map implementations has not been investigated. Using internal and external metrics, we define and check two basic SOM properties. First, determinism: A given SOM implementation should produce the same SOM when run repeatedly on the same training dataset. Second, consistency: Two SOM implementations should produce similar SOMs when presented with the same training dataset. We check these properties in four popular SOM implementations. We ran our approach on 381 popular datasets used in health, medicine, and other critical domains. We found that implementations violate these basic properties. For example, 375 out of 381 datasets have nondeterministic outcomes; for 51-92% of datasets, toolkits yield significantly different SOM clusterings; and clustering accuracy might be so inconsistent as to vary by a factor of four between toolkits. This undermines SOM reliability, and the reliability of results obtained via SOMs. Our study shines a light on what to expect, in practice, when running actual SOM implementations. Our findings suggest that for critical applications, SOM users should not take reliability for granted; rather, multiple runs and different toolkits should be considered and compared.
KW - AI reliability
KW - AI testing
KW - Self-organizing maps
KW - neural networks
KW - nondeterminism
KW - validation
UR - http://www.scopus.com/inward/record.url?scp=85118797852&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85118797852&partnerID=8YFLogxK
U2 - 10.1109/AITEST52744.2021.00026
DO - 10.1109/AITEST52744.2021.00026
M3 - Conference contribution
AN - SCOPUS:85118797852
T3 - Proceedings - 3rd IEEE International Conference on Artificial Intelligence Testing, AITest 2021
SP - 85
EP - 92
BT - Proceedings - 3rd IEEE International Conference on Artificial Intelligence Testing, AITest 2021
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 3rd IEEE International Conference on Artificial Intelligence Testing, AITest 2021
Y2 - 23 August 2021 through 26 August 2021
ER -