{"date_created":"2018-12-11T12:07:04Z","publication":"Journal of Classification","date_updated":"2022-01-27T14:16:27Z","article_type":"original","month":"01","user_id":"ea97e931-d5af-11eb-85d4-e6957dddbf17","title":"Efficient algorithms for agglomerative hierarchical clustering methods","publisher":"Springer","article_processing_charge":"No","page":"7 - 24","doi":"10.1007/BF01890115","status":"public","main_file_link":[{"url":"https://link.springer.com/article/10.1007%2FBF01890115"}],"publication_status":"published","day":"01","year":"1984","quality_controlled":"1","volume":1,"extern":"1","intvolume":" 1","date_published":"1984-01-01T00:00:00Z","oa_version":"None","citation":{"ieee":"W. Day and H. Edelsbrunner, “Efficient algorithms for agglomerative hierarchical clustering methods,” Journal of Classification, vol. 1. Springer, pp. 7–24, 1984.","ista":"Day W, Edelsbrunner H. 1984. Efficient algorithms for agglomerative hierarchical clustering methods. Journal of Classification. 1, 7–24.","ama":"Day W, Edelsbrunner H. Efficient algorithms for agglomerative hierarchical clustering methods. Journal of Classification. 1984;1:7-24. doi:10.1007/BF01890115","mla":"Day, William, and Herbert Edelsbrunner. “Efficient Algorithms for Agglomerative Hierarchical Clustering Methods.” Journal of Classification, vol. 1, Springer, 1984, pp. 7–24, doi:10.1007/BF01890115.","short":"W. Day, H. Edelsbrunner, Journal of Classification 1 (1984) 7–24.","apa":"Day, W., & Edelsbrunner, H. (1984). Efficient algorithms for agglomerative hierarchical clustering methods. Journal of Classification. Springer. https://doi.org/10.1007/BF01890115","chicago":"Day, William, and Herbert Edelsbrunner. “Efficient Algorithms for Agglomerative Hierarchical Clustering Methods.” Journal of Classification. Springer, 1984. https://doi.org/10.1007/BF01890115."},"publist_id":"1998","abstract":[{"lang":"eng","text":"Whenevern objects are characterized by a matrix of pairwise dissimilarities, they may be clustered by any of a number of sequential, agglomerative, hierarchical, nonoverlapping (SAHN) clustering methods. These SAHN clustering methods are defined by a paradigmatic algorithm that usually requires 0(n 3) time, in the worst case, to cluster the objects. An improved algorithm (Anderberg 1973), while still requiring 0(n 3) worst-case time, can reasonably be expected to exhibit 0(n 2) expected behavior. By contrast, we describe a SAHN clustering algorithm that requires 0(n 2 logn) time in the worst case. When SAHN clustering methods exhibit reasonable space distortion properties, further improvements are possible. We adapt a SAHN clustering algorithm, based on the efficient construction of nearest neighbor chains, to obtain a reasonably general SAHN clustering algorithm that requires in the worst case 0(n 2) time and space.\r\nWhenevern objects are characterized byk-tuples of real numbers, they may be clustered by any of a family of centroid SAHN clustering methods. These methods are based on a geometric model in which clusters are represented by points ink-dimensional real space and points being agglomerated are replaced by a single (centroid) point. For this model, we have solved a class of special packing problems involving point-symmetric convex objects and have exploited it to design an efficient centroid clustering algorithm. Specifically, we describe a centroid SAHN clustering algorithm that requires 0(n 2) time, in the worst case, for fixedk and for a family of dissimilarity measures including the Manhattan, Euclidean, Chebychev and all other Minkowski metrics."}],"author":[{"first_name":"William","full_name":"Day, William","last_name":"Day"},{"last_name":"Edelsbrunner","full_name":"Edelsbrunner, Herbert","first_name":"Herbert","id":"3FB178DA-F248-11E8-B48F-1D18A9856A87","orcid":"0000-0002-9823-6833"}],"type":"journal_article","language":[{"iso":"eng"}],"_id":"4121","publication_identifier":{"eissn":["1432-1343"],"issn":["0176-4268"]}}