True, high-size data cannot be easily visualized in Euclidean high-dimensional data, but it is not true that there are no visualization methods for them.
In addition to this statement, I will add that with only four functions (your measurements) you can easily try the parallel coordinates of the visualization method . Or just try multivariate data analysis using two functions at a time (6 times in total) to try to figure out what kind of relationship is there between the two (correlation and dependence in general). Or you can even use three-dimensional space three at a time.
Then, how to get information from these visualizations? Well, it's not as simple as in Euclidean space, but the point of view is visually displayed if the data cluster in some groups (for example, next to some values ββon the axis for a parallel coordinate diagram) and thinks that the data is somehow separable (for example, if it forms areas like circles or lines shared on scatterplots).
A small digression: the diagram you published does not indicate the power or capabilities of each algorithm, taking into account some specific data distributions, it simply emphasizes the nature of some algorithms: for example, the k-tool allows you to separate only convex and ellipsoidal regions (and keep in mind that bulge and ellipsoids exist even in the nth dimension). I mean, there is no rule that says: given the distributions shown in this diagram, you must choose the right clustering algorithm.
I suggest using a data mining toolbar that allows you to examine and visualize data (and easily transform it, since you can change your topology with transformations, forecasts and abbreviations, check another lejlot answer for this), for example Weka (plus you no need to run all the algorithms yourself.
In the end, I will show you this resource for various parameters of the Q factor and suitability of the cluster so that you can compare the results with other algorithms.
source share