I like to stop the execution when Batch SOM becomes converged. What error function can I use to determine the convergence?
When talking about convergence for SOMs, for a given map size (n x m), you want to know whether sufficient iterations of the algorithm have run to ensure the map is "Stable". This means, loosely speaking, do new inputs (observations) to the map get placed at the same neurons /codebook vectors if the map is retrained many times?(Ignoring the issue of the fact that the arrangement of the map may switch around when it is trained each time, which is fine as long as the clusters are still arranged in a stable way).
To assist in answering the question of whether enough iterations have run, see the academic papers listed below. Both papers also touch on the issue of what map size is appropriate (what n x m values help ensure convergence of the SOM?).
One of the traditional approaches that has been popular in papers is given here:
Statistical tools to assess the reliability of self-organizing maps (Bodt, Cottrell, Verleysen)
More recently, this method has come about, which looks rather promising:
This thesis, in my opinion, was really well written and a pleasure to read. What's also nice is that this research has been written up as a SOM convergence test in a (rather unknown) package in R, called popsom
. Check it out:
I'm pretty sure you mean cost function rather than error function.
SOM does not require an error function (nor a cost function).
At the top level of the Machine Learning taxonomy, SOM is an unsupervised learning technique--no target vector, and therefore no "target-vector" minus "value_at_the_current_iteration" to minimize.
Another way to think of it: The role of a cost function is to minimize some cost; in ML, it's the delta between model calculation and supplied data. In SOM, no data is supplied to the algorithm for this purpose.
(I realize that this is somewhat confusing because the input data from which the network is created is often referred to as "training data"--probably that's the role of input data in supervised ML techniques, which are far more common than unsupervised ones. It's probably also confusing because Teuvo Kohonen, the person credited with 'inventing' SOM, originally referred to them as a class of neural networks--and of course NN is a supervised technique and does rely on a cost function (often gradient descent.))
Finally, just to make sure, i checked my own SOM code as well as the code from the ML textbook by Marsland, "Machine Learning: An Algorithmic Perspective". In both my code and his, the only stopping criterion whatever value for "maximum iterations" the user passed in when he called the main function.
class
(basic version), kohonen
, popsom
... –
Swage When talking about convergence for SOMs, for a given map size (n x m), you want to know whether sufficient iterations of the algorithm have run to ensure the map is "Stable". This means, loosely speaking, do new inputs (observations) to the map get placed at the same neurons /codebook vectors if the map is retrained many times?(Ignoring the issue of the fact that the arrangement of the map may switch around when it is trained each time, which is fine as long as the clusters are still arranged in a stable way).
To assist in answering the question of whether enough iterations have run, see the academic papers listed below. Both papers also touch on the issue of what map size is appropriate (what n x m values help ensure convergence of the SOM?).
One of the traditional approaches that has been popular in papers is given here:
Statistical tools to assess the reliability of self-organizing maps (Bodt, Cottrell, Verleysen)
More recently, this method has come about, which looks rather promising:
This thesis, in my opinion, was really well written and a pleasure to read. What's also nice is that this research has been written up as a SOM convergence test in a (rather unknown) package in R, called popsom
. Check it out:
© 2022 - 2024 — McMap. All rights reserved.