Template talk:Machine learning
This is the talk page for discussing improvements to the Machine learning template. |
|
This template does not require a rating on Wik.ipedia.Pro's content assessment scale. It is of interest to the following WikiProjects: | ||||||||||||||||||||||
|
"Models"
This section title and contents seem pretty much random to me. How are contents chosen? One regression, one random clustering algorithm, 4 standard classificators; but no decision tree; which is probably the grandfather of all classificators. --Chire (talk) 12:41, 22 October 2013 (UTC)
In general, one may argue that k-means is NOT machine learning, but plain old statistics. And clustering is at most a step child of the machine learning world; it's a data mining / knowledge discovery domain, just like outlier detection and freuqent itemset mining. If you look at the communities, I would not call data mining part of machine learning either; it's living in parallel (unfortunately). Machine learners don't get or like unsupervised methods, actually. The "theory" section in this template is also pretty random, isn't it? --Chire (talk) 12:45, 22 October 2013 (UTC)
- This template is brand new and very incomplete. You're welcome to add it. k-means clustering is a very widely employed method in the machine learning community, e.g. by computer vision folks who use it as a feature learning method, by neural nets folks for booststrapping their RBF networks and by text mining people. New papers employing or improving k-means appear regularly in the ML literature. I can dig up some references if you like.
- AFAIC, the sidebar can be renamed something like "Data mining/machine learning/pattern recognition" -- the three overlap to such a degree that they're impossible to demarcate. QVVERTYVS (hm?) 14:59, 22 October 2013 (UTC)
- Re: "one regression algorithm": wrong. Logistic regression is in fact a classification algorithm. It is very popular in esp. the natural language processing community and form the basis for much recent neural nets and structured prediction work. Neural nets, k-NN and SVMs are all used for regression, though, even if this is not reflected in their Wik.ipedia.Pro articles. QVVERTYVS (hm?) 15:04, 22 October 2013 (UTC)
- I agree that they are hard to separate and it thus may be a good idea to merge them into one template. I know that k-means is used a lot in machine learning, as it is a statistical optimization problem; not so much actually a structure discovery thing. Maybe instead of the "Models" block, make one for each "Problem" above then? I.e. regression, classification, clustering, anomaly detection, etc.? --Chire (talk) 09:05, 23 October 2013 (UTC)
Maybe we need to add Markovian models?
Hidden Markov Models (HMM) has successfully been used (there are dozens are articles, just do a Google scholar search) where HMM have been used for NLP amongst other Machine learning tasks. I believe it should be added as one of the models. — Preceding unsigned comment added by 150.135.223.128 (talk) 19:12, 28 January 2014 (UTC)
- I've added CRFs, HMMs and a link to the more general article graphical model. QVVERTYVS (hm?) 22:00, 28 January 2014 (UTC)
Two problems are the same: "classification" and "clustering"
there are simply two general approaches to solve the same problem, supervised and unsupervised -- but the problem is one and the same. Fgnievinski (talk) 23:55, 3 May 2014 (UTC)
- Applications are also overlapping if not coincident. Fgnievinski (talk) 12:32, 5 May 2014 (UTC)
- As discussed in Talk:Statistical classification#Terminology: "classification" is supervised, "clustering" is unsupervised -- Really?, I disagree that they are the same thing.
- The objectives are different in the sense that classification tries to minimize the prediction error. Clustering however tries to discover some meaningful structure, without knowing what to look out for (which is also why clustering more often than not returns crap results - too little guideance on what you are looking for). They are related, but clearly not the same thing. IMHO, the applications as well as the methods differ fundamentally, too. You can't easily take one method and transfer it to the other problem; not even naive bayes, or kNN classification. There are some cases where you have similar ideas - k-means also minimizes squared errors - but these occur in many other areas, too. And there are many clustering approaches not based on minimizing some statistical quantity. The big problem with clustering is evaluation: usually you evaluate by some statistical quantity (internal), or by class labels (external); both of which look a lot like classification.
- Either way; we are not truth finders. There is plenty of literature that distinguishes these approaches, so we should not merge them. The rule of thumb in literature is that classification and regression are supervised, and this is well resembled by the ML bar template. --Chire (talk) 13:36, 5 May 2014 (UTC)
- I'm sorry, this is not a restatement of the previous talk. The methods are outside the scope of the present discussion. What is inside the scope is that both methodological approaches aim to cluster, group, segment, partition, and classify input variates. All the problems addressed by unsupervised methods could be tackled by supervised ones if additional information is given. Fgnievinski (talk) 13:48, 5 May 2014 (UTC)
- I also disagree on that. If you added labels to a data set, it would become a different problem: how to predict the labels of new instances, given the training data set, i.e. it becomes class prediction, whereas it was structure discovery before. That is IMHO a quite different task. --Chire (talk) 13:57, 5 May 2014 (UTC)
- The algorithms and performance metrics for clustering are radically different from those for classification, which is enough reason not to conflate them. I also challenge the statement that the applications are "coincident". QVVERTYVS (hm?) 14:50, 5 May 2014 (UTC)
- I also disagree on that. If you added labels to a data set, it would become a different problem: how to predict the labels of new instances, given the training data set, i.e. it becomes class prediction, whereas it was structure discovery before. That is IMHO a quite different task. --Chire (talk) 13:57, 5 May 2014 (UTC)
- I'm sorry, this is not a restatement of the previous talk. The methods are outside the scope of the present discussion. What is inside the scope is that both methodological approaches aim to cluster, group, segment, partition, and classify input variates. All the problems addressed by unsupervised methods could be tackled by supervised ones if additional information is given. Fgnievinski (talk) 13:48, 5 May 2014 (UTC)
Reenforcement learning & Terminology
Seems confusing for an outsider to call it 'supervised learning' but then not talk about unsupervised learning or reenforcement ? Not sure what the best approach here would be, since clustering is (in some way) unsupervised learning -- would a rename be warranted? Perhaps worth renaming to "Unsupervised Learning / Clustering" ? Dm1911 (talk) 17:50, 27 May 2015 (UTC)
- Reinforcement learning is currently missing, we should add it. Unsupervised learning is much broader than clustering: it also encompasses dimensionality reduction and feature learning. QVVERTYVS (hm?) 18:46, 27 May 2015 (UTC)
- Added Reinforcement learning as per this discussion. Situphobos (talk) 07:34, 4 July 2016 (UTC)
How about adding list of datasets for machine learning research to the bar? Any thoughts on this? --Datakeeper (talk) 19:04, 25 February 2016 (UTC)
Collapsable version
Is there a way to make this template have collapsable sections? It's pretty large and a little unwieldy to put on pages. --Datakeeper (talk) 21:15, 22 February 2016 (UTC)
- Good point. Made it collapsible. QVVERTYVS (hm?) 20:15, 25 February 2016 (UTC)
- @Qwertyus: Excellent - looks great! Thank you!--Datakeeper (talk) 21:17, 25 February 2016 (UTC)
spam
I see this all over the place as intro diagram and it is not helpful. Diagrams about the topic instead of machine learning template on every subtopic would be better. Daniel.Cardenas (talk) 23:03, 1 May 2016 (UTC)
- Suggest putting after intro or in other words the first topic. This will encourage others to create a more specific and more helpful diagram for intro. Daniel.Cardenas (talk) 00:08, 2 May 2016 (UTC)
Per WP:NAVBOX: "The collection of articles in a sidebar template should be fairly tightly related... If the articles are not tightly related, a footer template (a navbox, located at the bottom of the article) may be more appropriate.". This definitely applies here. I cannot see this template being much use to anybody for navigation, and certainly not one of the first things a reader will look for on arriving on an article. I think conversion to a footer would be ideal. Bigbluefish (talk) 10:18, 13 February 2018 (UTC)
Existing template image is not illustrative
@Snus-kin: Hi, template image now is
File:Multi-Layer_Neural_Network-Vector-Blank.svg
I think this image is not conveying any helpful information about the concept of "Machine Learning". I proposed to use this image:
File:Machine Learning Icon.jpg
which is now reverted, but this image is part of the cover of this book: https://books.google.com/books?id=Ex8_tAEACAAJ&source=gbs_book_other_versions , it is a robot that is graduated from some knowledge course. However, existing template image, which is a graph (neural networks), is not suitable for this purpose and a regular person does not understand "Machine Learning" or "Data Mining" from it. @Qwertyus: Any idea? Thanks, Hooman Mallahzadeh (talk) 08:15, 28 September 2021 (UTC)
- Hi Hooman, I think if we were to change the image we should aim to have something of a high quality, perhaps drawn in SVG, currently this image is rough and pixelated. I'd also be careful about recreating images from books as they are typically copywritten material, but I am not an expert on this. Snus-kin (talk) 09:15, 28 September 2021 (UTC)
- I feel the current image File:Kernel Machine.svg is both distracting and unhelpful. I would support an improvement. The grey neural network File:Multi-Layer_Neural_Network-Vector-Blank.svg one above was better. The owl with mortarboard isn't to my taste. Chumpih t 08:28, 5 August 2023 (UTC), tweaked 2023-08-08.
- Aaand I see this has now (2023-08-08) been changed to File:Neural network with dark background.png, which is better than the Kernel Machine classification one in my eyes, but I would still prefer File:Multi-Layer_Neural_Network-Vector-Blank.svg. Then again, de gustibus. Chumpih t 00:37, 8 August 2023 (UTC)
- There are pros and cons for each. The one that you prefer (the one with the neurons being emphasized as big dark grey circles and the arrows illustrating the inputs for each layer) illustrates the basic blueprint introduced forty years ago (since then known under a misnomer, multilayer perceptron.), but it does not show any progress made (i.e. stochastic gradient descent introduced in 1969; backpropagation method used in 1970; different activation function used; transformer architectures used since 2018, etc). I'm not fan of the "File:Kernel Machine.svg" either, but at least it is about the (somewhat modern) 1990s kernel trick. I was thinking about all that, but the only low-hanging fruit that I can see is a large size, i.e. the largeness itself characterizing large language models, for example. DancingPhilosopher (talk) 09:57, 8 August 2023 (UTC)
- The new one is much better. I was always distracted by that old thumbnail, as my brain would briefly try to understand the ML article as if the image was a visual representation of the concepts of the article. 100DashSix (talk) 17:00, 17 August 2023 (UTC)
- There are pros and cons for each. The one that you prefer (the one with the neurons being emphasized as big dark grey circles and the arrows illustrating the inputs for each layer) illustrates the basic blueprint introduced forty years ago (since then known under a misnomer, multilayer perceptron.), but it does not show any progress made (i.e. stochastic gradient descent introduced in 1969; backpropagation method used in 1970; different activation function used; transformer architectures used since 2018, etc). I'm not fan of the "File:Kernel Machine.svg" either, but at least it is about the (somewhat modern) 1990s kernel trick. I was thinking about all that, but the only low-hanging fruit that I can see is a large size, i.e. the largeness itself characterizing large language models, for example. DancingPhilosopher (talk) 09:57, 8 August 2023 (UTC)
- Aaand I see this has now (2023-08-08) been changed to File:Neural network with dark background.png, which is better than the Kernel Machine classification one in my eyes, but I would still prefer File:Multi-Layer_Neural_Network-Vector-Blank.svg. Then again, de gustibus. Chumpih t 00:37, 8 August 2023 (UTC)
See what we do next...
OR
By submitting your email or phone number, you're giving mschf permission to send you email and/or recurring marketing texts. Data rates may apply. Text stop to cancel, help for help.
Success: You're subscribed now !