User:Ian Helmke/Sandbox: Difference between revisions
imported>Ian Helmke mNo edit summary |
imported>Ian Helmke mNo edit summary |
||
Line 1: | Line 1: | ||
= Clustering Additions = | = Clustering Additions = | ||
This is text I'd like to add to the [[machine learning]] page (pending heavy editing). This text would go underneath the "Issues in Training and Evaluation" section. | This is text I'd like to add to the [[machine learning]] page (pending | ||
heavy editing). This text would go underneath the "Issues in Training | |||
and Evaluation" section. | |||
== Overfitting == | == Overfitting == | ||
Overfitting occurs when a classifier trains so closely to model data | Overfitting occurs when a classifier trains so closely to model data | ||
that it is not useful for classifying things outside of the | that it is not useful for classifying things outside of the model. | ||
model | |||
This can sometimes occur because a classifier looks at irrelevant data | This can sometimes occur because a classifier looks at irrelevant data | ||
Line 15: | Line 14: | ||
features seen less commonly in the training data, and if data given to | features seen less commonly in the training data, and if data given to | ||
the classifier shares similar uncommon values, it is grouped | the classifier shares similar uncommon values, it is grouped | ||
accordingly. | accordingly. As a result, the classifier is extremely accurate when classifying the | ||
training data and appears to be useful, but when it is | |||
used to organize data outside of the model, it becomes inaccurate. | |||
Machine learning techniques can prevent overfitting to some extent by | Machine learning techniques can prevent overfitting to some extent by | ||
imposing a penalty upon itself for complicated models. A | |||
more consistent with the data in question. | simpler model is oftentimes more consistent with the trends of the | ||
data in question. | |||
== Imbalance of Data == | == Imbalance of Data == | ||
Line 33: | Line 35: | ||
One way to prevent this imbalance of data is to use a technique called | One way to prevent this imbalance of data is to use a technique called | ||
active example selection. Active example selection builds the | active example selection<ref>{{cite journal | ||
|author=Sangyoon, Oh et al. | |||
|title=Ensemble Learning with Active Example Selection for Imbalanced Biomedical Data Classification | |||
|journal=IEEE/ACM Trans. Comput. Biol. Bioinformatics | |||
|publisher=IEEE Computer Society Press | |||
|year=2011 | |||
|id=http://dx.doi.org/10.1109/TCBB.2010.96}}</ref>. | |||
Active example selection builds the | |||
classifier model slowly, adding a few pieces of sample data of each | classifier model slowly, adding a few pieces of sample data of each | ||
class at a time. The model is tested at every stage, and documents | class at a time. The model is tested at every stage, and documents | ||
Line 39: | Line 48: | ||
while ones that do not change the model or make its performance worse | while ones that do not change the model or make its performance worse | ||
are removed. This ensures that only meaningful pieces of data are used | are removed. This ensures that only meaningful pieces of data are used | ||
in the classifier. | in the classifier. | ||
== Evaluating Results == | == Evaluating Results == | ||
Line 54: | Line 63: | ||
machine learning algorithm is compared to a set of data classified by | machine learning algorithm is compared to a set of data classified by | ||
experts. In this case, the machine learning algorithm is given a set | experts. In this case, the machine learning algorithm is given a set | ||
of training data, and then classifies a second sample set of data. A group of | of training data, and then classifies a second sample set of data. A | ||
experts also annotate this second set of data, marking it according to how | group of experts also annotate this second set of data, marking it | ||
they believe it should be classified (not according to how they | according to how they believe it should be classified (not according | ||
believe a machine would classify it). This training and evaluation | to how they believe a machine would classify it). This training and | ||
data is generally a tiny subset of the data available. If a classifier | evaluation data is generally a tiny subset of the data available. If a | ||
is able to produce good results for a subset of data, it should also | classifier is able to produce good results for a subset of data, it | ||
be successful at classifying a larger set of similar data. | should also be successful at classifying a larger set of similar data. | ||
The results of the algorithm are compared to the results of the | The results of the algorithm are compared to the results of the | ||
experts and arranged into two scores: precision and recall. Precision | experts and arranged into two scores: precision and recall. Precision | ||
Line 83: | Line 92: | ||
the ability of an algorithm to be able to run in parallel and scale | the ability of an algorithm to be able to run in parallel and scale | ||
accordingly (so that an algorithm running on two cores runs twice as | accordingly (so that an algorithm running on two cores runs twice as | ||
fast, for example) allows it to | fast, for example) allows it to run faster on modern hardware, so that | ||
run faster on modern hardware, so that larger quantities of data can | larger quantities of data can be processed. | ||
be processed. | |||
= Biclustering = | = Biclustering = | ||
The following material would go on a new page of the above title (probably linked to the ML page) | ''The following material would go on a new page of the above title | ||
(probably linked to the ML page)'' | |||
'''Biclustering''' is an unsupervised [[machine learning]] method which searches | '''Biclustering''' is an unsupervised [[machine learning]] method | ||
for similarities in specific subsections of the input data. It is unique among | which searches for similarities in specific subsections of the input | ||
machine learning methods because it searches for similarities in small | data. It is unique among machine learning methods because it searches | ||
parts of the data, instead of putting a piece of data into a single | for similarities in small parts of the data, instead of putting a | ||
group. | piece of data into a single group. | ||
Biclustering was first discovered in 1970. Today, it is a commonly | Biclustering was first discovered in 1970. Today, it is a commonly | ||
Line 104: | Line 113: | ||
== Processing == | == Processing == | ||
Clustering algorithms take [[vector|vectors]] as input. They sort the vectors | Clustering algorithms take [[vector|vectors]] as input. They sort the | ||
according to how similar they are by comparing all of the features (data) in | vectors according to how similar they are by comparing all of the | ||
the vector, and the result of the clustering is several groups that | features (data) in the vector, and the result of the clustering is | ||
each contain a bunch of (hopefully) similar vectors. Biclustering | several groups that each contain a bunch of (hopefully) similar | ||
looks at all of the vectors as a single input matrix, and attempts to | vectors. Biclustering looks at all of the vectors as a single input | ||
find regions of the input which look similar. | matrix, and attempts to find regions of the input which look similar. | ||
Biclustering is a useful technique for finding trends in data when | Biclustering is a useful technique for finding trends in data when | ||
Line 122: | Line 131: | ||
Biclustering is particularly useful in the medical field, where it can | Biclustering is particularly useful in the medical field, where it can | ||
be used, for example, to find genes related to a specific disease in a | be used, for example, to find genes related to a specific disease in a | ||
group of patients. If each vector represents how a person expresses | group of patients<ref>{{cite journal | ||
|author=Still, Martin et al. | |||
|title=Robust biclustering by sparse singular value decomposition incorporating stability selection | |||
|publisher=Oxford University Press | |||
|journal=Bioinformatics | |||
|year=2011 | |||
|id=http://dx.doi.org/10.1093/bioinformatics/btr322}}</ref>. | |||
If each vector represents how a person expresses | |||
traits, biclustering can be used to determine a set of genes which is | traits, biclustering can be used to determine a set of genes which is | ||
associated with cancer. It can even be used to find similarities and | associated with cancer. It can even be used to find similarities and | ||
differences between different varieties of cancers. | differences between different varieties of cancers. | ||
== References == | |||
<references /> |
Revision as of 17:53, 1 August 2011
Clustering Additions
This is text I'd like to add to the machine learning page (pending heavy editing). This text would go underneath the "Issues in Training and Evaluation" section.
Overfitting
Overfitting occurs when a classifier trains so closely to model data that it is not useful for classifying things outside of the model.
This can sometimes occur because a classifier looks at irrelevant data points in the training data. Classifiers often give more weight to features seen less commonly in the training data, and if data given to the classifier shares similar uncommon values, it is grouped accordingly. As a result, the classifier is extremely accurate when classifying the training data and appears to be useful, but when it is used to organize data outside of the model, it becomes inaccurate.
Machine learning techniques can prevent overfitting to some extent by imposing a penalty upon itself for complicated models. A simpler model is oftentimes more consistent with the trends of the data in question.
Imbalance of Data
At times, data within training sets is imbalanced, where one sample category has many more points of data than another. Oftentimes, the category with less available data is the more interesting one, and we want to find characteristics that are common to the minority group that are also not shared by the majority group. Classifiers assume that the ratio of the different categories of training data that they receive are roughly equivalent to the ratio of real data that will belong in each category.
One way to prevent this imbalance of data is to use a technique called active example selection[1]. Active example selection builds the classifier model slowly, adding a few pieces of sample data of each class at a time. The model is tested at every stage, and documents which improve the accuracy of the model are kept in the classifier, while ones that do not change the model or make its performance worse are removed. This ensures that only meaningful pieces of data are used in the classifier.
Evaluating Results
There are a number of techniques for evaluating the results of a machine learning algorithm. Some of these techniques are also used in natural-language processing. Machine learning techniques are generally evaluated by their results, and some methods (such as neural networks) are considered "black box" forms of classification, since it is not easy to understand how or why the underlying implementation is sorting the way it is.
In some cases, particularly with classifiers, the outcome of the machine learning algorithm is compared to a set of data classified by experts. In this case, the machine learning algorithm is given a set of training data, and then classifies a second sample set of data. A group of experts also annotate this second set of data, marking it according to how they believe it should be classified (not according to how they believe a machine would classify it). This training and evaluation data is generally a tiny subset of the data available. If a classifier is able to produce good results for a subset of data, it should also be successful at classifying a larger set of similar data. The results of the algorithm are compared to the results of the experts and arranged into two scores: precision and recall. Precision accounts for situations where the classifier put something in a category where it did not belong. Recall accounts for situations where the classifier did not put something in a category it should have.
Classification and clustering algorithms can also be measured against other algorithms. This is useful when an algorithm is attempting to improve performance (speedwise, for example) while providing similar levels of precision and recall relative to another algorithm. It can also be used to show that an algorithm is an improvement over a previous generation, or to show which algorithm is most useful for organizing data for a particular problem.
Scalability
The ability of machine learning algorithms to take advantage of modern computers, which tend to have the ability to multitask exceptionally well through the use of multiple cores or CPUs, is important. Since these algorithms are often used to process large quantities of data, the ability of an algorithm to be able to run in parallel and scale accordingly (so that an algorithm running on two cores runs twice as fast, for example) allows it to run faster on modern hardware, so that larger quantities of data can be processed.
Biclustering
The following material would go on a new page of the above title (probably linked to the ML page)
Biclustering is an unsupervised machine learning method which searches for similarities in specific subsections of the input data. It is unique among machine learning methods because it searches for similarities in small parts of the data, instead of putting a piece of data into a single group.
Biclustering was first discovered in 1970. Today, it is a commonly used technique in bioinformatics, particularly in the area of gene expression, or identifying groups of genes that are similar between different people.
Processing
Clustering algorithms take vectors as input. They sort the vectors according to how similar they are by comparing all of the features (data) in the vector, and the result of the clustering is several groups that each contain a bunch of (hopefully) similar vectors. Biclustering looks at all of the vectors as a single input matrix, and attempts to find regions of the input which look similar.
Biclustering is a useful technique for finding trends in data when each vector of data is very large because it can spot trends in specific parts of data that clustering cannot. A normal clustering algorithm sorts pieces of data according to features that the majority of them share. Biclustering is able to organize data according to parts of them that seem similar.
Applications
Biclustering is particularly useful in the medical field, where it can be used, for example, to find genes related to a specific disease in a group of patients[2]. If each vector represents how a person expresses traits, biclustering can be used to determine a set of genes which is associated with cancer. It can even be used to find similarities and differences between different varieties of cancers.
References
- ↑ Sangyoon, Oh et al. (2011). "Ensemble Learning with Active Example Selection for Imbalanced Biomedical Data Classification". IEEE/ACM Trans. Comput. Biol. Bioinformatics. http://dx.doi.org/10.1109/TCBB.2010.96.
- ↑ Still, Martin et al. (2011). "Robust biclustering by sparse singular value decomposition incorporating stability selection". Bioinformatics. http://dx.doi.org/10.1093/bioinformatics/btr322.