Center for Machine Learning and Intelligent Systems
About  Citation Policy  Donate a Data Set  Contact


Repository Web            Google
View ALL Data Sets

MONK's Problems Data Set

Below are papers that cite this data set, with context shown. Papers were automatically harvested and associated with this data set, in collaboration with Rexa.info.

Return to MONK's Problems data set page.


Jianbin Tan and David L. Dowe. MML Inference of Decision Graphs with Multi-way Joins and Dynamic Attributes. Australian Conference on Artificial Intelligence. 2003.

Each 10-fold cross-validation consists of 10 tests. In each test, we trained on nine-tenths of the data and tested on the remaining one-tenth. This amounted to 10x10=100 tests. 1st monk s data set: The 1st monk's data set is in the UCI machine learning repository [10, 1], and constructed from the noiseless function ( Jacket Color = Red ) V ( Head Shape = body Shape ) 10 independent tests were


Wl/odzisl/aw Duch and Karol Grudzinski. Ensembles of Similarity-based Models. Intelligent Information Systems. 2001.

range. A single weight corresponding to a highly-ranked feature is fixed at 1 to establish an absolute scale for distances. First the ensemble selection method has been used with two artificial datasets, Monk 1 and Monk-3 [13]. These problems are designed for rule-based symbolic machine learning algorithms and the nearest neighbor algorithms usually do not work well in such cases. 6 symbolic


Mark A. Hall. Department of Computer Science Hamilton, NewZealand Correlation-based Feature Selection for Machine Learning. Doctor of Philosophy at The University of Waikato. 1999.

a one-third/two-thirds split on credit and one-eighth of the instances were used for training on mushroom (the largest 80 dataset). In the case of the Monk s problems, testing is performed on the full dataset (as was done originally by Thrun et al. [TBB + 91]). Various different train and test set sizes are used with the


Alexey Tsymbal and Seppo Puuronen and Vagan Y. Terziyan. Arbiter Meta-Learning with Dynamic Selection of Classifiers and Its Experimental Investigation. ADBIS. 1999.

from the UCI machine learning repository: three MONK s problem datasets donated by Sebastian Thrun and the Tic-Tac-Toe Endgame dataset donated by David W. Aha [8]. The MONK's problems are a collection of three artificial binary classification problems over the same


Blai Bonet and Hector Geffner. Learning Sorting and Decision Trees with POMDPs. ICML. 1998.

figures for ID3 and C4.5 were taken from (Friedman, Kohavi, & Yun 1996). The column named `Test' in the table indicates how the generalization performance of the algorithms was measured. The Monk n datasets come with separate training and test data; on the other two problems the test data was generated by 5-fold cross validation: the data were partitioned into five segments, and fives runs were


Jan C. Bioch and D. Meer and Rob Potharst. Bivariate Decision Trees. PKDD. 1997.

in which interactions between two variables occur. We will test our method on two artificial data sets. The first is the monk 1 data set [Thr91]. This data set contains 6 attributes and two classes. The rule that generates the data is: if (x 1 = x 2 or x 5 = 1) then yes else no. Note that x 5 takes


Geoffrey I. Webb. OPUS: An Efficient Admissible Algorithm for Unordered Search. J. Artif. Intell. Res. (JAIR, 3. 1995.

with respect to minimizing the number of nodes expanded under depth-first search. 458 An Efficient Admissible Algorithm for Unordered Search Nonetheless, for only one search task, the Monk 2 data set, does OPUS o explore more nodes under depth-first search (16,345) than an alternative (both no optimistic reordering and fixed-order search that explore 12,879 and 12,791 nodes respectively). These


Ron Kohavi. The Power of Decision Tables. ECML. 1995.

Breiman et al. (1984), Devijver & Kittler (1982)). The results demonstrate that IDTM can achieve high accuracy in discrete domains using the simple hypothesis space of DTMs. In corral, dna, the Monk Dataset Features sizes Accuracy Accuracy Accuracy Accuracy australian 14 690 CV 55.5Sigma2.3 85.4Sigma1.1 84.9Sigma 1.7 89.4Sigma1.3 breast 10 699 CV 65.5Sigma1.7 95.4Sigma0.7 90.6Sigma 0.9


Karol Grudzi nski and Wl/odzisl/aw Duch. SBL-PM: A Simple Algorithm for Selection of Reference Instances in Similarity Based Methods. Department of Computer Methods, Nicholas Copernicus University.

by other classification systems. Due to the noisy character of the data the limit in the leave-one-out or crossvalidation tests is about 98% [10]. Another set of experiments was done on the 3 Monk datasets [11]. On this artificial data SBM gives good results (100% of correct answers on the first problem, 85% on problem 2, and over 97% on problem 3) only if feature selection and/or weighting is


Wl/odzisl/aw Duch and Karol Grudzinski. Meta-learning: searching in the model space. Department of Computer Methods, Nicholas Copernicus University.

The goal of further search for the best model should therefore include not only accuracy but also reduction of variance, i.e. stabilization of the classifier. 4.1 Monk problems The artificial dataset Monk-1 [11] is designed for rule-based symbolic machine learning algorithms (the data was taken from the UCI repository [12]). The nearest neighbor algorithms usually do not work well in such cases.


Ron Kohavi and Brian Frasca. Useful Feature Subsets and Rough Set Reducts. the Third International Workshop on Rough Sets and Soft Computing.

tic-tac-toe, breast-cancer, chess, mushroom, vote, and vote1, Holte-II has an average accuracy of 93.6%, much better than C4.5's average accuracy of 82.2%. If we ignore Monk 1, Monk 2, and parity---datasets that C4.5 does very badly on---the average accuracy for Holte-II is 91.2% and 88.5% for C4.5. Holte's 1R program (Holte 1993) built one-rules, that is, rules that test a single attribute, and was


Chotirat Ann and Dimitrios Gunopulos. Scaling up the Naive Bayesian Classifier: Using Decision Trees for Feature Selection. Computer Science Department University of California.

3,198 instances, 37 attributes, 2classes. Attributes selected by SBC = 4. Monk 85 88 91 94 97 100 10203040506070809099 Training Data NBC SBC C4.5 Figure 5. Monk dataset (prob.3). 554 instances, 6 attributes, 2 classes. Attributes selected by SBC = 4. Mushroom 90 92 94 96 98 100 10 20 30 40 50 60 70 80 90 99 Training Data (%) Accuracy (%) NBC SBC C4.5 Figure 6.


Wl odzisl/aw Duch and Rafal Adamczak and Krzysztof Grabczewski and Norbert Jankowski. Control and Cybernetics. Department of Computer Methods, Nicholas Copernicus University.

classifiers used in such problems, but they have an opinion of being opaque black boxes. Several neural methods have been compared experimentally on the mushroom and the 3 Monk problems benchmark datasets (Andrews et al. 1995), and recently comparison with some machine learning methods has been given (Duch et al. 2000). There is no reason why a simple classification model based on logical rules


Ron Kohavi and Dan Sommerfield. To Appear in KDD-98 Targeting Business Users with Decision Table Classifiers. Data Mining and Visualization Silicon Graphics, Inc.

available, mostly natural but also a few artificial ones (m-of-n and the monk problems). The artificial datasets were tested on the given training and test sets. The natural datasets were evaluated using 10fold cross-validation if the file size was less than 3,000 records (to ensure a small standard deviation


Wl odzisl and Rafal Adamczak and Krzysztof Grabczewski and Grzegorz Zal. A hybrid method for extraction of logical rules from data. Department of Computer Methods, Nicholas Copernicus University.

this hybrid method the simplest logical description for several benchmark problems (Iris, mushroom) has been found. Very good solutions were obtained for the three monk problems. For many medical datasets (only 3 were shown here) very simple and highly accurate results were obtained. It is not quite clear why logical rules work so well, for example in the hypothyroidor the Wisconsin breast cancer


Return to MONK's Problems data set page.

Supported By:

 In Collaboration With:

About  ||  Citation Policy  ||  Donation Policy  ||  Contact  ||  CML