Professional Documents
Culture Documents
Number of
Data Set Multivariate 155 Area: Life
Instances:
Characteristics:
Number of
Associated Tasks: Classification Missing Values? Yes 209076
Web Hits:
Source:
Creator:
unknown
Donor:
Attribute Information:
The BILIRUBIN attribute appears to be continuously-valued. I checked this with the donater, Bojan
Cestnik, who replied:
About the hepatitis database and BILIRUBIN problem I would like to say the following: BILIRUBIN is
continuous attribute (= the number of it's "values" in the ASDOHEPA.DAT file is negative!!!); "values" are
quoted because when speaking about the continuous attribute there is no such thing as all possible
values. However, they represent so called "boundary" values; according to these "boundary" values the
attribute can be discretized. At the same time, because of the continious attribute, one can perform some
other test since the continuous information is preserved. I hope that these lines have at least roughly
answered your question.
Relevant Papers:
Diaconis,P. & Efron,B. (1983). Computer-Intensive Methods in Statistics. Scientific American, Volume
248.
[Web Link]
Cestnik,G., Konenenko,I, & Bratko,I. (1987). Assistant-86: A Knowledge-Elicitation Tool for Sophisticated
Users. In I.Bratko & N.Lavrac (Eds.) Progress in Machine Learning, 31-45, Sigma Press.
[Web Link]
Amaury Habrard and Marc Bernard and Marc Sebban. IOS Press Detecting Irrelevant Subtrees to
Improve Probabilistic Learning from Tree-structured Data. Fundamenta Informaticae. 2004. [View
Context].
Zhi-Hua Zhou and Yuan Jiang and Shifu Chen. Extracting symbolic rules from trained neural network
ensembles. AI Commun, 16. 2003. [View Context].
Xiaoli Z. Fern and Carla Brodley. Boosting Lazy Decision Trees. ICML. 2003. [View Context].
Jinyan Li and Limsoon Wong. Using Rules to Analyse Bio-medical Data: A Comparison between C4.5
and PCL. WAIM. 2003. [View Context].
Michael L. Raymer and Travis E. Doom and Leslie A. Kuhn and William F. Punch. Knowledge discovery
in medical and biological datasets using a hybrid Bayes classifier/evolutionary algorithm. IEEE
Transactions on Systems, Man, and Cybernetics, Part B, 33. 2003. [View Context].
Takashi Matsuda and Hiroshi Motoda and Tetsuya Yoshida and Takashi Washio. Mining Patterns from
Structured Data by Beam-Wise Graph-Based Induction. Discovery Science. 2002. [View Context].
Petri Kontkanen and Petri Myllym and Tomi Silander and Henry Tirri and Peter Gr. On predictive
distributions and Bayesian networks. Department of Computer Science, Stanford University. 2000. [View
Context].
Gary M. Weiss and Haym Hirsh. A Quantitative Study of Small Disjuncts: Experiments and Results.
Department of Computer Science Rutgers University. 2000. [View Context].
David W. Opitz and Richard Maclin. Popular Ensemble Methods: An Empirical Study. J. Artif. Intell. Res.
(JAIR, 11. 1999. [View Context].
Yk Huhtala and Juha Kärkkäinen and Pasi Porkka and Hannu Toivonen. Efficient Discovery of Functional
and Approximate Dependencies Using Partitions. ICDE. 1998. [View Context].
Floriana Esposito and Donato Malerba and Giovanni Semeraro. A Comparative Analysis of Methods for
Pruning Decision Trees. IEEE Trans. Pattern Anal. Mach. Intell, 19. 1997. [View Context].
. Prototype Selection for Composite Nearest Neighbor Classifiers. Department of Computer Science
University of Massachusetts. 1997. [View Context].
Christophe Giraud and Tony Martinez and Christophe G. Giraud-Carrier. University of Bristol Department
of Computer Science ILA: Combining Inductive Learning with Prior Knowledge and Reasoning. 1995.
[View Context].
Gabor Melli. A Lazy Model-Based Approach to On-Line Classification. University of British Columbia.
1989. [View Context].
Rafael S. Parpinelli and Heitor S. Lopes and Alex Alves Freitas. An Ant Colony Based System for Data
Mining: Applications to Medical Data. CEFET-PR, CPGEI Av. Sete de Setembro, 3165. [View Context].
Suresh K. Choubey and Jitender S. Deogun and Vijay V. Raghavan and Hayri Sever. A comparison of
feature selection algorithms in the context of rough classifiers. [View Context].
Takao Mohri and Hidehiko Tanaka. An Optimal Weighting Criterion of Case Indexing for Both Numeric
and Symbolic Attributes. Information Engineering Course, Faculty of Engineering The University of
Tokyo. [View Context].
Wl/odzisl/aw Duch and Rafal/ Adamczak Email:duchraad@phys. uni. torun. pl. Statistical methods for
construction of neural networks. Department of Computer Methods, Nicholas Copernicus University.
[View Context].
Chris Drummond and Robert C. Holte. C4.5, Class Imbalance, and Cost Sensitivity: Why Under-
Sampling beats Over-Sampling. Institute for Information Technology, National Research Council
Canada. [View Context].
Ida G. Sprinkhuizen-Kuyper and Elena Smirnova and I. Nalbantis. Reliability yields Information Gain.
IKAT, Universiteit Maastricht. [View Context].
Zhi-Hua Zhou and Xu-Ying Liu. Training Cost-Sensitive Neural Networks with Methods Addressing the
Class Imbalance Problem. [View Context].
Rafael S. Parpinelli and Heitor S. Lopes and Alex Alves Freitas. PART FOUR: ANT COLONY
OPTIMIZATION AND IMMUNE SYSTEMS Chapter X An Ant Colony Algorithm for Classification Rule
Discovery. CEFET-PR, Curitiba. [View Context].
Wl/odzisl/aw Duch and Rafal Adamczak and Geerd H. F Diercksen. Neural Networks from Similarity
Based Perspective. Department of Computer Methods, Nicholas Copernicus University. [View Context].
Wl/odzisl/aw Duch and Karol Grudzinski and Geerd H. F Diercksen. Minimal distance neural methods.
Department of Computer Methods, Nicholas Copernicus University. [View Context].
Wl odzisl and Rafal Adamczak and Krzysztof Grabczewski. Optimization of Logical Rules Derived by
Neural Procedures. Department of Computer Methods, Nicholas Copernicus University. [View Context].
Wl/odzisl/aw Duch and Rafal Adamczak and Geerd H. F Diercksen. Classification, Association and
Pattern Completion using Neural Similarity Based Methods. Department of Computer Methods, Nicholas
Copernicus University. [View Context].
Elena Smirnova and Ida G. Sprinkhuizen-Kuyper and I. Nalbantis and b. ERIM and Universiteit
Rotterdam. Unanimous Voting using Support Vector Machines. IKAT, Universiteit Maastricht. [View
Context].
Citation Request:
Please refer to the Machine Learning Repository's citation policy