WO2000065480A2 - System and method for generating decision trees - Google Patents
System and method for generating decision trees Download PDFInfo
- Publication number
- WO2000065480A2 WO2000065480A2 PCT/US2000/010654 US0010654W WO0065480A2 WO 2000065480 A2 WO2000065480 A2 WO 2000065480A2 US 0010654 W US0010654 W US 0010654W WO 0065480 A2 WO0065480 A2 WO 0065480A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- cluster
- features
- data
- maximal
- fuzzy
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/35—Clustering; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/243—Classification techniques relating to the number of classes
- G06F18/24323—Tree-organised classifiers
Definitions
- the present invention relates to data analysis and more particularly to generating decision trees
- Data mining, knowledge discovery, and other forms of data analysis involve the extraction of useful information from vast amounts of accumulated data
- pharmaceutical companies are creating large databases listing drug compounds and their features, such as which diseases that are effectively treated by which drug compound and what are the drug compound's side-effects Given the large number of different drug compounds, it is difficult to manually analyze this data to ascertain useful patterns, such as determining what group of drugs are more or less effective in treating each of a group of diseases, especially when the desired groupings of drugs and diseases are not identified beforehand
- a decision tree is a data structure that contains a hierarchical arrangement of rules that successively indicates how to classify an object into a plurality of classes More specifically, each object is characterized by a number of attributes, and each rule in the decision tree tests the value of one of the attributes Decision trees separate out data into sets of rules that are likely to have a different effect on a target variable For example, one might want to find the characteristics of a drug compound and its method of administration that are likely to be effective in treating a particular disease These characteristics can be translated into a set of rules
- FIG 5 depicts an exemplary decision tree 500 that represents how to treat a hypothetical medical condition for a patient
- the exemplary decision tree 500 comprises two branch nodes 510 and 530, three leaf nodes 520, 540, and 550, and arcs 512, 514, 532, and 534
- Each of the branch nodes 510 and 530 represents a "rule" or condition that indicates how to choose between a number of possible values for a particular attribute of the patient
- the possible values that the attribute may take are indicated by the arcs 512, 514, 532, and 534
- the corresponding arc is taken to reach a leaf node or another branch node
- One of the branch nodes 510 is designated as the "root" node, which is the starting point of the decision tree
- root branch node 510 is labeled "AGE?" and indicates that the age of the patient is tested
- arc 514 connects branch node 510 to branch node 530 and is labeled "> 12?", which indicates that branch node 530 is to be reached if the age of the patient is greater than 12
- Branch node 530 is labeled "TEMP?" to indicate that the body temperature of the patient is tested If the body temperature of the patient is less than or equal to 102° (indicated by arc 532), then leaf node 540 is reached, otherwise, if the body temperature of the patient is greater than 102° (indicated by arc 5334), then leaf node 550 is reached
- the leaf nodes 520, 540, and 550 represent a "decision" or classification of the object
- the decision is the treatment to be administered to the patient
- the decision is to use 20 mg of drug X
- the decision is to use 40 mg of drug X
- the decision is to use 10 mg of drug Y
- the exemplary decision tree 500 may be used to determine which treatment to be administered to a patient by starting at the "root" node, testing the attribute of the patients to select an arc and follow the arc until a leaf node is reached.
- a 10 year old child with a temperature of 98 6° is to be treated Starting at root branch node 510
- the age of the patient is tested Since the 10 year old is less than 12 years of age
- arc 512 is followed to reach leaf node 520 Therefore, 20 mg of drug X is prescribed to the 10 year old
- the patient is a 32-year with a 105° fever Starting at root branch node 510, the age of the patient is tested Since the 32-year old's age is greater than 12, arc 514 is followed to branch node 530 where the body temperature of the patient is tested Since the patient has a 105° fever, arc 534 is followed to reach leaf node 550, which indicates that 10 mg of drug Y is to be administered
- Decision tree induction refers to the process of determining how to build the decision tree from a set of training data
- a decision tree is built by successively identifying which attributes to test first and which attributes to test later, if at all
- a common conventional approach to build decision trees is known as "Induction of Decision Trees" or ID3
- ID3 is a recursive algorithm that starts with a set of training objects that belong to a set of predefined classes If all the objects belong to a single class, then there is no decision to make and a leaf node is created and labeled with the class Otherwise, a branch node is created and the attribute with the highest "information gain" is selected if that attribute were used to discriminate objects at the branch node
- the information gain is calculated by finding the average entropy of each attribute
- a problem with conventional decision trees such as those produced by ID3 is that such decision trees are rigid, inflexible, and brittle
- conventional decision trees impose an "either-or" or binary approach to the data, even though different drugs have varying degrees of effectiveness
- data values close to the border of a crisp range in a decision tree are apt to be misclassified due to the imprecision of real-world data
- fuzzy logic employs a "membership function" between 0 0 and 1 0 to represent the degree to which the object belongs to the class For example, rather than categorize a patient's age as "twelve years and below" and "above twelve years," two fuzzy sets, Young and Old, can be employed, such that a two-year old may
- FID3 One attempt to combine fuzzy logic with classical, crisp decision trees is known as FID3, in which the user defines the membership functions in each of the predefined classes for all of the training data Each membership function can serve as an arc label of a fuzzy decision As in ID3, FID3 generates its decision tree by maximizing information gains
- the decision of the fuzzy decision tree is also a fuzzy variable, indicating the memberships of a tested object in each of the possible classifications
- the arcs 512 and 514 emanating from branch node 510 could be fuzzified by a membership function on a Young fuzzy set and an Old fuzzy set, respectively
- arc 512 could be the test ⁇ y oung (X,) ⁇ 0 5 or other value that maximizes the information gain
- the respective fuzzy sets could be Normal and Feverish, respectively
- FID3 One disadvantage with FID3 is that the membership functions in each of the attributes for all of the training data must be specified beforehand by the user For data with a high number of attributes or dimensions, however, determining the membership functions is typically a difficult task, requiring intensive involvement by experts In addition, the fuzzy sets themselves may not even be known beforehand and require further investigation
- the data are dynamically clustered while a decision tree is generated
- the data are clustered using a fuzzy clustering analysis, which generates the membership functions on the fly, without requiring the user to predefine sets or calculate the membership functions beforehand
- one aspect of the invention involves a method and software for generating a decision tree for data characterized by several features, in which several fuzzy cluster analyses are performed along each of the features to calculate a maximal partition coefficient and a corresponding set of one or more fuzzy clusters The feature corresponding to the maximal partition coefficient is selected, and the decision tree is built based on the corresponding set of one or more fuzzy clusters
- Another aspect of the invention relates to a method and software for generating a decision tree for data that is characterized by several features, in which several cluster analyses are performed along each of the features to calculate a maximal cluster validity measure One of the features corresponding to the maximal cluster validity measure is selected, and the data is subdivided into one or more groups based on the selected feature Then, the decision tree is built based on the one or more groups By performing cluster analyses to calculate a maximal cluster validity, the decision tree can correspond to an optimal cluster separability
- Still another aspect of the invention pertains to a method and software for generating a decision tree for data characterized by several features (e g dimensions or attributes), in which one of the features is selected A cluster analysis along the selected feature is performed to group the data into one or more clusters, and the decision tree is built based on the one or more clusters
- a cluster analysis along the selected feature is performed to group the data into one or more clusters, and the decision tree is built based on the one or more clusters
- FIG 1 depicts a computer system upon which an embodiment of the present invention can be implemented
- FIG 2 is a flowchart illustrating the operation of one embodiment of the present invention
- FIG 3 is a graph of an exemplary data set used to illustrate the operation of one embodiment of the present invention
- FIG 4 is a schematic diagram of an exemplary decision tree produced by one embodiment of the present invention.
- FIG 5 is a schematic diagram of a decision tree
- FIG 1 is a block diagram that illustrates a computer system 100 upon which an embodiment of the invention may be implemented
- Computer system 100 includes a bus 102 or other communication mechanism for communicating information, and a processor 104 coupled with bus 102 for processing information
- Computer system 100 also includes a main memory 106, such as a random access memory (RAM) or other dynamic storage device, coupled to bus 102 for storing information and instructions to be executed by processor 104
- Main memory 106 also may be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 104
- Computer system 100 further includes a read only memory (ROM) 108 or other static storage device coupled to bus 102 for storing static information and instructions for processor 104
- ROM read only memory
- a storage device 110 such as a magnetic disk or optical disk, is provided and coupled to bus 102 for storing information and instructions
- Computer system 100 may be coupled via bus 102 to a display 112, such as a cathode ray tube (CRT), for displaying information to a computer user
- a display 112 such as a cathode ray tube (CRT)
- An input device 114 is coupled to bus 102 for communicating information and command selections to processor 104
- cursor control 116 is Another type of user input device
- cursor control 116 such as a mouse, a trackball, or cursor direction keys
- This input device typically has two degrees of freedom in two axes, a first axis (e g , x) and a second axis (e g , y), that allows the device to specify positions in a plane
- the invention is related to the use of computer system 100 for generating decision trees
- generating decision trees is provided by computer system 100 in response to processor 104 executing one or more sequences of one or more instructions contained in main memory 106
- Such instructions may be read into main memory 106 from another computer- readable medium, such as storage device 110
- Execution of the sequences of instructions contained in main memory 106 causes processor 104 to perform the process steps described herein
- processors in a multi-processing arrangement may also be employed to execute the sequences of instructions contained in main memory 106
- hard-wired circuitry may be used in place of or in combination with software instructions to implement the invention
- embodiments of the invention are not limited to any specific combination of hardware circuitry and software
- Non-volatile media include for example optical or magnetic disks, such as storage device 110
- Volatile media include dynamic memory, such as main memory 106
- Transmission media include coaxial cables, copper wire and fiber optics, including the wires that comprise bus 102 Transmission media can also take the form of acoustic or light waves, such as those generated during radio frequency (RF) and infrared (IR) data communications
- RF radio frequency
- IR infrared
- Common forms of computer-readable media include for example, a floppy disk, a flexible disk, hard disk, magnetic tape, any other magnetic medium a CD-ROM, DVD, any other optical medium, punch cards, paper tape, any other physical medium with patterns of holes, a RAM, a PROM, and EPROM, a FLASH-EPROM, any other memory chip or cartridge, a carrier wave as described hereinafter, or any other medium from which
- Various forms of computer readable media may be involved in carrying one or more sequences of one or more instructions to processor 104 for execution
- the instructions may initially be borne on a magnetic disk of a remote computer
- the remote computer can load the instructions into its dynamic memory and send the instructions over a telephone line using a modem
- a modem local to computer system 100 can receive the data on the telephone line and use an infrared transmitter to convert the data to an infrared signal
- An infrared detector coupled to bus 102 can receive the data carried in the infrared signal and place the data on bus 102
- Bus 102 carries the data to main memory 106, from which processor 104 retrieves and executes the instructions
- the instructions received by main memory 106 may optionally be stored on storage device 1 10 either before or after execution by processor 104
- Computer system 100 also includes a communication interface 1 18 coupled to bus 102
- Communication interface 118 provides a two-way data communication coupling to a network link 120 that is connected to a local network 122
- communication interface 1 18 may be an integrated services digital network (ISDN) card or a modem to provide a data communication connection to a corresponding type of telephone line
- ISDN integrated services digital network
- communication interface 118 may be a local area network (LAN) card to provide a data communication connection to a compatible LAN Wireless links may also be implemented
- communication interface 118 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information
- Network link 120 typically provides data communication through one or more networks to other data devices
- network link 120 may provide a connection through local network 122 to a host computer 124 or to data equipment operated by an Internet Service Provider (ISP) 126
- ISP 126 in turn provides data communication services through the worldwide packet data communication network now commonly referred to as the "Internet" 128
- Internet Internet Service Provider
- Computer system 100 can send messages and receive data, including program code, through the network(s), network link 120, and communication interface 1 18
- a server 130 might transmit a requested code for an application program through Internet 128, ISP 126, local network 122 and communication interface 118
- one such downloaded application provides for generating decision trees as described herein
- the received code may be executed by processor 104 as it is received, and/or stored in storage device 1 10, or other non-volatile storage for later execution In this manner, computer system 100 may obtain application code in the form of a carrier wave
- FUZZY CLUSTERING One aspect of the present invention stems from the realization that data set itself contains the information that can be used to obtain reasonable membership functions Thus, both the decision tree and the membership functions can be automatically extracted from the data set
- FCM fuzzy c-means
- the continuous membership functions ⁇ , (/) IR - [0 1 ], / e 1 c, I e 1 p, can be obtained by projection and subsequent interpolation or approximation of the u lk memberships, or simply by inserting the projections v of the cluster centers v, into
- the extraction of membership functions is obtained by clustering while generating a decision tree by induction
- the decision tree starts with a root node that is associated with the data set
- the steps illustrated in FIG 2 are performed
- the result of this working example is depicted as decision tree 400 of FIG 4
- the number of remaining dimensions p is checked If zero dimensions remain, then there is no further subdivision to be made and the current node is made a leaf node
- dimension is used in this discussion because the point data in the working example are continuous, the present invention is not so limited, and the described approach may be applied more gene ⁇ cally to features of the data, which can be continuous dimension data as well as nominal attribute data
- Dimensional data typically refers to continuous or quantized measurements, such as length, time, charge, temperature, mass, energy, etc
- Nominal attribute data typically refers to one element chosen from a finite set, for example, male/female, etc
- data set 300 is comprises 29 two-dimensional points, which form four clearly visible cluster, cluster 310 of six points in the upper-left corner, cluster 320 of nine points in the lower-left corner, cluster 330 of six points in the middle, and cluster 340 of four points in the upper-right corner
- there are two remaining dimensions, x and y thus execution proceeds to step 202
- the data in each dimension are clustered and a partition coefficient, which quantifies the goodness of the clustering, is computed for each dimension as a measure of cluster validity or how well separated the clusters are
- fuzzy c-means clustering may be employed, but other forms of fuzzy clustering such as fuzzy k-means may be employed
- fuzzy c- means clustering approach the objective function J FCM IS minimized for a given number of clusters c
- PC(l/ c (/) ) is calculated for c > 1 as follows
- the data components might not possess any cluster structure at all Thus, there is a need for testing the data to determine if the best clustering is a single clustering
- the domain limits of the data are tested, and if the domain of the data in a cluster in the dimension ( ⁇ m ax - ⁇ mm) fall within a predetermined fraction of the of the domain of the entire data set (Xmax - Xmin), then the data are considered to constitute a single cluster More specifically, the following test can be performed
- the clustering in the x-dimension with the maximal partition coefficient of 0 94 has three clusters, and the clustering in the y-dimensio ⁇ with the maximal partition coefficient of 0 93 also has three clusters
- the x-dimension is chosen because the partition coefficient of the x-dimensio ⁇ (0 94) is greater than the partition coefficient of the y-dimension (0 93)
- the data is then partitioned into the c * clusters along the selected dimension / * These subsets are inserted (step 208) into the decision tree by constructing c * arcs at the current node N with the labels x (r) ⁇ ⁇ (r) , b ⁇ ⁇ x ( ⁇ ⁇ ⁇ 2 (r) ,
- branch node 410 at the root of the decision tree is labeled to indicate that the x-dimension is tested at node 410
- Each arc 412, 413, and 414 from node 410 to respective nodes 420, 430, and 440 is labeled with the borders (8 00, 22 31 ), (22 31 , 41 19), and (41 19, 53 00), respectively
- the x- coordinates are removed from the data, leaving only the y-coordinates, because the x-dimension was the dimension selected in step 204
- step 212 the process of steps 200-210 is repeated recursively for each cluster, until there is no dimension or data left
- the recursive application of this process results in two additional leaf nodes 450 and 460, corresponding to clusters 320 and 310 respectively
- the arcs 425 and 426 from branch node 420 to respective leaf nodes 450 and 460 are labeled (8 00, 31 32) and (31 32 58 00), respectively
- the recursive calls for the other subsets corresponding to branch nodes 430 and 440 do not result in further splitting into clusters because the test of inequality (6) held Rather, a single arc 437 emanates from the branch node 430 to leaf node 470, and a single arc 448 emanates from the branch node 440 to leaf node 480
- the result of this embodiment on the exemplary data set 300 is therefore a decision tree 400 with four leaf nodes 450, 460, 470, and 480, each corresponding to the four clusters 320, 310, 330, and 340, respectively, in the data set 300 that was evident by visual inspection
- the corresponding crisp partition moreover, is exactly the same as the visually obtained partitions, / e the points x k _ X belong to the same (crisp) classes
- the automatically generated borderlines are set up to produce optimal cluster separability
- a decision tree clustering procedure which employs a unified approach to extracting both the decision tree and the (crisp or fuzzy) clusters
- the decision tree is built by subsequent clustering of single dimensions or features, and the choice of the winning separation is based on cluster validity
- the clustering employs a fuzzy c-means (FCM) model and the partition coefficient (PC) to determine the selected separations
- FCM fuzzy c-means
- PC partition coefficient
- clustering models other than FCM can be employed for generating decisions trees
- HCM hard c-means
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Databases & Information Systems (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
- Breeding Of Plants And Reproduction By Means Of Culturing (AREA)
Abstract
Description
Claims
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP00928239A EP1212698A2 (en) | 1999-04-23 | 2000-04-21 | System and method for generating decision trees |
JP2000614155A JP4817502B2 (en) | 1999-04-23 | 2000-04-21 | System and method for generating a decision tree |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13063699P | 1999-04-23 | 1999-04-23 | |
US60/130,636 | 1999-04-23 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2000065480A2 true WO2000065480A2 (en) | 2000-11-02 |
WO2000065480A3 WO2000065480A3 (en) | 2001-04-05 |
Family
ID=22445618
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2000/010654 WO2000065480A2 (en) | 1999-04-23 | 2000-04-21 | System and method for generating decision trees |
Country Status (3)
Country | Link |
---|---|
EP (1) | EP1212698A2 (en) |
JP (1) | JP4817502B2 (en) |
WO (1) | WO2000065480A2 (en) |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2004037448A2 (en) * | 2002-10-18 | 2004-05-06 | Giesecke & Devrient Gmbh | Method and system for processing banknotes |
WO2005024648A1 (en) * | 2003-09-05 | 2005-03-17 | Agency For Science, Technology And Research | Methods of processing biological data |
KR100498651B1 (en) * | 2002-05-03 | 2005-07-01 | 이영섭 | statistic classification method of data using one-sided purity splitting criteria for classification trees in data mining |
WO2007005547A1 (en) * | 2005-07-01 | 2007-01-11 | Square D Company | Automated hierarchy classification in utility monitoring systems |
US7349815B2 (en) | 2005-07-01 | 2008-03-25 | Square D Company | Automated integration of data in utility monitoring systems |
US7499908B2 (en) | 2003-04-23 | 2009-03-03 | International Business Machines Corporation | Method for identifying a workload type for a given workload of database requests |
US7684441B2 (en) | 2005-07-01 | 2010-03-23 | Bickel Jon A | Automated precision alignment of data in a utility monitoring system |
US8024390B2 (en) | 2005-07-01 | 2011-09-20 | Schneider Electric USA, Inc. | Automated data alignment based upon indirect device relationships |
US8081726B2 (en) | 2007-05-10 | 2011-12-20 | Schneider Electric USA, Inc. | Method and apparatus for synchronizing data in utility system |
US8260579B2 (en) | 2009-07-01 | 2012-09-04 | Schneider Electric USA, Inc. | Automatic identification of multiple power grids using data synchronization |
US8583582B2 (en) | 2010-01-28 | 2013-11-12 | Schneider Electric USA, Inc. | Robust automated hierarchical determination for power monitoring systems |
US9077208B2 (en) | 2011-12-30 | 2015-07-07 | Schneider Electric USA, Inc. | Method of detecting instability in islanded electrical systems |
US9176171B2 (en) | 2010-11-19 | 2015-11-03 | Schneider Electric USA, Inc. | Data alignment in large scale electrical system applications |
CN113487182A (en) * | 2021-07-06 | 2021-10-08 | 新智数字科技有限公司 | Equipment health state evaluation method and device, computer equipment and medium |
CN113627788A (en) * | 2021-08-10 | 2021-11-09 | 中国电信股份有限公司 | Service policy determination method and device, electronic equipment and storage medium |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
AU2012387666B2 (en) | 2012-08-15 | 2016-02-11 | Entit Software Llc | Validating a metadata tree using a metadata integrity validator |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5574910A (en) * | 1992-07-02 | 1996-11-12 | Bay Networks, Inc. | Method for segmenting data packets to form binary decision trees which determine filter masks combined to filter the packets for forwarding |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3363996B2 (en) * | 1994-05-17 | 2003-01-08 | 三菱電機株式会社 | Judgment procedure generation method and judgment procedure generation device |
-
2000
- 2000-04-21 WO PCT/US2000/010654 patent/WO2000065480A2/en active Application Filing
- 2000-04-21 JP JP2000614155A patent/JP4817502B2/en not_active Expired - Fee Related
- 2000-04-21 EP EP00928239A patent/EP1212698A2/en not_active Withdrawn
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5574910A (en) * | 1992-07-02 | 1996-11-12 | Bay Networks, Inc. | Method for segmenting data packets to form binary decision trees which determine filter masks combined to filter the packets for forwarding |
Non-Patent Citations (3)
Title |
---|
CEZARY Z.JANIKOW: "Fuzzy Decision Trees : Issues and Methods" IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS, vol. 28, no. 1, February 1998 (1998-02), XP002156702 * |
JUNG S -W ET AL: "A DESIGN SCHEME FOR A HIERARCHICAL FUZZY PATTERN MATCHING CLASSIFIER AND ITS APPLICATION TO THE TIRE TREAD PATTERN RECOGNITION" FUZZY SETS AND SYSTEMS,NL,ELSEVIER SCIENCE PUBLISHERS, AMSTERDAM, vol. 65, no. 2/03, 10 August 1994 (1994-08-10), pages 311-322, XP000461892 ISSN: 0165-0114 * |
LUO R C ET AL: "AUTOMATED DECISION TREE GENERATION FOR OBJECT RECOGNITION AND CLASSIFICATION" PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON INDUSTRIAL ELECTRONICS,CONTROL AND INSTRUMENTATION. (IECON),US,NEW YORK, IEEE, 29 September 1986 (1986-09-29), pages 357-362, XP002056809 * |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100498651B1 (en) * | 2002-05-03 | 2005-07-01 | 이영섭 | statistic classification method of data using one-sided purity splitting criteria for classification trees in data mining |
WO2004037448A3 (en) * | 2002-10-18 | 2004-08-12 | Giesecke & Devrient Gmbh | Method and system for processing banknotes |
WO2004037448A2 (en) * | 2002-10-18 | 2004-05-06 | Giesecke & Devrient Gmbh | Method and system for processing banknotes |
US7499908B2 (en) | 2003-04-23 | 2009-03-03 | International Business Machines Corporation | Method for identifying a workload type for a given workload of database requests |
WO2005024648A1 (en) * | 2003-09-05 | 2005-03-17 | Agency For Science, Technology And Research | Methods of processing biological data |
US7684441B2 (en) | 2005-07-01 | 2010-03-23 | Bickel Jon A | Automated precision alignment of data in a utility monitoring system |
US7349815B2 (en) | 2005-07-01 | 2008-03-25 | Square D Company | Automated integration of data in utility monitoring systems |
US7272518B2 (en) | 2005-07-01 | 2007-09-18 | Square D Company | Automated hierarchy classification in utility monitoring systems |
WO2007005547A1 (en) * | 2005-07-01 | 2007-01-11 | Square D Company | Automated hierarchy classification in utility monitoring systems |
US8024390B2 (en) | 2005-07-01 | 2011-09-20 | Schneider Electric USA, Inc. | Automated data alignment based upon indirect device relationships |
US8081726B2 (en) | 2007-05-10 | 2011-12-20 | Schneider Electric USA, Inc. | Method and apparatus for synchronizing data in utility system |
US8260579B2 (en) | 2009-07-01 | 2012-09-04 | Schneider Electric USA, Inc. | Automatic identification of multiple power grids using data synchronization |
US8583582B2 (en) | 2010-01-28 | 2013-11-12 | Schneider Electric USA, Inc. | Robust automated hierarchical determination for power monitoring systems |
US9176171B2 (en) | 2010-11-19 | 2015-11-03 | Schneider Electric USA, Inc. | Data alignment in large scale electrical system applications |
US9077208B2 (en) | 2011-12-30 | 2015-07-07 | Schneider Electric USA, Inc. | Method of detecting instability in islanded electrical systems |
CN113487182A (en) * | 2021-07-06 | 2021-10-08 | 新智数字科技有限公司 | Equipment health state evaluation method and device, computer equipment and medium |
CN113487182B (en) * | 2021-07-06 | 2024-04-26 | 新奥新智科技有限公司 | Device health state evaluation method, device, computer device and medium |
CN113627788A (en) * | 2021-08-10 | 2021-11-09 | 中国电信股份有限公司 | Service policy determination method and device, electronic equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
JP4817502B2 (en) | 2011-11-16 |
EP1212698A2 (en) | 2002-06-12 |
JP2002543497A (en) | 2002-12-17 |
WO2000065480A3 (en) | 2001-04-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7197504B1 (en) | System and method for generating decision trees | |
Langley | Induction of recursive Bayesian classifiers | |
EP1212698A2 (en) | System and method for generating decision trees | |
US10073906B2 (en) | Scalable tri-point arbitration and clustering | |
Tan et al. | Evolutionary computing for knowledge discovery in medical diagnosis | |
Hruschka et al. | A genetic algorithm for cluster analysis | |
Yang et al. | A non-revisiting quantum-behaved particle swarm optimization based multilevel thresholding for image segmentation | |
Looney | A fuzzy clustering and fuzzy merging algorithm | |
Zhu et al. | Medical image segmentation using fruit fly optimization and density peaks clustering | |
Hruschka et al. | Feature selection by Bayesian networks | |
Kim et al. | Uncertain reasoning and learning for feature grouping | |
Wilkins et al. | Comparison of five clustering algorithms to classify phytoplankton from flow cytometry data | |
Jiao et al. | [Retracted] An Improved Cuckoo Search Algorithm for Multithreshold Image Segmentation | |
CN116483932A (en) | Hierarchical classification management method and device based on coronary heart disease drug library of endocardial department | |
Houben et al. | Coupling of K-NN with decision trees for power system transient stability assessment | |
CN116150690A (en) | DRGs decision tree construction method and device, electronic equipment and storage medium | |
Dehuri et al. | Revisiting evolutionary algorithms in feature selection and nonfuzzy/fuzzy rule based classification | |
Ishibuchi et al. | Designing fuzzy rule-based classifiers that can visually explain their classification results to human users | |
Tahir et al. | Shock graph for representation and modeling of posture | |
Li et al. | A multi‐classification model for non‐small cell lung cancer subtypes based on independent subtask learning | |
Wei et al. | An entropy clustering analysis based on genetic algorithm | |
Tareq et al. | A new density-based method for clustering data stream using genetic algorithm | |
US20030023575A1 (en) | System and method of automatic object classification by tournament strategy | |
Dharmalingam et al. | Hybrid feature selection model for classification of lung disorders | |
CN112364896B (en) | Method and device for determining health information distribution based on machine learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A2 Designated state(s): JP |
|
AL | Designated countries for regional patents |
Kind code of ref document: A2 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
AK | Designated states |
Kind code of ref document: A3 Designated state(s): JP |
|
AL | Designated countries for regional patents |
Kind code of ref document: A3 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE |
|
ENP | Entry into the national phase |
Ref country code: JP Ref document number: 2000 614155 Kind code of ref document: A Format of ref document f/p: F |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2000928239 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 2000928239 Country of ref document: EP |