Home

Running RF++ from the command line ( do we want this

image

Contents

1. P RP RP RP RP RPP Pe J N N hN2 NJ Im HD RPP RP RP RP RP Re INP PRP RPP RP RP RP PP ee 5o OOS OTOSTSOOOOSPDS 3 Full clustered with numsubjects 100 Elapsed Time 72 seconds The Variable Importance tab displays variable importance scores for 2 variable importance measures side by side Variables in each column are sorted in decreasing order of importance RF File Options Help Training Data Forest Parameters Training Error Variable Importance L Mean Decrease in Margin MDM ID Score 36 0 0752 0623 0444 0027 0014 0013 0013 0013 0012 0012 0010 0010 0010 Elapsed Time 73 seconds c c Occ ood eo O00 000 o oeoooo eo oo oo Testing a Forest and Making Predictions The Training Prediction Data tab is used to test the performance of the trained forest or to make predictions for unknown cases This tab is similar to the Training Data tab but the Number of Variables and Number of Classes fields are automatically filled in form the training dataset and are unchangeable greyed out The user needs to provide the number of samples in the testing prediction data file and check the Have outcomes checkbox if the column of outcomes is present in the file This column should be present only in the testing file not in the file where prediction of the unknown cases are to be made Note that this column is never included in the count of variables RF
2. Subject level Error Rate 0 0700 Subject level Confusion Matrix rows predicted cols actual 48 5 2 45 Elapsed Time 1 seconds The following columns are displayed sample subject IDs true class predicted class followed by the proportions of votes for each of the classes J lo 0 1758 0 1339 9 1558 Sample level Error Rate 0 0000 Sample level Confusion Marix rows predicted cols actual 500 0 O 500 Subject level CLassifications ID True Predicted Class 0 Class Class a 0 0 8057 0 1772 0 8237 0 1349 0 8622 0 1625 0 A 8694 1771 1 2 3 4 5 6 7 a 20000000 1 iL 0 iC l 1 a a 1 1 0 a 1 1 Elapsed Time 6 seconds When making predictions a true outcome column should not be present Error rate s and confusion matrices will not be computed Saving forest variable importance scores and classifications To save a forest select on of the following tabs Training Data Forest Parameters or Training Error then select the Save menu item from the File pull down menu Forests are saved with the rff file extension Training error and confusion matrices are also saved in this file RF E foesut oo me 35 rj Browse for other folders 4 D leclerc yuliya GUI tmp Create Folder eme seines D leclerc D 20_3_1_train rff Yesterday D Desktop D 20_3_1_train_nowt rff Yesterday 7 File System 1D 20 3
3. 1 train prox rff Yesterday D iris 100 rff Yesterday X Cancel Variable Importance measures are saved by first clicking on the Variable Importance tab and then selecting the Save menu item from the File pull down menu This file is identical to the text displayed to the user in RF window RF Training Data Forest Parameters Training Error uu Save Training Variable Importance Oo x important_vars bd Wc 0 0 7 Browse for other folders 4 a leclerc yuliya cu Create Folder 07 03 07 icons smolka 07 03 07 unused 06 24 07 D iris testcl b t Tuesday v 10 Classifications predictions can be saved by first clicking on the Testing Prediction Classifications tab and then selecting the Save menu item from the File pull down menu from the RF Tum Wadi lh af at Save Testing Prediction Classifications Name some classification resultsbt e Browse for other folders 4 D leclerc yuliya GUI tmp Create Folder ame oea D 100 3 1 testcl bxt Yesterday DR RRR EB LI RRP Re D Desktop JD 100 3 1 testcl nowt b t Yesterday B File System 1D 100_3_1_testcl_prox bt Yesterday D iris 100 testel txt Yesterday a tos jo X Cancel se Loading an Existing Forest A previously forest can be loaded by selecting the Open menu item from the File pull
4. down menu After a forest is loaded 4 tabs that describe the forest are displayed The first 3 tabs describe the forest Training Data Forest Parameters and Training Error The Testing Prediction Data tab is opened for the user to specify testing or prediction dataset parameters in order to test performance of the forest or to make predictions for unknown cases 11 Forest File Description XML File The XML schema for a forest is defined in the file forest xsd This schema defines a forest consisting of forest attributes such as number of samples used to grow a forest number of trees etc followed by a sequence of trees Each tree is a sequence of nodes Nodes are listed in level order though knowledge of this ordering is unnecessary for parsing The correct linkage structure of a tree can be determined by using the node attribute id which uniquely identifies a node Each non terminal node also contains ids of the left and right children nodes An example of a forest in XML format is listed in the supplementary file forest xml Text File The forest is saved as a text file with the rff extension It is best to read the descriptions provided for the XML file syntax files forest xsd and forest xml The rff file contains more information then the XML file such as the confusion matrices that is useful when a trained forest is loaded into RF GUI An example is available in the supplementary file fores
5. File Help 4 Forest Parameters Training Error Variable Importance Testing Prediction Data Number of Samples 1000 Number of Variables es Number of Classes b Have outcomes Testing Prediction Data File trunk gensamples simulations sim2 ts 100 10 1 bt Browse Test Classify Full clustered with numsubjects 100 Elapsed Time 59 seconds Classifications are displayed in the Testing Prediction Classifications tab When testing sample classifications are displayed first then sample error rate and the sample confusion matrix If subject level classification is appropriate i e the data is cluster correlated and outcomes for each subject replicates belong to the same class the subject level classifications then the subject error rate and subject confusion matrix are displayed RF File Options Help 4 Variable Importance Testing Prediction Data Testing Prediction Classifications Sample level Classifications E ID True Predicted Class 1 Class 2 E Class Class 1 6180 3820 5800 4200 7035 2965 7465 2535 5995 4005 7470 2530 2615 7385 2330 7670 h2h2 BR RRP rrr 1 1 1 1 1 9 2 2 Sample level Error Rate 0 1310 Sample level Confusion Matrix rows predicted cols actual 454 85 46 415 Subject level CLassifications ID True Predicted Class 1 Class 2 Class Class 6769 3231 4870 5131 5971 4030 3558 6442 0 3386 0 6614 9 5931 4070 0 3185 6815
6. RFT User Manual Version 1 0 Home Pages http sourceforge net projects rfpp Contact Yuliya Karpievitch yuliya stat tamu edu Anthony Leclerc leclerc cs cofc edu Table of Contents Training growing a new Forest scsvccsscsssecseceneccvocesvcasesevnscoscevscconecevccsycevsessugeossebetvovncssdeveasbonendensnsevesende Training Data Information Pc P teia da eua ninndos Forest Parameters eese ec eere eed e see eo sod tou vede sao koe c ooa vede sto cadsd eben vo e eas caca d censoa desde oS ud ete ea eeu odassdocasoados 3 3 4 crire Eu Testing a Forest and Making Predictions cene ecce eee eee eee eee eee eo see ease essa sees csse eo sese es seeeess Saving forest variable importance scores and classifications eeeeeeeeeeeeeeeeeeeeeeee D Loading an Existing Forest RR RTT LL Forest File Description iier ere o ESO ia Gigas INNO OO onsen e iate L2 GUI Training growing a new Forest From the initial RF window we can select to train a new forest or open an existing forest To train a forest select the Train menu item from the File pull down menu This will open a training data tab Open Alt O Save Alt S Exit Alt X Welcome Training Data Information Two files are provided with the executable tran 100 10 1 txt and test 100 10 1 txt These are trainin
7. class in different row Class labels are not printed here classes are ordered as class1 2 in rows and columns 0 subject level confusion matrix identifier flag 0 1 O if no subject level confusion matrix follows 1 if subject level confusion matrix printed Forest file output with the subject level confusion matrix More information is stored in the forest file if subject level classification is done In the following example 2 classes were available 2 100 0 516667 0 433333 0 433333 0 4 0 513333 0 39 0 39 0 39 0 546667 0 433333 0 576667 122 36 28 114 1 43 11 7 39 End of file gt 2 100 2 proximity based weights are used 100 number of trees in the forest small for the purpose of the example 0 516667 0 433333 0 433333 0 4 0 513333 0 39 0 39 0 39 0 546667 0 433333 0 576667 tree weights 100 weights 122 36 28 114 sample level confusion matrix 1 flag indicator if subject level confusion matrix follows 1 means subject level matrix is present 43 11 7 39 subject level confusion matrix Generally the users do not need to be familiar with the syntax of these files RF will save and load these files automatically If someone wants to further investigate the rules used to grow trees and consequently a forest the rff and XML files will provide valuable insight into the rules 14
8. g and testing files respectively RF File Help Training Data Forest Parameters Number of Samples Number of Variables Number of Classes Have ids Training Data Bef 770 0 0 0 0 Browse Welcome In the Training Data tab information describing the dataset must be entered number of samples in the file number of variables and number of classes Samples in the training data file must be organized in rows 1 row per sample Columns must be arranged as such an ID is required in the in the first column if the data is clustered this ID is optional for non clustered data Next Number of Variables values for each variable The last column must be the outcome classification column Note if IDs are present RF will do subject level bootstrapping based on the values in the ID columns where all samples with matching IDs belong to the same subject 3 If the data are not clustered then subject level bootstrapping consists of subject clusters of size one Thus in this case subject level bootstrapping is equivalent to sample level bootstrapping The Number of Variables field is the number of variables in each sample and must not include the IDs and outcomes columns Outcomes classifications are integer values in the range 1 number of classes IDs are integers and are used primarily to identify clusters within the data Note no double or character values are allowed If the data does no
9. n all fields have been filled the forest can be grown by clicking the Generate Forest button The progress report including any error messages will be displayed in the status bar at the bottom of the RF window frame A Error reading in training samples message may appear if the data parameters are improperly specified This usually indicates a mismatch of the specified numbers of variables and or samples with the corresponding values read from the training file The progress report will change from Growing Forest to Calculating Statistics and finally to Done Training Results After a forest is built the Out of Bag OOB statistics are computed and 3 new tabs appear Training Error Variable Importance and Testing Predicting Data tabs The Training Error tab displays OOB sample level and subject level when applicable error rate s and confusion matrices RF File Options Help Training Data Forest Parameters Training Error Variable Importance Sample level Error Rate 0 141 Sample level Confusion Matrix rows predicted cols actual 434 75 66 425 Subject level Error Rate 0 05 Subject level Confusion Matrix rows predicted cols actual 48 3 2 47 Sample level OOB Classifications ID True Predicted Class 1 Class Class 1 1 0 6622 0 6243 9 5541 0 6608 0 6811 0 6932 0 6041 0 6243 9 5905 9 5622 0 6774 0 6145 90 5656 0 4316 A17 A0 NJ NJ NJ NM
10. ses Sample size that reached a particular node is printed for purposes of seeing how the splits are made and how many samples a split separates from the rest of the samples in a particular node The 1 node is the root node that has all the samples so we see the total number of the samples 150 which is the same as in the 1 line of the file Variables are numbered 1 2 3 Class last value in the line is in the range 1 2 for terminal nodes nodes that produce classifications and 1 for nodes inside the tree that do not produce the classification and are split further Trees are separated by a new line After all the trees are written out more forest information is added 04 1111 5074 0415 0241 0 End of file 0 100 0 which tree weight weight vector size which tree weight 0 1 O if no tree weights were used 1 if proximity based weights used 100 weight vector size number equal to the number of trees in the forest as each tree has a weight 1111 111 Tree weights as one long line with number of trees values in it Here the weights are set to 1 and thus do not affect the voting 13 50 04 0 462 0 448 Sample level confusion matrix Matrix shows the numbers of the Out of bag OOB samples classified into each of the classes Rows are the predicted outcome columns are true outcome Values on the diagonal are the correct classifications all values off diagonal were classified incorrectly as the
11. t comply with the standard RF will output an error message in the status bar located at the bottom of the GUI window Forest Parameters To generate a forest specify forest parameters by clicking on the Forest Parameters tab RF File Options Help Training Data Forest Parameters Number of variables to try at each split 4 4 Recommend value near 14 Number of trees 2000 Recommend value of 5000 or more Random Seed 1 Use Proximity based weights Generate Forest The 1 parameter Number of variables to try at each split will be automatically filled in with the square root of the number of variables entered in the Training Data tab This default value prevents overfitting the forest to the training data The user can experiment with different values but should not increase this number too much or overfitting may occur The next parameter is Number of trees to grow For best results it is advised to grow between 2 000 and 10 000 trees Smaller values can be used for quick experimentation with RF but larger numbers of trees should be used for effective analysis The Random Seed parameter is used to seed the pseudo random number generator This value is useful when reproducing the results of prior experiments Proximity based weights can be used for cluster correlated data with the same replicate outcomes within a subject This can be done by selecting the Proximity based weights checkbox 4 Whe
12. t rff Example 1504003 2 1234567 100 302539 1262150 T 43 7579 2 1 9 95 1 55 1 0 55 2 45 1 0 45 0 42 64 2 4 9 50 1 46 10 46 1 4 1 0 42 502 4 8 150 1 47 3958 3 0 6 96 1 50 1481 3 1 7 54 1 39 1 0 390 53 1404 0 4 9 57 1 gt 2 5 19 e 46 1 0 46 2 Last 0 1 2 54 0357 3 1 6 56 1 1666607 2 92 52 3 I I o0 1 32 Loeb Q2 2 1 0 2 1 1504003 2 4 50 30 6 150 1 46 10460 53 2308 3 1 7 104 1 52 5077 2 4 9 65 1 35 2051 0 5 9 39 1 12 the rest of the trees The numbers in the file are as follows 1504003 number of samples number of variables have ids clustered outcomes number of classes have ids 0 1 O if no IDs were present in the training data 1 if IDs were present clustered outcomes 0 1 O if training data is not clustered and or outcomes were not the same for all replicates within a subject 1 if training data was clustered with outcomes for all replicates within a subject belonging to the same class If this is 1 then subject level classifications and error rates are produced in addition to the sample level classification and error rate 2 number of variables to split at each node 1234567 random seed 100 number of trees 50 3 1 6 150 1 Gini score split variable split variable value sample size reached this node class Gini score is used to decide which variable and its value will produce the best separation of training data into distinct clas

Download Pdf Manuals

image

Related Search

Related Contents

  DeLOCK 2 x USB 3.0 19-pin - 4 x USB 3.0-A  Samsung SH12ZWHD/XFO Инструкция по использованию  Impex WM-354 User's Manual  Avvitatore pneumo-idraulico ad arresto Istruzioni per l`uso P2119BA/IT  iQUBE2 Quick-Start Guide - Rice Lake Weighing Systems  Toshiba Satellite P870/05P  TOUGH GUN TA3 Robotic Air-Cooled Owner`s Manual  Dossier de Spécification de Projet Webmestre  

Copyright © All rights reserved.
Failed to retrieve file