DaniFrame commited on
Commit
2a57d88
·
verified ·
1 Parent(s): 1de6268

Upload 3 files

Browse files
abalone_xgboost_seed_290/extra_params.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "Lpls_beta": 0.6870062831574866
3
+ }
abalone_xgboost_seed_290/feature_importance.csv ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ,L1s_importance,L1_importance,L2s_importance,L2_importance,Lpls_importance,Lpl_importance
2
+ Length,0.046116106,0.059359178,0.09559635,0.08101684,0.04700958,0.047475114
3
+ Diameter,0.08347633,0.083421156,0.09657221,0.081145264,0.0735461,0.0685057
4
+ Height,0.095758766,0.082271226,0.10904176,0.12482568,0.0820814,0.0652053
5
+ Whole_weight,0.14149736,0.10289427,0.09803597,0.10497068,0.10299344,0.08895267
6
+ Shucked_weight,0.113597356,0.111627914,0.10594202,0.09149923,0.12201429,0.10130253
7
+ Viscera_weight,0.09624309,0.08488683,0.16501197,0.09503233,0.10133789,0.084858485
8
+ Shell_weight,0.12606904,0.1357582,0.12762637,0.10907371,0.12054574,0.14377396
9
+ Sex_F,0.13824679,0.16443609,0.052349646,0.16653058,0.13603318,0.13090624
10
+ Sex_I,0.06553346,0.07296358,0.04576923,0.036288325,0.084431335,0.13967492
11
+ Sex_M,0.08053033,0.08184645,0.055465773,0.057772715,0.11351491,0.09908052
12
+ cpu_training_time,0.0004123493,0.0009450985,0.0030869574,0.0033165414,0.0005483471,0.0008271249
13
+ cpu_prediction_time,0.00084361684,0.0016405184,0.0031036562,0.0043830452,0.00094981136,0.0018285697
14
+ memory_usage,0.00071206293,0.0011046223,0.003288013,0.003175999,0.0009821351,0.0019112405
15
+ max_depth,0.00084254955,0.0012418531,0.005844905,0.003735444,0.0005734921,0.001572156
16
+ learning_rate,0.002743278,0.0037915446,0.009919717,0.010913548,0.002896176,0.004176276
17
+ n_estimators,0.0018578945,0.0027561309,0.0069090463,0.008278209,0.0038587295,0.0062657716
18
+ model_CatBoostRegressor,0.0011453587,0.0019247631,0.0025640593,0.002565437,0.0020484088,0.0031329282
19
+ model_DecisionTreeRegressor,0.0,0.0,0.0,0.0,0.0,0.0
20
+ model_LGBMRegressor,0.0022861054,0.0025855661,0.004588634,0.0034922657,0.0026682643,0.004976525
21
+ model_RandomForestRegressor,0.00066435145,0.0013605943,0.0026431407,0.0021846676,0.0007676625,0.0016871105
22
+ model_XGBRegressor,0.0014238374,0.0031844915,0.0066405446,0.009799519,0.0011991067,0.0038867183
abalone_xgboost_seed_290/metrics.json ADDED
@@ -0,0 +1,140 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "L1s": {
3
+ "train": {
4
+ "RMSE": 2.0935023727791857,
5
+ "MAPE": 166438453271.0572,
6
+ "R2": 0.11158685923416167,
7
+ "Pearson": 0.6158135307261174,
8
+ "Spearman": 0.5788932809364057
9
+ },
10
+ "validation": {
11
+ "RMSE": 2.2512418812373576,
12
+ "MAPE": 283869471508.3639,
13
+ "R2": 0.004466261028571061,
14
+ "Pearson": 0.1368178445645586,
15
+ "Spearman": 0.15792130051791844
16
+ },
17
+ "test": {
18
+ "RMSE": 2.263817152069494,
19
+ "MAPE": 153009063454.0765,
20
+ "R2": 0.01653331676724823,
21
+ "Pearson": 0.14349039493331098,
22
+ "Spearman": 0.12634328681963605
23
+ }
24
+ },
25
+ "L1": {
26
+ "train": {
27
+ "RMSE": 1.1292430415336119,
28
+ "MAPE": 2051285956564.513,
29
+ "R2": 0.4639736642595673,
30
+ "Pearson": 0.83547799228978,
31
+ "Spearman": 0.6654701454660048
32
+ },
33
+ "validation": {
34
+ "RMSE": 1.440392265819184,
35
+ "MAPE": 2957708023046.919,
36
+ "R2": 0.15167597708796576,
37
+ "Pearson": 0.39547994903364103,
38
+ "Spearman": 0.2750141587023082
39
+ },
40
+ "test": {
41
+ "RMSE": 1.5236750772075849,
42
+ "MAPE": 1923477733846.6677,
43
+ "R2": 0.0677536387712282,
44
+ "Pearson": 0.26277276278228867,
45
+ "Spearman": 0.32684629214048194
46
+ }
47
+ },
48
+ "L2s": {
49
+ "train": {
50
+ "RMSE": 10.294708575419621,
51
+ "MAPE": 1661421470120.9082,
52
+ "R2": 0.23390979634061526,
53
+ "Pearson": 0.7742496077219196,
54
+ "Spearman": 0.4331005508949738
55
+ },
56
+ "validation": {
57
+ "RMSE": 12.23174807287727,
58
+ "MAPE": 1365532803327.7688,
59
+ "R2": 0.030083578208156436,
60
+ "Pearson": 0.22546474307583905,
61
+ "Spearman": 0.1612823990280642
62
+ },
63
+ "test": {
64
+ "RMSE": 11.429551958867165,
65
+ "MAPE": 1712646993977.2002,
66
+ "R2": 0.014953094207992068,
67
+ "Pearson": 0.12681621009155855,
68
+ "Spearman": 0.039821671665158646
69
+ }
70
+ },
71
+ "L2": {
72
+ "train": {
73
+ "RMSE": 6.835865961688753,
74
+ "MAPE": 5402096208562.108,
75
+ "R2": 0.597473031389564,
76
+ "Pearson": 0.8936859978752215,
77
+ "Spearman": 0.5768850543834181
78
+ },
79
+ "validation": {
80
+ "RMSE": 10.746042742273191,
81
+ "MAPE": 9698662997228.479,
82
+ "R2": 0.0992712522445337,
83
+ "Pearson": 0.31860741760215017,
84
+ "Spearman": 0.29415577280933397
85
+ },
86
+ "test": {
87
+ "RMSE": 10.308094457504874,
88
+ "MAPE": 4813369657308.078,
89
+ "R2": 0.023895346752281665,
90
+ "Pearson": 0.20760949916635318,
91
+ "Spearman": 0.35095723401202994
92
+ }
93
+ },
94
+ "Lpls": {
95
+ "train": {
96
+ "RMSE": 0.4949139885794366,
97
+ "MAPE": 40294987191.24139,
98
+ "R2": 0.02065403643381991,
99
+ "Pearson": 0.46055139959952246,
100
+ "Spearman": 0.390851056749001
101
+ },
102
+ "validation": {
103
+ "RMSE": 0.4982910882748077,
104
+ "MAPE": 44031114935.136894,
105
+ "R2": -0.017699252362459506,
106
+ "Pearson": -0.08906291258464431,
107
+ "Spearman": -0.1230321604342047
108
+ },
109
+ "test": {
110
+ "RMSE": 0.508996916353969,
111
+ "MAPE": 36345412545.637085,
112
+ "R2": 0.00250850962538407,
113
+ "Pearson": 0.07289499838790674,
114
+ "Spearman": 0.09835065604346571
115
+ }
116
+ },
117
+ "Lpl": {
118
+ "train": {
119
+ "RMSE": 0.21938048119670445,
120
+ "MAPE": 557715155401.5345,
121
+ "R2": 0.3750460316299302,
122
+ "Pearson": 0.7774319608769594,
123
+ "Spearman": 0.733977406880339
124
+ },
125
+ "validation": {
126
+ "RMSE": 0.25075618095991975,
127
+ "MAPE": 712904530023.2345,
128
+ "R2": 0.08516267902584329,
129
+ "Pearson": 0.2958661148553663,
130
+ "Spearman": 0.24310324831733923
131
+ },
132
+ "test": {
133
+ "RMSE": 0.26860093357902654,
134
+ "MAPE": 528333070735.6516,
135
+ "R2": 0.09307541029081912,
136
+ "Pearson": 0.31046689154329266,
137
+ "Spearman": 0.28800103851210784
138
+ }
139
+ }
140
+ }