-
Notifications
You must be signed in to change notification settings - Fork 1.9k
/
LightGbmArguments.cs
414 lines (348 loc) · 19.9 KB
/
LightGbmArguments.cs
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
// Licensed to the .NET Foundation under one or more agreements.
// The .NET Foundation licenses this file to you under the MIT license.
// See the LICENSE file in the project root for more information.
using System.Collections.Generic;
using System.Text;
using System.Reflection;
using Microsoft.ML.Runtime;
using Microsoft.ML.Runtime.CommandLine;
using Microsoft.ML.Runtime.EntryPoints;
using Microsoft.ML.Runtime.Internal.Internallearn;
using Microsoft.ML.Runtime.LightGBM;
[assembly: LoadableClass(typeof(LightGbmArguments.TreeBooster), typeof(LightGbmArguments.TreeBooster.Arguments),
typeof(SignatureLightGBMBooster), LightGbmArguments.TreeBooster.FriendlyName, LightGbmArguments.TreeBooster.Name)]
[assembly: LoadableClass(typeof(LightGbmArguments.DartBooster), typeof(LightGbmArguments.DartBooster.Arguments),
typeof(SignatureLightGBMBooster), LightGbmArguments.DartBooster.FriendlyName, LightGbmArguments.DartBooster.Name)]
[assembly: LoadableClass(typeof(LightGbmArguments.GossBooster), typeof(LightGbmArguments.GossBooster.Arguments),
typeof(SignatureLightGBMBooster), LightGbmArguments.GossBooster.FriendlyName, LightGbmArguments.GossBooster.Name)]
[assembly: EntryPointModule(typeof(LightGbmArguments.TreeBooster.Arguments))]
[assembly: EntryPointModule(typeof(LightGbmArguments.DartBooster.Arguments))]
[assembly: EntryPointModule(typeof(LightGbmArguments.GossBooster.Arguments))]
namespace Microsoft.ML.Runtime.LightGBM
{
public delegate void SignatureLightGBMBooster();
[TlcModule.ComponentKind("BoosterParameterFunction")]
public interface ISupportBoosterParameterFactory : IComponentFactory<IBoosterParameter>
{
}
public interface IBoosterParameter
{
void UpdateParameters(Dictionary<string, object> res);
}
/// <summary>
/// Parameters names comes from LightGBM library.
/// See https://github.com/Microsoft/LightGBM/blob/master/docs/Parameters.rst.
/// </summary>
public sealed class LightGbmArguments : LearnerInputBaseWithGroupId
{
public abstract class BoosterParameter<TArgs> : IBoosterParameter
where TArgs : class, new()
{
protected TArgs Args { get; }
protected BoosterParameter(TArgs args)
{
Args = args;
}
/// <summary>
/// Update the parameters by specific Booster, will update parameters into "res" directly.
/// </summary>
public virtual void UpdateParameters(Dictionary<string, object> res)
{
FieldInfo[] fields = Args.GetType().GetFields();
foreach (var field in fields)
res[GetArgName(field.Name)] = field.GetValue(Args).ToString();
}
}
private static string GetArgName(string name)
{
StringBuilder strBuf = new StringBuilder();
bool first = true;
foreach (char c in name)
{
if (char.IsUpper(c))
{
if (first)
first = false;
else
strBuf.Append('_');
strBuf.Append(char.ToLower(c));
}
else
strBuf.Append(c);
}
return strBuf.ToString();
}
public sealed class TreeBooster : BoosterParameter<TreeBooster.Arguments>
{
public const string Name = "gbdt";
public const string FriendlyName = "Tree Booster";
[TlcModule.Component(Name = Name, FriendlyName = FriendlyName, Desc = "Traditional Gradient Boosting Decision Tree.")]
public class Arguments : ISupportBoosterParameterFactory
{
[Argument(ArgumentType.AtMostOnce, HelpText = "Use for binary classification when classes are not balanced.", ShortName = "us")]
public bool UnbalancedSets = false;
[Argument(ArgumentType.AtMostOnce,
HelpText = "Minimum loss reduction required to make a further partition on a leaf node of the tree. the larger, " +
"the more conservative the algorithm will be.")]
[TlcModule.Range(Min = 0.0)]
public double MinSplitGain = 0;
[Argument(ArgumentType.AtMostOnce,
HelpText = "Maximum depth of a tree. 0 means no limit. However, tree still grows by best-first.")]
[TlcModule.Range(Min = 0, Max = int.MaxValue)]
public int MaxDepth = 0;
[Argument(ArgumentType.AtMostOnce,
HelpText = "Minimum sum of instance weight(hessian) needed in a child. If the tree partition step results in a leaf " +
"node with the sum of instance weight less than min_child_weight, then the building process will give up further partitioning. In linear regression mode, " +
"this simply corresponds to minimum number of instances needed to be in each node. The larger, the more conservative the algorithm will be.")]
[TlcModule.Range(Min = 0.0)]
public double MinChildWeight = 0.1;
[Argument(ArgumentType.AtMostOnce,
HelpText = "Subsample frequency. 0 means no subsample. "
+ "If subsampleFreq > 0, it will use a subset(ratio=subsample) to train. And the subset will be updated on every Subsample iteratinos.")]
[TlcModule.Range(Min = 0, Max = int.MaxValue)]
public int SubsampleFreq = 0;
[Argument(ArgumentType.AtMostOnce,
HelpText = "Subsample ratio of the training instance. Setting it to 0.5 means that LightGBM randomly collected " +
"half of the data instances to grow trees and this will prevent overfitting. Range: (0,1].")]
[TlcModule.Range(Inf = 0.0, Max = 1.0)]
public double Subsample = 1;
[Argument(ArgumentType.AtMostOnce,
HelpText = "Subsample ratio of columns when constructing each tree. Range: (0,1].",
ShortName = "ff")]
[TlcModule.Range(Inf = 0.0, Max = 1.0)]
public double FeatureFraction = 1;
[Argument(ArgumentType.AtMostOnce,
HelpText = "L2 regularization term on weights, increasing this value will make model more conservative.",
ShortName = "l2")]
[TlcModule.Range(Min = 0.0)]
[TGUI(Label = "Lambda(L2)", SuggestedSweeps = "0,0.5,1")]
[TlcModule.SweepableDiscreteParam("RegLambda", new object[] { 0f, 0.5f, 1f })]
public double RegLambda = 0.01;
[Argument(ArgumentType.AtMostOnce,
HelpText = "L1 regularization term on weights, increase this value will make model more conservative.",
ShortName = "l1")]
[TlcModule.Range(Min = 0.0)]
[TGUI(Label = "Alpha(L1)", SuggestedSweeps = "0,0.5,1")]
[TlcModule.SweepableDiscreteParam("RegAlpha", new object[] { 0f, 0.5f, 1f })]
public double RegAlpha = 0;
[Argument(ArgumentType.AtMostOnce,
HelpText = "Control the balance of positive and negative weights, useful for unbalanced classes." +
" A typical value to consider: sum(negative cases) / sum(positive cases).")]
public double ScalePosWeight = 1;
public virtual IBoosterParameter CreateComponent(IHostEnvironment env) => new TreeBooster(this);
}
public TreeBooster(Arguments args)
: base(args)
{
Contracts.CheckUserArg(Args.MinSplitGain >= 0, nameof(Args.MinSplitGain), "must be >= 0.");
Contracts.CheckUserArg(Args.MinChildWeight >= 0, nameof(Args.MinChildWeight), "must be >= 0.");
Contracts.CheckUserArg(Args.Subsample > 0 && Args.Subsample <= 1, nameof(Args.Subsample), "must be in (0,1].");
Contracts.CheckUserArg(Args.FeatureFraction > 0 && Args.FeatureFraction <= 1, nameof(Args.FeatureFraction), "must be in (0,1].");
Contracts.CheckUserArg(Args.ScalePosWeight > 0 && Args.ScalePosWeight <= 1, nameof(Args.ScalePosWeight), "must be in (0,1].");
}
public override void UpdateParameters(Dictionary<string, object> res)
{
base.UpdateParameters(res);
res["boosting_type"] = Name;
}
}
public class DartBooster : BoosterParameter<DartBooster.Arguments>
{
public const string Name = "dart";
public const string FriendlyName = "Tree Dropout Tree Booster";
[TlcModule.Component(Name = Name, FriendlyName = FriendlyName, Desc = "Dropouts meet Multiple Additive Regresion Trees. See https://arxiv.org/abs/1505.01866")]
public class Arguments : TreeBooster.Arguments
{
[Argument(ArgumentType.AtMostOnce, HelpText = "Drop ratio for trees. Range:(0,1).")]
[TlcModule.Range(Inf = 0.0, Max = 1.0)]
public double DropRate = 0.1;
[Argument(ArgumentType.AtMostOnce, HelpText = "Max number of dropped tree in a boosting round.")]
[TlcModule.Range(Inf = 0, Max = int.MaxValue)]
public int MaxDrop = 1;
[Argument(ArgumentType.AtMostOnce, HelpText = "Probability for not perform dropping in a boosting round.")]
[TlcModule.Range(Inf = 0.0, Max = 1.0)]
public double SkipDrop = 0.5;
[Argument(ArgumentType.AtMostOnce, HelpText = "True will enable xgboost dart mode.")]
public bool XgboostDartMode = false;
[Argument(ArgumentType.AtMostOnce, HelpText = "True will enable uniform drop.")]
public bool UniformDrop = false;
public override IBoosterParameter CreateComponent(IHostEnvironment env) => new DartBooster(this);
}
public DartBooster(Arguments args)
: base(args)
{
Contracts.CheckUserArg(Args.DropRate > 0 && Args.DropRate < 1, nameof(Args.DropRate), "must be in (0,1).");
Contracts.CheckUserArg(Args.MaxDrop > 0, nameof(Args.MaxDrop), "must be > 0.");
Contracts.CheckUserArg(Args.SkipDrop >= 0 && Args.SkipDrop < 1, nameof(Args.SkipDrop), "must be in [0,1).");
}
public override void UpdateParameters(Dictionary<string, object> res)
{
base.UpdateParameters(res);
res["boosting_type"] = Name;
}
}
public class GossBooster : BoosterParameter<GossBooster.Arguments>
{
public const string Name = "goss";
public const string FriendlyName = "Gradient-based One-Size Sampling";
[TlcModule.Component(Name = Name, FriendlyName = FriendlyName, Desc = "Gradient-based One-Side Sampling.")]
public class Arguments : TreeBooster.Arguments
{
[Argument(ArgumentType.AtMostOnce,
HelpText = "Retain ratio for large gradient instances.")]
[TlcModule.Range(Inf = 0.0, Max = 1.0)]
public double TopRate = 0.2;
[Argument(ArgumentType.AtMostOnce,
HelpText =
"Retain ratio for small gradient instances.")]
[TlcModule.Range(Inf = 0.0, Max = 1.0)]
public double OtherRate = 0.1;
public override IBoosterParameter CreateComponent(IHostEnvironment env) => new GossBooster(this);
}
public GossBooster(Arguments args)
: base(args)
{
Contracts.CheckUserArg(Args.TopRate > 0 && Args.TopRate < 1, nameof(Args.TopRate), "must be in (0,1).");
Contracts.CheckUserArg(Args.OtherRate >= 0 && Args.OtherRate < 1, nameof(Args.TopRate), "must be in [0,1).");
Contracts.Check(Args.TopRate + Args.OtherRate <= 1, "Sum of topRate and otherRate cannot be larger than 1.");
}
public override void UpdateParameters(Dictionary<string, object> res)
{
base.UpdateParameters(res);
res["boosting_type"] = Name;
}
}
public enum EvalMetricType
{
DefaultMetric,
Rmse,
Mae,
Logloss,
Error,
Merror,
Mlogloss,
Auc,
Ndcg,
Map
};
[Argument(ArgumentType.AtMostOnce, HelpText = "Number of iterations.", SortOrder = 1, ShortName = "iter")]
[TGUI(Label = "Number of boosting iterations", SuggestedSweeps = "10,20,50,100,150,200")]
[TlcModule.SweepableDiscreteParam("NumBoostRound", new object[] { 10, 20, 50, 100, 150, 200 })]
public int NumBoostRound = 100;
[Argument(ArgumentType.AtMostOnce,
HelpText = "Shrinkage rate for trees, used to prevent over-fitting. Range: (0,1].",
SortOrder = 2, ShortName = "lr", NullName = "<Auto>")]
[TGUI(Label = "Learning Rate", SuggestedSweeps = "0.025-0.4;log")]
[TlcModule.SweepableFloatParamAttribute("LearningRate", 0.025f, 0.4f, isLogScale: true)]
public double? LearningRate;
[Argument(ArgumentType.AtMostOnce, HelpText = "Maximum leaves for trees.",
SortOrder = 2, ShortName = "nl", NullName = "<Auto>")]
[TGUI(Description = "The maximum number of leaves per tree", SuggestedSweeps = "2-128;log;inc:4")]
[TlcModule.SweepableLongParamAttribute("NumLeaves", 2, 128, isLogScale: true, stepSize: 4)]
public int? NumLeaves;
[Argument(ArgumentType.AtMostOnce, HelpText = "Minimum number of instances needed in a child.",
SortOrder = 2, ShortName = "mil", NullName = "<Auto>")]
[TGUI(Label = "Min Documents In Leaves", SuggestedSweeps = "1,10,20,50 ")]
[TlcModule.SweepableDiscreteParamAttribute("MinDataPerLeaf", new object[] { 1, 10, 20, 50 })]
public int? MinDataPerLeaf;
[Argument(ArgumentType.AtMostOnce, HelpText = "Max number of bucket bin for features.", ShortName = "mb")]
public int MaxBin = 255;
[Argument(ArgumentType.Multiple, HelpText = "Which booster to use, can be gbtree, gblinear or dart. gbtree and dart use tree based model while gblinear uses linear function.", SortOrder = 3)]
public ISupportBoosterParameterFactory Booster = new TreeBooster.Arguments();
[Argument(ArgumentType.AtMostOnce, HelpText = "Verbose", ShortName = "v")]
public bool VerboseEval = false;
[Argument(ArgumentType.AtMostOnce, HelpText = "Printing running messages.")]
public bool Silent = true;
[Argument(ArgumentType.AtMostOnce, HelpText = "Number of parallel threads used to run LightGBM.", ShortName = "nt")]
public int? NThread;
[Argument(ArgumentType.AtMostOnce,
HelpText = "Evaluation metrics.",
ShortName = "em")]
public EvalMetricType EvalMetric = EvalMetricType.DefaultMetric;
[Argument(ArgumentType.AtMostOnce, HelpText = "Use softmax loss for the multi classification.")]
[TlcModule.SweepableDiscreteParam("UseSoftmax", new object[] { true, false})]
public bool? UseSoftmax;
[Argument(ArgumentType.AtMostOnce, HelpText = "Rounds of early stopping, 0 will disable it.",
ShortName = "es")]
public int EarlyStoppingRound = 0;
[Argument(ArgumentType.AtMostOnce, HelpText = "Comma seperated list of gains associated to each relevance label.", ShortName = "gains")]
[TGUI(Label = "Ranking Label Gain")]
public string CustomGains = "0,3,7,15,31,63,127,255,511,1023,2047,4095";
[Argument(ArgumentType.AtMostOnce, HelpText = "Number of entries in a batch when loading data.", Hide = true)]
public int BatchSize = 1 << 20;
[Argument(ArgumentType.AtMostOnce, HelpText = "Enable categorical split or not.", ShortName = "cat")]
[TlcModule.SweepableDiscreteParam("UseCat", new object[] { true, false })]
public bool? UseCat;
[Argument(ArgumentType.AtMostOnce, HelpText = "Enable missing value auto infer or not.")]
[TlcModule.SweepableDiscreteParam("UseMissing", new object[] { true, false })]
public bool UseMissing = false;
[Argument(ArgumentType.AtMostOnce, HelpText = "Min number of instances per categorical group.", ShortName = "mdpg")]
[TlcModule.Range(Inf = 0, Max = int.MaxValue)]
[TlcModule.SweepableDiscreteParam("MinDataPerGroup", new object[] { 10, 50, 100, 200 })]
public int MinDataPerGroup = 100;
[Argument(ArgumentType.AtMostOnce, HelpText = "Max number of categorical thresholds.", ShortName = "maxcat")]
[TlcModule.Range(Inf = 0, Max = int.MaxValue)]
[TlcModule.SweepableDiscreteParam("MaxCatThreshold", new object[] { 8, 16, 32, 64 })]
public int MaxCatThreshold = 32;
[Argument(ArgumentType.AtMostOnce, HelpText = "Lapalace smooth term in categorical feature spilt. Avoid the bias of small categories.")]
[TlcModule.Range(Min = 0.0)]
[TlcModule.SweepableDiscreteParam("CatSmooth", new object[] { 1, 10, 20 })]
public double CatSmooth = 10;
[Argument(ArgumentType.AtMostOnce, HelpText = "L2 Regularization for categorical split.")]
[TlcModule.Range(Min = 0.0)]
[TlcModule.SweepableDiscreteParam("CatL2", new object[] { 0.1, 0.5, 1, 5, 10 })]
public double CatL2 = 10;
[Argument(ArgumentType.Multiple, HelpText = "Parallel LightGBM Learning Algorithm", ShortName = "parag")]
public ISupportParallel ParallelTrainer = new SingleTrainerFactory();
internal Dictionary<string, object> ToDictionary(IHost host)
{
Contracts.CheckValue(host, nameof(host));
Contracts.CheckUserArg(MaxBin > 0, nameof(MaxBin), "must be > 0.");
Dictionary<string, object> res = new Dictionary<string, object>();
var boosterParams = Booster.CreateComponent(host);
boosterParams.UpdateParameters(res);
res[GetArgName(nameof(MaxBin))] = MaxBin.ToString();
res["verbose"] = Silent ? "-1" : "1";
if (NThread.HasValue)
res["nthread"] = NThread.Value.ToString();
res["seed"] = host.Rand.Next().ToString();
string metric = null;
switch (EvalMetric)
{
case EvalMetricType.DefaultMetric:
break;
case EvalMetricType.Mae:
metric = "l1";
break;
case EvalMetricType.Logloss:
metric = "binary_logloss";
break;
case EvalMetricType.Error:
metric = "binary_error";
break;
case EvalMetricType.Merror:
metric = "multi_error";
break;
case EvalMetricType.Mlogloss:
metric = "multi_logloss";
break;
case EvalMetricType.Rmse:
case EvalMetricType.Auc:
case EvalMetricType.Ndcg:
case EvalMetricType.Map:
metric = EvalMetric.ToString().ToLower();
break;
}
if (!string.IsNullOrEmpty(metric))
res["metric"] = metric;
res["sigmoid"] = "0.5";
res["label_gain"] = CustomGains;
res[GetArgName(nameof(UseMissing))] = UseMissing.ToString();
res[GetArgName(nameof(MinDataPerGroup))] = MinDataPerGroup.ToString();
res[GetArgName(nameof(MaxCatThreshold))] = MaxCatThreshold.ToString();
res[GetArgName(nameof(CatSmooth))] = CatSmooth.ToString();
res[GetArgName(nameof(CatL2))] = CatL2.ToString();
return res;
}
}
}