shuDaoNan9

Results 31 comments of shuDaoNan9

> @imatiach-msft Thank you for your Reply! In this case, my dataset have 26320507 rows * 38 columns. And I used a single machine(48 vCore, 96 GiB memory) for test,...

booleans value in ValidationIndicatorCol, emm. But where can we set what kind of metric for our validation? @imatiach-msft thanks!

> hi @JWenBin can you please try: useSingleDatasetMode = True numThreads = num cores - 1 These two PRs should resolve this: > > #1222 #1282 > > In performance...

> hi @JWenBin "the speed improved a lot, but AUC and accuracy becomes too low (less than 0.6)" That is very interesting. In our benchmarking this didn't affect accuracy at...

Every input data is about 48MB±2MB in each task while trainning('setUseBarrierExecutionMode(true)'), But spark history server indicate only 575989/26320507 rows were trained enough time. ![图片](https://user-images.githubusercontent.com/49393828/149073889-41ca10cb-ad3d-431d-ab21-3bdac9e1b364.png)

When I tried to update lightgbmlib from 2.2.350 to 3.2.110, I got this error too. ``` com.microsoft.ml.lightgbm lightgbmlib 2.2.350 ```

直接按照训练时候的样子组织数据传给模型预测分数不就好了吗?不过可以问下PB格式模型无法导出来咋搞

删了历史保存的模型,重跑就跑起来了,o(╯□╰)o。。。。。。。。。。。。。。。。。。

> > > 可能输入里没有 tfrecords 文件。把 train_input_fn 的函数第一句改成 dataset = tf.data.TFRecordDataset(filenames),然后 filename 里包含 tfrecords ,比如”../../cvr_train.tfrecords“ ,cvr 结果就不为 0 你有数据集吗?