Hi,
I’m looking into the concrete-ml version of xgboost, and the impact of the quantization on the functional performance of the model.
During my tests the algorithm needs a lot more memory than the original algorithm during the training. The size of the original training dataset is ~10 M, I had to reduce it to 2M elements otherwise the system kills the process, even with small hyper-parameters (max_depth=2
, n_estimators=25
). While for the orignial algorithm it works fine with max_depth=3
, n_estimators=250
.
Is this an expected behavior?
The notebook is available at https://github.com/BastienVialla/concrete-xgboost/blob/main/debug_concrete.ipynb