79627287

Date: 2025-05-18 09:10:10
Score: 8.5 🚩
Natty: 6.5
Report link

I’m using sparse vectors with about 10 features out of a possible 50 million. However, the conversion to dense vectors is causing heap exhaustion. Is there a way to disable the sparse-to-dense conversion?

Right now, I can’t even train on a small batch of vectors without running into memory issues — but I ultimately need to train on 200 million rows.

Any help would be greatly appreciated. I’m using XGBoost4j-Spark version 3.0.0 with the Java.

Thanks!

Reasons:
  • Blacklisted phrase (0.5): Thanks
  • Blacklisted phrase (1): appreciated
  • Blacklisted phrase (1): Any help
  • Blacklisted phrase (1): Is there a way
  • RegEx Blacklisted phrase (3): Any help would be greatly appreciated
  • No code block (0.5):
  • Contains question mark (0.5):
  • Low reputation (1):
Posted by: Kristijan Stepanov