Where communities thrive

  • Join over 1.5M+ people
  • Join over 100K+ communities
  • Free without limits
  • Create your own community
Repo info
    Tommy Yang
    Tommy Yang
    I used xgboost4j-0.80.jar, xgboost train parameter of round is 800 and train data is 2000000. When I use predictleaf to get leafIndex, the jvm crashed.


    A fatal error has been detected by the Java Runtime Environment:


    SIGSEGV (0xb) at pc=0x00007f42160bf902, pid=880, tid=0x00007f42175f2700


    JRE version: Java(TM) SE Runtime Environment (8.0_171-b11) (build 1.8.0_171-b11)

    Java VM: Java HotSpot(TM) 64-Bit Server VM (25.171-b11 mixed mode linux-amd64 compressed oops)

    Problematic frame:

    V [libjvm.so+0x6d6902] jni_SetFloatArrayRegion+0xc2


    Core dump written. Default location: /data/suzhe/suzhe-1.0-SNAPSHOT/core or core.880


    If you would like to submit a bug report, please visit:



    Java frames: (J=compiled Java code, j=interpreted, Vv=VM code)
    j ml.dmlc.xgboost4j.java.XGBoostJNI.XGBoosterPredict(JJII[[F)I+0
    j ml.dmlc.xgboost4j.java.Booster.predict(Lml/dmlc/xgboost4j/java/DMatrix;ZIZZ)[[F+45
    j ml.dmlc.xgboost4j.java.Booster.predictLeaf(Lml/dmlc/xgboost4j/java/DMatrix;I)[[F+6
    j com.jianshu.suzhe.LRTrainer.train()V+23
    j com.jianshu.suzhe.LRTrainer.main([Ljava/lang/String;)V+30
    v ~StubRoutines::call_stub
    Stack: [0x00007f42174f2000,0x00007f42175f3000], sp=0x00007f42175f1590, free space=1021k
    Native frames: (J=compiled Java code, j=interpreted, Vv=VM code, C=native code)
    V [libjvm.so+0x6d6902] jni_SetFloatArrayRegion+0xc2
    C [libxgboost4j8098523902211486429.so+0x9001c] Java_ml_dmlc_xgboost4j_java_XGBoostJNI_XGBoosterPredict+0x5c
    Can someone suggest how to solve this problem?
    Adrian Nembach
    Hi, I am having a problem with xgboost4j on linux. More specifically, if I specify eta to be smaller than 1 e.g. 0.3 (which is the default), than the model doesn't seem to learn anything. It almost seems like eta is set to zero somewhere along the line. Does anyone else experience this problem or maybe knows a solution? Thanks in advance.
    Lijo Varghese

    Hi i am stuck in my work in submitting a spark job to hadoop yarn master in cluster mode
    please find my environment setup below

    i have a linux machine having 128 GB of RAM, 2TB Hard disk, 2x16 cores.
    i have set up cloud era hadoop containers in a docker mount point having 50 GB(this mount point is almost full). i have one datanode, namenode and yarnmaster containers runnings.
    i am submitting spark job from my host machine to run Rscript in cluster mode. R server and libraries are set up in datanode.
    When i submit the spark job it remains in the accepted state for long time. please find the spark submit command i am using below
    spark-submit --master yarn --name RechargeModel --deploy-mode cluster --executor-memory 3G --num-executors 4 rechargemodel.R

    Hi currently seeing
    a segmentation fault when I try to import xgboost
    Are there developers here? Is there interest in adding conformal predictions to the library? The error on the predicted value is an often desired quantity. Here is a link to the paper about how it was done with Random Forrest. https://link.springer.com/article/10.1007/s10994-014-5453-0 This method should work with anything which has our of bag samples.
    Lukas Heumos

    Hi everyone,

    odd request: I need a non reproducible dataset & xgboost model.
    Does anybody have any pointers or in the best case both?

    Apoorv Shrivastava
    has anybody worked on dask-xgboost...need some help
    Harshit Gupta
    @apoorv22 Yeah, What’s the issue ?
    Apoorv Shrivastava
    @harshit-2115 not able to convert a model create from dask-xgboost to PMML using sklearn2pmml
    Peng Yu
    has anyone encountered problems with negative hessian ? for some reason i felt it’s being swallowed
    Fabio "Draco" Fonseca
    Hi, is there a way to use the spark version using pure java? I cannot use scala in my company..
    Apoorv Shrivastava
    posting it here too
    Hello! I've a question about XGBRegressor. I'm training a model with 1M observations and I get an r2 of 0.80. If I partition the train set in two (500K observation each) and train them independently with same hyperparameters used for full training set I get an r2 of 0.95 and 0.94. Question is: how can I improve the r2 for the full training set?
    Hamza Mohd. Zubair
    import xgboost throwing segmentation fault
    Hemang Joshi
    please paste whole error @hamzamohdzubair
    Hamza Mohd. Zubair
    @hemangjoshi37a The problem was resolved on Github issues page, within the hour. See issue: dmlc/xgboost#6480. Would like to thank Philip Hyunsu Cho: https://github.com/hcho3
    Joey Gao
    Hi, is any one
    Hi, Why is the node gain output from xgboost different from that calculated manually? Can anyone answer this question? Thanks a lot!