• Login
    View Item 
    •   MINDS@UW Home
    • MINDS@UW Madison
    • College of Letters and Science, University of Wisconsin–Madison
    • Department of Computer Sciences, UW-Madison
    • CS Technical Reports
    • View Item
    •   MINDS@UW Home
    • MINDS@UW Madison
    • College of Letters and Science, University of Wisconsin–Madison
    • Department of Computer Sciences, UW-Madison
    • CS Technical Reports
    • View Item
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Explore Optimal Degree of Parallelism for Distributed XGBoost Training

    Thumbnail
    File(s)
    TR1867 Junda Chen 1.pdf (1.525Mb)
    Date
    2021-09-01
    Author
    Chen, Junda
    Akash, Aditya Kumar
    Suzuki, Yukiko
    Metadata
    Show full item record
    Abstract
    The XGBoost has been an extremely popular and effective machine learning method which gained its fame throughwinning multiple Kaggle competitions. One of its strengths lies in parallel processing which makes the computationsalable and faster than its counterparts. On the other hand, there are system configurations and model tuning param-eters which need to be adjusted in order to achieve its full potential cost-effectively. In this paper, we explore how thetraining duration changes under different workloads, system configurations and framework parameters. By running mul-tiple of these experiments, practical insights can be learned and applied for the future applications of XGBoost methods.
    Subject
    XGBoost
    Distributed Machine Learning
    Big DataSystem
    Permanent Link
    http://digital.library.wisc.edu/1793/82297
    Type
    Technical Report
    Citation
    TR1867
    Part of
    • CS Technical Reports

    Contact Us | Send Feedback
     

     

    Browse

    All of MINDS@UWCommunities & CollectionsBy Issue DateAuthorsTitlesSubjectsThis CollectionBy Issue DateAuthorsTitlesSubjects

    My Account

    Login

    Contact Us | Send Feedback