Skip to content

Practical Federated Gradient Boosting Decision Trees (AAAI 2020)

Notifications You must be signed in to change notification settings

Xtra-Computing/SimFL

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

This is the code of paper Practical Federated Gradient Boosting Decision Trees. The implementation is based on a previous version of ThunderGBM. Only Linux-based operating systems are supported.

Installation

Prerequisites

  • CMake
  • CUDA

Build

git submodule init
git submodule update
mkdir build
cd build
cmake ..
make -j

Usage

SimFL currently only works for binary classification tasks with labels 0 and 1 and requires GPUs.

Parameteres

* -p: int, number of parties (default:2)
* -t: int, number of lsh tables (default:40)
* -b: int, number of buckets (default:500)
* -r: float, r value of LSH function (default:4.0)
* -s: int, init seed for LSH
* -f: string, path to the dataset file
* -d: int, the maximum dimension of the datasets

Datasets

Please rename all the local datasets in such format: name+'_train'+party_id, e.g., a9a_train0, a9a_train1. For the test dataset, please rename it in such format: name+'_test', e.g., a9a_test.

Sample command:

Under build directory

./src/test/thundergbm-test -p 2 -t 30 -b 500 -r 4 -s -1 -f ../dataset/a9a/a9a -d 123 -n 50 -e 8

About

Practical Federated Gradient Boosting Decision Trees (AAAI 2020)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages