We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
On p3.2xlarge (V100) (current benchmark):
On g5.4xlarge (A10G) (newer, but lower-end GPU):
Roughly the same speed.
On g6.4xlarge (L4) (newer, but lower-end GPU):
On lambdalabs gpu_1x_a100_sxm4 (A100) (newer GPU):
Faster (1.4-1.9x) on the largest dataset, but about the same speed on the small/medium sized data.
On lambdalabs gpu_1x_h100_pcie (H100) (newest, most powerful GPU):
Faster (1.3-2x) on the largest dataset, but about the same speed on the small/medium sized data.
The text was updated successfully, but these errors were encountered:
Summary XGBoost 10M size:
GPU specs (table by ChatGPT):
GPU Benchmark by LambdaLabs:
data from the bar plot:
LambdaCloud H100 80GB PCIe Gen5: ~5.2 LambdaCloud A100 40GB PCIe: ~3.5 LambdaCloud A10: ~1.3 LambdaCloud V100 16GB: 1.0 (reference)
Sorry, something went wrong.
No branches or pull requests
On p3.2xlarge (V100) (current benchmark):
On g5.4xlarge (A10G) (newer, but lower-end GPU):
Roughly the same speed.
On g6.4xlarge (L4) (newer, but lower-end GPU):
Roughly the same speed.
On lambdalabs gpu_1x_a100_sxm4 (A100) (newer GPU):
Faster (1.4-1.9x) on the largest dataset, but about the same speed on the small/medium sized data.
On lambdalabs gpu_1x_h100_pcie (H100) (newest, most powerful GPU):
Faster (1.3-2x) on the largest dataset, but about the same speed on the small/medium sized data.
The text was updated successfully, but these errors were encountered: