Skip to content

Commit

Permalink
FIX-modin-project#2601: reduce data size for some asv tests
Browse files Browse the repository at this point in the history
Signed-off-by: Anatoly Myachev <anatoly.myachev@intel.com>
  • Loading branch information
anmyachev committed Jan 13, 2021
1 parent 477c5f6 commit c11f6f6
Showing 1 changed file with 17 additions and 2 deletions.
19 changes: 17 additions & 2 deletions asv_bench/benchmarks/benchmarks.py
Original file line number Diff line number Diff line change
Expand Up @@ -75,7 +75,16 @@ def setup(self, data_size, ncols=1):

class TimeMultiColumnGroupby(BaseTimeGroupBy):
param_names = ["data_size", "ncols"]
params = [UNARY_OP_DATA_SIZE, [6]]
params = [
[
(5000, 5000),
(10_000, 10),
# TODO: after optimization try to use UNARY_OP_DATA_SIZE here
]
if ASV_DATASET_SIZE == "Big"
else UNARY_OP_DATA_SIZE,
[6],
]

def time_groupby_agg_quan(self, data_size, ncols):
execute(self.df.groupby(by=self.groupby_columns).agg("quantile"))
Expand Down Expand Up @@ -150,7 +159,13 @@ def time_join(self, data_size, how, sort):
class TimeMerge:
param_names = ["data_size", "how", "sort"]
params = [
BINARY_OP_DATA_SIZE,
[
(5000, 5000, 5000, 5000),
(125_000, 15, 100_000, 10),
# TODO: after optimization try to use BINARY_OP_DATA_SIZE here
]
if ASV_DATASET_SIZE == "Big"
else BINARY_OP_DATA_SIZE,
["left", "inner"],
[False],
]
Expand Down

0 comments on commit c11f6f6

Please sign in to comment.