Efficiency of tar_make() with {crew} vs tar_make_clustermq() #1079
Closed
wlandau-lilly
started this conversation in
General
Replies: 2 comments
-
Ultimately, targets that run quickly are much better off with |
Beta Was this translation helpful? Give feedback.
0 replies
-
So after further profiling studies, I think the limitation is in |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Help
Description
This
crew
test with lots of small targets:targets/tests/hpc/test-crew_local.R
Lines 85 to 121 in 7463ece
is slower than the equivalent
clustermq
test:targets/tests/hpc/test-clustermq_local.R
Lines 150 to 181 in 7463ece
proffer
flame graphs do not show an overt bottleneck, so it's hard to say what might be slowing things down. It might have to do with my solution to #1074. Reverting it seems to shave a couple seconds off.On the other hand, maybe the above test is not truly representative of the work that takes place in real projects. In a proprietary internal simulation study,
tar_make_clustermq()
took 2.007 minutes, whereastar_make()
with the equivalent number of persistentcrew
workers took about 2.661 minutes. From previous profiling studies and from looking at what is in the pipeline, I believe I can explain this difference in terms of serialization.clustermq
supports "common data", whereascrew
andmirai
do not (c.f. wlandau/crew#33).Beta Was this translation helpful? Give feedback.
All reactions