You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
there seems to be a bug for the inner join for DNDSparse; NDSparse join works and can be tested with join(Ascalar, Bscalar)
for a join, ideally the names of the data columns should be different in the two ndsparse (if you do not want Bscalar or Ascalar data values to replace one another, in which case it is correct)
The example above works for join(ddb, Bscalar; how=:outer)
EDIT: after your fix, an empty table would be the correct result.
@zgornel Thank you! I can't believe I didn't notice I was doing the wrong kind of join.
I do get the correct result with an outer join, with one caveat: No matter how many tables I join to the original distributed table, the number of chunks remains the same. I guess I'm not sure what the default behavior should be, so this behavior makes sense, even though I didn't expect it. However, I'm not sure how to join to a distributed table and increase the number of chunks, which is important in my use-case because each joined table individually approaches the memory limit of my machine. Does anyone know how to do this? Does it warrant its own issue? At the very least, I would expect documentation on this and I'm happy to provide that if I understand how to do it.
I'll leave this issue open because of the remaining problem that join errors uninformatively when it should return an empty distributed table.
gives
I tried the naive solution of adding a
Base.convert(::Type{T}, x::Nullable{T}) where T = x.value
but then the join returns an empty table.The text was updated successfully, but these errors were encountered: