Skip to content

Commit

Permalink
feat(flink): implement array operators
Browse files Browse the repository at this point in the history
  • Loading branch information
mfatihaktas committed Jan 31, 2024
1 parent 2a92306 commit 31b2f61
Show file tree
Hide file tree
Showing 4 changed files with 188 additions and 40 deletions.
42 changes: 39 additions & 3 deletions ibis/backends/flink/registry.py
Original file line number Diff line number Diff line change
Expand Up @@ -282,7 +282,11 @@ def _floor_divide(translator: ExprTranslator, op: ops.Node) -> str:
return f"FLOOR(({left}) / ({right}))"


def _array_index(translator: ExprTranslator, op: ops.arrays.ArrayIndex):
def _array(translator: ExprTranslator, op: ops.Array) -> str:
return f"ARRAY[{', '.join(map(translator.translate, op.exprs))}]"


def _array_index(translator: ExprTranslator, op: ops.ArrayIndex):
table_column = op.arg
index = op.index

Expand All @@ -292,10 +296,35 @@ def _array_index(translator: ExprTranslator, op: ops.arrays.ArrayIndex):
return f"{table_column_translated} [ {index_translated} + 1 ]"


def _array_length(translator: ExprTranslator, op: ops.arrays.ArrayLength) -> str:
def _array_length(translator: ExprTranslator, op: ops.ArrayLength) -> str:
return f"CARDINALITY({translator.translate(op.arg)})"


def _array_position(translator: ExprTranslator, op: ops.ArrayPosition) -> str:
arg = translator.translate(op.arg)
other = translator.translate(op.other)
return f"ARRAY_POSITION({arg}, {other}) - 1"


def _array_slice(translator: ExprTranslator, op: ops.ArraySlice) -> str:
array = translator.translate(op.arg)
start = op.start.value
# The offsets are 1-based for ARRAY_SLICE.
# Ref: https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/functions/systemfunctions
if start >= 0:
start += 1

if op.stop is None:
return f"ARRAY_SLICE({array}, {start})"

stop = op.stop.value
if stop >= 0:
return f"ARRAY_SLICE({array}, {start}, {stop})"
else:
# To imitate the behavior of pandas array slicing.
return f"ARRAY_SLICE({array}, {start}, CARDINALITY({array}) - {abs(stop)})"


def _json_get_item(translator: ExprTranslator, op: ops.json.JSONGetItem) -> str:
arg_translated = translator.translate(op.arg)
if op.index.dtype.is_integer():
Expand Down Expand Up @@ -532,9 +561,16 @@ def _struct_field(translator, op):
# Binary operations
ops.Power: fixed_arity("power", 2),
ops.FloorDivide: _floor_divide,
# Collection functions
# Collection operations
ops.Array: _array,
ops.ArrayContains: fixed_arity("ARRAY_CONTAINS", 2),
ops.ArrayDistinct: fixed_arity("ARRAY_DISTINCT", 1),
ops.ArrayIndex: _array_index,
ops.ArrayLength: _array_length,
ops.ArrayPosition: _array_position,
ops.ArrayRemove: fixed_arity("ARRAY_REMOVE", 2),
ops.ArraySlice: _array_slice,
ops.ArrayUnion: fixed_arity("ARRAY_UNION", 2),
ops.JSONGetItem: _json_get_item,
ops.Map: _map,
ops.MapGet: _map_get,
Expand Down
3 changes: 2 additions & 1 deletion ibis/backends/flink/tests/conftest.py
Original file line number Diff line number Diff line change
Expand Up @@ -52,12 +52,13 @@ def connect(*, tmpdir, worker_id, **kw: Any):
def _load_data(self, **_: Any) -> None:
import pandas as pd

from ibis.backends.tests.data import json_types, struct_types, win
from ibis.backends.tests.data import array_types, json_types, struct_types, win

for table_name in TEST_TABLES:
path = self.data_dir / "parquet" / f"{table_name}.parquet"
self.connection.create_table(table_name, pd.read_parquet(path), temp=True)

self.connection.create_table("array_types", array_types, temp=True)
self.connection.create_table("json_t", json_types, temp=True)
self.connection.create_table("struct", struct_types, temp=True)
self.connection.create_table("win", win, temp=True)
Expand Down
Loading

0 comments on commit 31b2f61

Please sign in to comment.