❗ This repository is still experimental. No API-compatibility is guaranteed.
Rikai is a framework specifically designed for AI workflows focused around large scale unstructured datasets (e.g., images, videos, sensor data (future), text (future), and more). Through every stage of the AI modeling workflow, Rikai strives to offer a great developer experience when working with real-world AI datasets.
The quality of an AI dataset can make or break an AI project, but tooling for AI data is sorely lacking in ergonomics. As a result, practitioners must spend most of their time and effort wrestling with their data instead of innovating on the models and use cases. Rikai alleviates the pain that AI practitioners experience on a daily basis dealing with the myriad of tedious data tasks, so they can focus again on model-building and problem solving.
To start trying Rikai right away, checkout the Quickstart Guide.
The core of Rikai is a data format ("rikai format") based on Apache Parquet. Rikai augments parquet with a rich collection of semantic types design specifically for unstructured data and annotations.
Rikai comes with an extensive set of I/O connectors. For ETL, Rikai is able to consume popular formats like ROS bags and Coco. For analysis, it's easy to read Rikai data into pandas/spark DataFrames (Rikai handles serde for the semantic types). And for training, Rikai allows direct creation of Pytorch/Tensorflow datasets without manual conversion.
Rikai extends Spark SQL with ML capability which allows users to analyze Rikai datasets using own models with SQL ("Bring your own model")
Carefully crafted data-visualization embedded with semantic types, especially in Jupyter notebooks, to help you visualize and inspect your AI data without having to remember complicated raw image manipulations.
- Improved video support
- Text / sensors / geospatial support
- Versioning support built into the dataset
- Better Rikai UDT-support
- Declarative annotation API (think vega-lite for annotating images/videos)
- Integrations into dbt and BI tools
from pyspark.sql import Row
from pyspark.ml.linalg import DenseMatrix
from rikai.types import Image, Box2d
from rikai.numpy import wrap
import numpy as np
df = spark.createDataFrame(
[
{
"id": 1,
"mat": DenseMatrix(2, 2, range(4)),
"image": Image("s3://foo/bar/1.png"),
"annotations": [
Row(
label="cat",
mask=wrap(np.random.rand(256, 256)),
bbox=Box2d(xmin=1.0, ymin=2.0, xmax=3.0, ymax=4.0),
)
],
}
]
)
df.write.format("rikai").save("s3://path/to/features")
Train dataset in Pytorch
from torch.utils.data import DataLoader
from torchvision import transforms as T
from rikai.pytorch.vision import Dataset
transform = T.Compose([
T.Resize(640),
T.ToTensor(),
T.Normalize((0.485, 0.456, 0.406), (0.229, 0.224, 0.225))
])
dataset = Dataset(
"s3://path/to/features",
image_column="image",
transform=transform
)
loader = DataLoader(
dataset,
batch_size=32,
num_workers=8,
)
for batch in loader:
predicts = model(batch.to("cuda"))
Using a ML model in Spark SQL (experiemental)
CREATE MODEL yolo5
OPTIONS (min_confidence=0.3, device="gpu", batch_size=32)
USING "s3://bucket/to/yolo5_spec.yaml";
SELECT id, ML_PREDICT(yolo5, image) FROM my_dataset
WHERE split = "train" LIMIT 100;
Rikai can use MLflow as its model registry. This allows you to automatically pickup the latest model version if you're using the mlflow model registry. Here is a list of supported model flavors:
- PyTorch (pytorch)
- Tensorflow (tensorflow)
- Scikit-learn (sklearn)
CREATE MODEL yolo5
OPTIONS (min_confidence=0.3, device="gpu", batch_size=32)
USING "mlflow:///yolo5_model/";
SELECT id, ML_PREDICT(yolo5, image) FROM my_dataset
WHERE split = "train" LIMIT 100;
For more details on the model spec, see SQL-ML documentation
Currently Rikai is maintained for Scala 2.12 and Python 3.7, 3.8, 3.9
There are multiple ways to install Rikai:
- Try it using the included Dockerfile.
- Install via pip
pip install rikai
, with extras for gcp, pytorch/tf, and others. - Install from source
Note: if you want to use Rikai with your own pyspark, please consult rikai documentation for tips.
The included Dockerfile creates a standalone demo image with Jupyter, Pytorch, Spark, and rikai preinstalled with notebooks for you to play with the capabilities of the rikai feature store.
To build and run the docker image from the current directory:
# Clone the repo
git clone git@github.com:eto-ai/rikai rikai
# Build the docker image
docker build --tag rikai --network host .
# Run the image
docker run -p 0.0.0.0:8888:8888/tcp rikai:latest jupyter lab -ip 0.0.0.0 --port 8888
If successful, the console should then print out a clickable link to JupyterLab. You can also
open a browser tab and go to localhost:8888
.
Base rikai library can be installed with just pip install rikai
. Dependencies for supporting
pytorch (pytorch and torchvision), jupyter (matplotlib and jupyterlab) are all part of
optional extras. Many open-source datasets also use Youtube videos so we've also added pafy and
youtube-dl as optional extras as well.
For example, if you want to use pytorch in Jupyter to train models on rikai datasets in s3 containing Youtube videos you would run:
pip install rikai[pytorch,jupyter,youtube]
If you're not sure what you need and don't mind installing some extra dependencies, you can simply install everything:
pip install rikai[all]
To build from source you'll need python as well as Scala with sbt installed:
# Clone the repo
git clone git@github.com:eto-ai/rikai rikai
# Build the jar
sbt publishLocal
# Install python package
cd python
pip install -e . # pip install -e .[all] to install all optional extras (see "Install from pypi")
pre-commit can be helpful in keep consistent code format with the repository. It can trigger reformat and extra things in your local machine before the CI force you to do it.
If you want it, install and enable pre-commit
pip install pre-commit
pre-commit install #in your local development directory
#pre-commit installed at .git/hooks/pre-commit
If you want to uninstall it, it would be easy, too.
pre-commit uninstall