whylogs is an open source standard for data and ML logging
whylogs logging agent is the easiest way to enable logging, testing, and monitoring in an ML/AI application. The lightweight agent profiles data in real time, collecting thousands of metrics from structured data, unstructured data, and ML model predictions with zero configuration.
whylogs can be installed in any Python, Java or Spark environment; it can be deployed as a container and run as a sidecar; or invoked through various ML tools (see integrations).
whylogs is designed by data scientists, ML engineers and distributed systems engineers to log data in the most cost-effective, scalable and accurate manner. No sampling. No post-processing. No manual configurations.
whylogs is released under the Apache 2.0 open source license. It supports many languages and is easy to extend. This repo contains the whylogs CLI, language SDKs, and individual libraries are in their own repos.
This is a Python implementation of whylogs. The Java implementation can be found here.
If you have any questions, comments, or just want to hang out with us, please join our Slack channel.
Install whylogs using the pip package manager by running
pip install whylogs
-
Download the source code by cloning the repository or by pressing Download ZIP on this page.
-
You'll need to install poetry in order to install dependencies using the lock file in this project. Follow their docs to get it set up.
-
Run the following comand at the root of the source code:
make install
make
whylogs is easy to get up and runnings
from whylogs import get_or_create_session
import pandas as pd
session = get_or_create_session()
df = pd.read_csv("path/to/file.csv")
with session.logger(dataset_name="my_dataset") as logger:
#dataframe
logger.log_dataframe(df)
#dict
logger.log({"name": 1})
#images
logger.log_images("path/to/image.png")
whylogs collects approximate statistics and sketches of data on a column-basis into a statistical profile. These metrics include:
- Simple counters: boolean, null values, data types.
- Summary statistics: sum, min, max, median, variance.
- Unique value counter or cardinality: tracks an approximate unique value of your feature using HyperLogLog algorithm.
- Histograms for numerical features. whyLogs binary output can be queried to with dynamic binning based on the shape of your data.
- Top frequent items (default is 128). Note that this configuration affects the memory footprint, especially for text features.
To view your logger profiles you can use, methods within whylogs.viz
:
vizualization = ProfileVisualizer()
vizualization.set_profiles([profile_day_1, profile_day_2])
figure= vizualization.plot_distribution("<feature_name>")
figure.savefig("/my/image/path.png")
Individual profiles are saved to disk, AWS S3, or WhyLabs API, automatically when loggers are closed, per the configuration found in the Session configuration.
Current profiles from active loggers can be loaded from memory with:
profile = logger.profile()
You can also load a local profile viewer, where you upload the json
summary file. The default path for the json files is set as output/{dataset_name}/{session_id}/json/dataset_profile.json
.
from whylogs.viz import profile_viewer
profile_viewer()
This will open a viewer on your default browser where you can load a profile json summary, using the Select JSON profile
button:
Once the json is selected you can view your profile's features and
associated and statistics.
The documentation of this package is generated automatically.
- Accurate data profiling: whylogs calculates statistics from 100% of the data, never requiring sampling, ensuring an accurate representation of data distributions
- Lightweight runtime: whylogs utilizes approximate statistical methods to achieve minimal memory footprint that scales with the number of features in the data
- Any architecture: whylogs scales with your system, from local development mode to live production systems in multi-node clusters, and works well with batch and streaming architectures
- Configuration-free: whylogs infers the schema of the data, requiring zero manual configuration to get started
- Tiny storage footprint: whylogs turns data batches and streams into statistical fingerprints, 10-100MB uncompressed
- Unlimited metrics: whylogs collects all possible statistical metrics about structured or unstructured data
Whylogs supports both structured and unstructured data, specifically:
Data type | Features | Notebook Example |
---|---|---|
Structured Data | Distribution, cardinality, schema, counts, missing values | Getting started with structured data |
Images | exif metadata, derived pixels features, bounding boxes | Getting started with images |
Video | In development | Github Issue #214 |
Tensors | derived 1d features (more in developement) | Github Issue #216 |
Text | top k values, counts, cardinality | String Features |
Audio | In developement | Github Issue #212 |
Integration | Features | Resources |
---|---|---|
Spark | Run whylogs in Apache Spark environment | |
Pandas | Log and monitor any pandas dataframe | |
Kafka | Log and monitor Kafka topics with whylogs | |
MLflow | Enhance MLflow metrics with whylogs: | |
Github actions | Unit test data with whylogs and github actions | |
RAPIDS | Use whylogs in RAPIDS environment | |
Java | Run whylogs in Java environment | |
Docker | Run whylogs as in Docker | |
AWS S3 | Store whylogs profiles in S3 |
For a full set of our examples, please check out whylogs-examples.
Check out our example notebooks with Binder:
whylogs is maintained by WhyLabs.
If you have any questions, comments, or just want to hang out with us, please join our Slack channel.
We welcome contributions to whylogs. Please see our contribution guide and our developement guide for details.