Skip to content

Commit

Permalink
docs(datafusion): capitalize the 'F' in DataFusion
Browse files Browse the repository at this point in the history
  • Loading branch information
deepyaman committed Jul 20, 2024
1 parent 7e152c4 commit a21de5b
Show file tree
Hide file tree
Showing 3 changed files with 13 additions and 10 deletions.
2 changes: 1 addition & 1 deletion .github/workflows/ibis-backends.yml
Original file line number Diff line number Diff line change
Expand Up @@ -126,7 +126,7 @@ jobs:
extras:
- sqlite
- name: datafusion
title: Datafusion
title: DataFusion
extras:
- datafusion
- name: polars
Expand Down
9 changes: 6 additions & 3 deletions gen_redirects.py
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,8 @@
"/backends/{version}/BigQuery/": "/backends/bigquery/",
"/backends/{version}/Clickhouse/": "/backends/clickhouse/",
"/backends/{version}/Dask/": "/backends/dask/",
"/backends/{version}/Datafusion/": "/backends/datafusion/",
"/backends/{version}/DataFusion/": "/backends/datafusion/",
"/backends/{version}/Datafusion/": "/backends/datafusion/", # For backwards compatibility
"/backends/{version}/Druid/": "/backends/druid/",
"/backends/{version}/DuckDB/": "/backends/duckdb/",
"/backends/{version}/Impala/": "/backends/impala/",
Expand All @@ -30,7 +31,8 @@
"/docs/{version}/backends/BigQuery/": "/backends/bigquery/",
"/docs/{version}/backends/Clickhouse/": "/backends/clickhouse/",
"/docs/{version}/backends/Dask/": "/backends/dask/",
"/docs/{version}/backends/Datafusion/": "/backends/datafusion/",
"/docs/{version}/backends/DataFusion/": "/backends/datafusion/",
"/docs/{version}/backends/Datafusion/": "/backends/datafusion/", # For backwards compatibility
"/docs/{version}/backends/Druid/": "/backends/druid/",
"/docs/{version}/backends/DuckDB/": "/backends/duckdb/",
"/docs/{version}/backends/Impala/": "/backends/impala/",
Expand Down Expand Up @@ -73,7 +75,8 @@
"/backends/BigQuery/": "/backends/bigquery/",
"/backends/Clickhouse/": "/backends/clickhouse/",
"/backends/Dask/": "/backends/dask/",
"/backends/Datafusion/": "/backends/datafusion/",
"/backends/DataFusion/": "/backends/datafusion/",
"/backends/Datafusion/": "/backends/datafusion/", # For backwards compatibility
"/backends/Druid/": "/backends/druid/",
"/backends/DuckDB/": "/backends/duckdb/",
"/backends/Impala/": "/backends/impala/",
Expand Down
12 changes: 6 additions & 6 deletions ibis/backends/datafusion/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,7 @@ def version(self):
def do_connect(
self, config: Mapping[str, str | Path] | SessionContext | None = None
) -> None:
"""Create a Datafusion `Backend` for use with Ibis.
"""Create a DataFusion `Backend` for use with Ibis.
Parameters
----------
Expand Down Expand Up @@ -342,7 +342,7 @@ def register(
table_name
The name of the table
kwargs
Datafusion-specific keyword arguments
DataFusion-specific keyword arguments
Examples
--------
Expand Down Expand Up @@ -436,7 +436,7 @@ def read_csv(
An optional name to use for the created table. This defaults to
a sequentially generated name.
**kwargs
Additional keyword arguments passed to Datafusion loading function.
Additional keyword arguments passed to DataFusion loading function.
Returns
-------
Expand Down Expand Up @@ -464,7 +464,7 @@ def read_parquet(
An optional name to use for the created table. This defaults to
a sequentially generated name.
**kwargs
Additional keyword arguments passed to Datafusion loading function.
Additional keyword arguments passed to DataFusion loading function.
Returns
-------
Expand Down Expand Up @@ -589,7 +589,7 @@ def create_table(
temp: bool = False,
overwrite: bool = False,
):
"""Create a table in Datafusion.
"""Create a table in DataFusion.
Parameters
----------
Expand Down Expand Up @@ -710,7 +710,7 @@ def truncate_table(
def _create_and_drop_memtable(_conn, table_name, tmp_name, overwrite):
"""Workaround inability to overwrite tables in dataframe API.
Datafusion has helper methods for loading in-memory data, but these methods
DataFusion has helper methods for loading in-memory data, but these methods
don't allow overwriting tables.
The SQL interface allows creating tables from existing tables, so we register
the data as a table using the dataframe API, then run a
Expand Down

0 comments on commit a21de5b

Please sign in to comment.