-
Notifications
You must be signed in to change notification settings - Fork 67
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Merge pull request #104 from databrickslabs/feature/starter-notebooks…
…-main Feature/starter notebooks main
- Loading branch information
Showing
12 changed files
with
918 additions
and
362 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,40 @@ | ||
# Databricks notebook source | ||
# MAGIC %md | ||
# MAGIC ## Setup temporary data location | ||
# MAGIC We will setup a temporary location to store our New York City Neighbourhood shapes. </br> | ||
|
||
# COMMAND ---------- | ||
|
||
user_name = dbutils.notebook.entry_point.getDbutils().notebook().getContext().userName().get() | ||
|
||
raw_path = f"dbfs:/tmp/mosaic/{user_name}" | ||
raw_taxi_zones_path = f"{raw_path}/taxi_zones" | ||
|
||
print(f"The raw data will be stored in {raw_path}") | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC ## Download taxi zones GeoJSON | ||
|
||
# COMMAND ---------- | ||
|
||
import requests | ||
import os | ||
import pathlib | ||
|
||
taxi_zones_url = 'https://data.cityofnewyork.us/api/geospatial/d3c5-ddgc?method=export&format=GeoJSON' | ||
|
||
# The DBFS file system is mounted under /dbfs/ directory on Databricks cluster nodes | ||
|
||
|
||
local_taxi_zones_path = pathlib.Path(raw_taxi_zones_path.replace('dbfs:/', '/dbfs/')) | ||
local_taxi_zones_path.mkdir(parents=True, exist_ok=True) | ||
|
||
req = requests.get(taxi_zones_url) | ||
with open(local_taxi_zones_path / f'nyc_taxi_zones.geojson', 'wb') as f: | ||
f.write(req.content) | ||
|
||
# COMMAND ---------- | ||
|
||
display(dbutils.fs.ls(raw_taxi_zones_path)) |
This file was deleted.
Oops, something went wrong.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,266 @@ | ||
# Databricks notebook source | ||
# MAGIC %md | ||
# MAGIC ## Setup NYC taxi zones | ||
# MAGIC In order to setup the data please run the notebook available at "../../data/DownloadNYCTaxiZones". </br> | ||
# MAGIC DownloadNYCTaxiZones notebook will make sure we have New York City Taxi zone shapes available in our environment. | ||
|
||
# COMMAND ---------- | ||
|
||
user_name = dbutils.notebook.entry_point.getDbutils().notebook().getContext().userName().get() | ||
|
||
raw_path = f"dbfs:/tmp/mosaic/{user_name}" | ||
raw_taxi_zones_path = f"{raw_path}/taxi_zones" | ||
|
||
print(f"The raw data is stored in {raw_path}") | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC ## Enable Mosaic in the notebook | ||
# MAGIC To get started, you'll need to attach the wheel to your cluster and import instances as in the cell below. | ||
|
||
# COMMAND ---------- | ||
|
||
from pyspark.sql.functions import * | ||
import mosaic as mos | ||
mos.enable_mosaic(spark, dbutils) | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md ## Read polygons from GeoJson | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC With the functionality Mosaic brings we can easily load GeoJSON files using spark. </br> | ||
# MAGIC In the past this required GeoPandas in python and conversion to spark dataframe. </br> | ||
|
||
# COMMAND ---------- | ||
|
||
neighbourhoods = ( | ||
spark.read | ||
.option("multiline", "true") | ||
.format("json") | ||
.load(raw_taxi_zones_path) | ||
.select("type", explode(col("features")).alias("feature")) | ||
.select("type", col("feature.properties").alias("properties"), to_json(col("feature.geometry")).alias("json_geometry")) | ||
.withColumn("geometry", mos.st_aswkt(mos.st_geomfromgeojson("json_geometry"))) | ||
) | ||
|
||
display( | ||
neighbourhoods | ||
) | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC ## Compute some basic geometry attributes | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC Mosaic provides a number of functions for extracting the properties of geometries. Here are some that are relevant to Polygon geometries: | ||
|
||
# COMMAND ---------- | ||
|
||
display( | ||
neighbourhoods | ||
.withColumn("calculatedArea", mos.st_area(col("geometry"))) | ||
.withColumn("calculatedLength", mos.st_length(col("geometry"))) | ||
# Note: The unit of measure of the area and length depends on the CRS used. | ||
# For GPS locations it will be square radians and radians | ||
.select("geometry", "calculatedArea", "calculatedLength") | ||
) | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC ## Read points data | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC We will load some Taxi trips data to represent point data. </br> | ||
# MAGIC We already loaded some shapes representing polygons that correspond to NYC neighbourhoods. </br> | ||
|
||
# COMMAND ---------- | ||
|
||
tripsTable = spark.table("delta.`/databricks-datasets/nyctaxi/tables/nyctaxi_yellow`") | ||
|
||
# COMMAND ---------- | ||
|
||
trips = ( | ||
tripsTable | ||
.drop("vendorId", "rateCodeId", "store_and_fwd_flag", "payment_type") | ||
.withColumn("pickup_geom", mos.st_astext(mos.st_point(col("pickup_longitude"), col("pickup_latitude")))) | ||
.withColumn("dropoff_geom", mos.st_astext(mos.st_point(col("dropoff_longitude"), col("dropoff_latitude")))) | ||
) | ||
|
||
# COMMAND ---------- | ||
|
||
display(trips.select("pickup_geom", "dropoff_geom")) | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC ## Spatial Joins | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC We can use Mosaic to perform spatial joins both with and without Mosaic indexing strategies. </br> | ||
# MAGIC Indexing is very important when handling very different geometries both in size and in shape (ie. number of vertices). </br> | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC ### Getting the optimal resolution | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC We can use Mosaic functionality to identify how to best index our data based on the data inside the specific dataframe. </br> | ||
# MAGIC Selecting an apropriate indexing resolution can have a considerable impact on the performance. </br> | ||
|
||
# COMMAND ---------- | ||
|
||
from mosaic import MosaicFrame | ||
|
||
neighbourhoods_mosaic_frame = MosaicFrame(neighbourhoods, "geometry") | ||
optimal_resolution = neighbourhoods_mosaic_frame.get_optimal_resolution(sample_fraction=0.75) | ||
|
||
print(f"Optimal resolution is {optimal_resolution}") | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC Not every resolution will yield performance improvements. </br> | ||
# MAGIC By a rule of thumb it is always better to under-index than over-index - if not sure select a lower resolution. </br> | ||
# MAGIC Higher resolutions are needed when we have very imballanced geometries with respect to their size or with respect to the number of vertices. </br> | ||
# MAGIC In such case indexing with more indices will considerably increase the parallel nature of the operations. </br> | ||
# MAGIC You can think of Mosaic as a way to partition an overly complex row into multiple rows that have a balanced amount of computation each. | ||
|
||
# COMMAND ---------- | ||
|
||
display( | ||
neighbourhoods_mosaic_frame.get_resolution_metrics(sample_rows=150) | ||
) | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC ### Indexing using the optimal resolution | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC We will use mosaic sql functions to index our points data. </br> | ||
# MAGIC Here we will use resolution 9, index resolution depends on the dataset in use. | ||
|
||
# COMMAND ---------- | ||
|
||
tripsWithIndex = (trips | ||
.withColumn("pickup_h3", mos.point_index_geom(col("pickup_geom"), lit(optimal_resolution))) | ||
.withColumn("dropoff_h3", mos.point_index_geom(col("dropoff_geom"), lit(optimal_resolution))) | ||
) | ||
|
||
# COMMAND ---------- | ||
|
||
display(tripsWithIndex) | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC We will also index our neighbourhoods using a built in generator function. | ||
|
||
# COMMAND ---------- | ||
|
||
neighbourhoodsWithIndex = (neighbourhoods | ||
|
||
# We break down the original geometry in multiple smoller mosaic chips, each with its | ||
# own index | ||
.withColumn("mosaic_index", mos.mosaic_explode(col("geometry"), lit(optimal_resolution))) | ||
|
||
# We don't need the original geometry any more, since we have broken it down into | ||
# Smaller mosaic chips. | ||
.drop("json_geometry", "geometry") | ||
) | ||
|
||
# COMMAND ---------- | ||
|
||
display(neighbourhoodsWithIndex) | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC ### Performing the spatial join | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC We can now do spatial joins to both pickup and drop off zones based on geolocations in our datasets. | ||
|
||
# COMMAND ---------- | ||
|
||
pickupNeighbourhoods = neighbourhoodsWithIndex.select(col("properties.borough").alias("pickup_zone"), col("mosaic_index")) | ||
|
||
withPickupZone = ( | ||
tripsWithIndex.join( | ||
pickupNeighbourhoods, | ||
tripsWithIndex["pickup_h3"] == pickupNeighbourhoods["mosaic_index.index_id"] | ||
).where( | ||
# If the borough is a core chip (the chip is fully contained within the geometry), then we do not need | ||
# to perform any intersection, because any point matching the same index will certainly be contained in | ||
# the borough. Otherwise we need to perform an st_contains operation on the chip geometry. | ||
col("mosaic_index.is_core") | mos.st_contains(col("mosaic_index.wkb"), col("pickup_geom")) | ||
).select( | ||
"trip_distance", "pickup_geom", "pickup_zone", "dropoff_geom", "pickup_h3", "dropoff_h3" | ||
) | ||
) | ||
|
||
display(withPickupZone) | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC We can easily perform a similar join for the drop off location. | ||
|
||
# COMMAND ---------- | ||
|
||
dropoffNeighbourhoods = neighbourhoodsWithIndex.select(col("properties.borough").alias("dropoff_zone"), col("mosaic_index")) | ||
|
||
withDropoffZone = ( | ||
withPickupZone.join( | ||
dropoffNeighbourhoods, | ||
withPickupZone["dropoff_h3"] == dropoffNeighbourhoods["mosaic_index.index_id"] | ||
).where( | ||
col("mosaic_index.is_core") | mos.st_contains(col("mosaic_index.wkb"), col("dropoff_geom")) | ||
).select( | ||
"trip_distance", "pickup_geom", "pickup_zone", "dropoff_geom", "dropoff_zone", "pickup_h3", "dropoff_h3" | ||
) | ||
.withColumn("trip_line", mos.st_astext(mos.st_makeline(array(mos.st_geomfromwkt(col("pickup_geom")), mos.st_geomfromwkt(col("dropoff_geom")))))) | ||
) | ||
|
||
display(withDropoffZone) | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC ## Visualise the results in Kepler | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %md | ||
# MAGIC For visualisation there simply arent good options in scala. </br> | ||
# MAGIC Luckily in our notebooks you can easily switch to python just for UI. </br> | ||
# MAGIC Mosaic abstracts interaction with Kepler in python. | ||
|
||
# COMMAND ---------- | ||
|
||
# MAGIC %python | ||
# MAGIC %%mosaic_kepler | ||
# MAGIC withDropoffZone "pickup_h3" "h3" 5000 | ||
|
||
# COMMAND ---------- | ||
|
||
|
Oops, something went wrong.