-
Notifications
You must be signed in to change notification settings - Fork 175
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
Jay Chia
committed
Feb 1, 2024
1 parent
59af587
commit 5fcc1cc
Showing
3 changed files
with
41 additions
and
13 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -3,4 +3,4 @@ Integrations | |
|
||
.. toctree:: | ||
|
||
integrations/data_catalogs | ||
integrations/iceberg |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,40 @@ | ||
Apache Iceberg | ||
============== | ||
|
||
`Apache Iceberg <https://iceberg.apache.org/>`_ is an open-sourced table format originally developed at Netflix for large-scale analytical datasets. | ||
|
||
To read from the Apache Iceberg table format, use the :func:`daft.read_iceberg` function. | ||
|
||
We integrate closely with `PyIceberg <https://py.iceberg.apache.org/>`_ (the official Python implementation for Apache Iceberg) and allow the reading of Daft dataframes easily from PyIceberg's Table objects. | ||
|
||
.. code:: python | ||
# Access a PyIceberg table as per normal | ||
from pyiceberg.catalog import load_catalog | ||
catalog = load_catalog("my_iceberg_catalog") | ||
table = catalog.load_table("my_namespace.my_table") | ||
# Create a Daft Dataframe | ||
import daft | ||
df = daft.read_iceberg(table) | ||
Daft currently natively supports: | ||
|
||
1. **Distributed Reads:** Daft will fully distribute the I/O of reads over your compute resources (whether Ray or on multithreading on the local PyRunner) | ||
2. **Skipping filtered data:** Daft uses ``df.where(...)`` filter calls to only read data that matches your predicates | ||
3. **All Catalogs From PyIceberg:** Daft is natively integrated with PyIceberg, and supports all the catalogs that PyIceberg does! | ||
|
||
Selecting a Table | ||
***************** | ||
|
||
Daft currently leverages PyIceberg for catalog/table discovery. Please consult `PyIceberg documentation <https://py.iceberg.apache.org/api/#load-a-table>`_ for more details on how to load a table! | ||
|
||
Roadmap | ||
******* | ||
|
||
Here are features of Iceberg that are works-in-progress. | ||
|
||
1. Iceberg V2 merge-on-read features | ||
2. Writing back to an Iceberg table (appends, overwrites, upserts) |