Skip to content

Commit

Permalink
Merge branch 'u/morriscb/debugForS3FS'
Browse files Browse the repository at this point in the history
  • Loading branch information
morriscb committed Jul 2, 2024
2 parents bb3f4c9 + 6d11195 commit 7247dab
Show file tree
Hide file tree
Showing 20 changed files with 157 additions and 27 deletions.
4 changes: 4 additions & 0 deletions intro.md
Original file line number Diff line number Diff line change
Expand Up @@ -58,6 +58,10 @@ Available notebooks:
(Kimberly Siletti)**

## Release Notes
* **[abc_atlas_access (v0.110)]**
* Fixed compatibility issue with read only local caches specifically for
fuse style file mounts such as those used on CodeOcean.
* Added local cache example to notebooks.
* **[abc_atlas_access (v0.1.0)]**
* Fixed issue with loading 10X mouse data with the `get_gene_data` function.
* **[Spring 2024 Public Beta (version 20240330)]**
Expand Down
8 changes: 7 additions & 1 deletion notebooks/10x_snRNASeq_tutorial_part_1.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -36,6 +36,8 @@
"source": [
"We will interact with the data using the **AbcProjectCache**. This cache object tracks which data has been downloaded and serves the path to the requsted data on disk. For metadata, the cache can also directly serve a up a Pandas Dataframe. See the ``getting_started`` notebook for more details on using the cache including installing it if it has not already been.\n",
"\n",
"The commented section that calls ``from_local_cache`` can be used instead when a download of the data already exists on disk or running on CodeOcean with the attached Allen Brain Cell Atlas or similar mounting of the AWS s3 bucket as a directory through [s3fs-fuse](https://github.com/s3fs-fuse/s3fs-fuse) or similar.\n",
"\n",
"**Change the download_base variable to where you have downloaded the data in your system.**"
]
},
Expand All @@ -58,6 +60,10 @@
"source": [
"download_base = Path('../../abc_download_root')\n",
"abc_cache = AbcProjectCache.from_s3_cache(download_base)\n",
"\n",
"# download_base = Path('../../data/abc_atlas') # Path to the already downloaded data or s3fs-fuse mount.\n",
"# abc_cache = AbcProjectCache.from_local_cache(download_base)\n",
"\n",
"abc_cache.current_manifest"
]
},
Expand Down Expand Up @@ -2421,7 +2427,7 @@
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.11.8"
"version": "3.11.9"
}
},
"nbformat": 4,
Expand Down
8 changes: 7 additions & 1 deletion notebooks/10x_snRNASeq_tutorial_part_2a.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -35,6 +35,8 @@
"source": [
"We will interact with the data using the **AbcProjectCache**. This cache object tracks which data has been downloaded and serves the path to the requsted data on disk. For metadata, the cache can also directly serve a up a Pandas Dataframe. See the ``getting_started`` notebook for more details on using the cache including installing it if it has not already been.\n",
"\n",
"The commented section that calls ``from_local_cache`` can be used instead when a download of the data already exists on disk or running on CodeOcean with the attached Allen Brain Cell Atlas or similar mounting of the AWS s3 bucket as a directory through [s3fs-fuse](https://github.com/s3fs-fuse/s3fs-fuse) or similar.\n",
"\n",
"**Change the download_base variable to where you have downloaded the data in your system.**"
]
},
Expand All @@ -57,6 +59,10 @@
"source": [
"download_base = Path('../../abc_download_root')\n",
"abc_cache = AbcProjectCache.from_s3_cache(download_base)\n",
"\n",
"# download_base = Path('../../data/abc_atlas') # Path to the already downloaded data or s3fs-fuse mount.\n",
"# abc_cache = AbcProjectCache.from_local_cache(download_base)\n",
"\n",
"abc_cache.current_manifest"
]
},
Expand Down Expand Up @@ -554,7 +560,7 @@
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.11.8"
"version": "3.11.9"
}
},
"nbformat": 4,
Expand Down
8 changes: 7 additions & 1 deletion notebooks/10x_snRNASeq_tutorial_part_2b.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -34,6 +34,8 @@
"source": [
"We will interact with the data using the **AbcProjectCache**. This cache object tracks which data has been downloaded and serves the path to the requsted data on disk. For metadata, the cache can also directly serve a up a Pandas Dataframe. See the ``getting_started`` notebook for more details on using the cache including installing it if it has not already been.\n",
"\n",
"The commented section that calls ``from_local_cache`` can be used instead when a download of the data already exists on disk or running on CodeOcean with the attached Allen Brain Cell Atlas or similar mounting of the AWS s3 bucket as a directory through [s3fs-fuse](https://github.com/s3fs-fuse/s3fs-fuse) or similar.\n",
"\n",
"**Change the download_base variable to where you have downloaded the data in your system.**"
]
},
Expand All @@ -56,6 +58,10 @@
"source": [
"download_base = Path('../../abc_download_root')\n",
"abc_cache = AbcProjectCache.from_s3_cache(download_base)\n",
"\n",
"# download_base = Path('../../data/abc_atlas') # Path to the already downloaded data or s3fs-fuse mount.\n",
"# abc_cache = AbcProjectCache.from_local_cache(download_base)\n",
"\n",
"abc_cache.current_manifest"
]
},
Expand Down Expand Up @@ -590,7 +596,7 @@
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.11.8"
"version": "3.11.9"
}
},
"nbformat": 4,
Expand Down
10 changes: 8 additions & 2 deletions notebooks/WHB-10x_snRNASeq_tutorial_part_1.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,9 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"We will interact with the data using the **AbcProjectCache**. This cache object tracks which data has been downloaded and serves the path to the requsted data on disk. For metadata, the cache can also directly serve a up a Pandas Dataframe from the stored csv file. See the ``getting_started`` notebook for more details on using the cache including installing it if it has not already been.\n",
"We will interact with the data using the **AbcProjectCache**. This cache object tracks which data has been downloaded and serves the path to the requsted data on disk. For metadata, the cache can also directly serve a up a Pandas Dataframe. See the ``getting_started`` notebook for more details on using the cache including installing it if it has not already been.\n",
"\n",
"The commented section that calls ``from_local_cache`` can be used instead when a download of the data already exists on disk or running on CodeOcean with the attached Allen Brain Cell Atlas or similar mounting of the AWS s3 bucket as a directory through [s3fs-fuse](https://github.com/s3fs-fuse/s3fs-fuse) or similar.\n",
"\n",
"**Change the download_base variable to where you have downloaded the data in your system.**"
]
Expand All @@ -60,6 +62,10 @@
"source": [
"download_base = Path('../../abc_download_root')\n",
"abc_cache = AbcProjectCache.from_s3_cache(download_base)\n",
"\n",
"# download_base = Path('../../data/abc_atlas') # Path to the already downloaded data or s3fs-fuse mount.\n",
"# abc_cache = AbcProjectCache.from_local_cache(download_base)\n",
"\n",
"abc_cache.current_manifest"
]
},
Expand Down Expand Up @@ -1346,7 +1352,7 @@
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.11.8"
"version": "3.11.9"
}
},
"nbformat": 4,
Expand Down
8 changes: 7 additions & 1 deletion notebooks/WHB-10x_snRNASeq_tutorial_part_2.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -34,6 +34,8 @@
"source": [
"We will interact with the data using the **AbcProjectCache**. This cache object tracks which data has been downloaded and serves the path to the requsted data on disk. For metadata, the cache can also directly serve a up a Pandas Dataframe. See the ``getting_started`` notebook for more details on using the cache including installing it if it has not already been.\n",
"\n",
"The commented section that calls ``from_local_cache`` can be used instead when a download of the data already exists on disk or running on CodeOcean with the attached Allen Brain Cell Atlas or similar mounting of the AWS s3 bucket as a directory through [s3fs-fuse](https://github.com/s3fs-fuse/s3fs-fuse) or similar.\n",
"\n",
"**Change the download_base variable to where you have downloaded the data in your system.**"
]
},
Expand All @@ -56,6 +58,10 @@
"source": [
"download_base = Path('../../abc_download_root')\n",
"abc_cache = AbcProjectCache.from_s3_cache(download_base)\n",
"\n",
"# download_base = Path('../../data/abc_atlas') # Path to the already downloaded data or s3fs-fuse mount.\n",
"# abc_cache = AbcProjectCache.from_local_cache(download_base)\n",
"\n",
"abc_cache.current_manifest"
]
},
Expand Down Expand Up @@ -1061,7 +1067,7 @@
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.11.8"
"version": "3.11.9"
}
},
"nbformat": 4,
Expand Down
8 changes: 7 additions & 1 deletion notebooks/WHB_cluster_annotation_tutorial.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -34,6 +34,8 @@
"source": [
"We will interact with the data using the **AbcProjectCache**. This cache object tracks which data has been downloaded and serves the path to the requsted data on disk. For metadata, the cache can also directly serve a up a Pandas Dataframe. See the ``getting_started`` notebook for more details on using the cache including installing it if it has not already been.\n",
"\n",
"The commented section that calls ``from_local_cache`` can be used instead when a download of the data already exists on disk or running on CodeOcean with the attached Allen Brain Cell Atlas or similar mounting of the AWS s3 bucket as a directory through [s3fs-fuse](https://github.com/s3fs-fuse/s3fs-fuse) or similar.\n",
"\n",
"**Change the download_base variable to where you have downloaded the data in your system.**"
]
},
Expand All @@ -56,6 +58,10 @@
"source": [
"download_base = Path('../../abc_download_root')\n",
"abc_cache = AbcProjectCache.from_s3_cache(download_base)\n",
"\n",
"# download_base = Path('../../data/abc_atlas') # Path to the already downloaded data or s3fs-fuse mount.\n",
"# abc_cache = AbcProjectCache.from_local_cache(download_base)\n",
"\n",
"abc_cache.current_manifest"
]
},
Expand Down Expand Up @@ -1605,7 +1611,7 @@
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.11.8"
"version": "3.11.9"
}
},
"nbformat": 4,
Expand Down
10 changes: 8 additions & 2 deletions notebooks/ccf_and_parcellation_annotation_tutorial.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -38,6 +38,8 @@
"source": [
"We will interact with the data using the **AbcProjectCache**. This cache object tracks which data has been downloaded and serves the path to the requsted data on disk. For metadata, the cache can also directly serve a up a Pandas Dataframe. See the ``getting_started`` notebook for more details on using the cache including installing it if it has not already been.\n",
"\n",
"The commented section that calls ``from_local_cache`` can be used instead when a download of the data already exists on disk or running on CodeOcean with the attached Allen Brain Cell Atlas or similar mounting of the AWS s3 bucket as a directory through [s3fs-fuse](https://github.com/s3fs-fuse/s3fs-fuse) or similar.\n",
"\n",
"**Change the download_base variable to where you have downloaded the data in your system.**"
]
},
Expand All @@ -58,8 +60,12 @@
}
],
"source": [
"download_base = pathlib.Path('../../abc_download_root')\n",
"download_base = Path('../../abc_download_root')\n",
"abc_cache = AbcProjectCache.from_s3_cache(download_base)\n",
"\n",
"# download_base = Path('../../data/abc_atlas') # Path to the already downloaded data or s3fs-fuse mount.\n",
"# abc_cache = AbcProjectCache.from_local_cache(download_base)\n",
"\n",
"abc_cache.current_manifest"
]
},
Expand Down Expand Up @@ -2659,7 +2665,7 @@
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.11.8"
"version": "3.11.9"
}
},
"nbformat": 4,
Expand Down
10 changes: 8 additions & 2 deletions notebooks/cluster_annotation_tutorial.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -36,6 +36,8 @@
"source": [
"We will interact with the data using the **AbcProjectCache**. This cache object tracks which data has been downloaded and serves the path to the requsted data on disk. For metadata, the cache can also directly serve a up a Pandas Dataframe. See the ``getting_started`` notebook for more details on using the cache including installing it if it has not already been.\n",
"\n",
"The commented section that calls ``from_local_cache`` can be used instead when a download of the data already exists on disk or running on CodeOcean with the attached Allen Brain Cell Atlas or similar mounting of the AWS s3 bucket as a directory through [s3fs-fuse](https://github.com/s3fs-fuse/s3fs-fuse) or similar.\n",
"\n",
"**Change the download_base variable to where you have downloaded the data in your system.**"
]
},
Expand All @@ -56,8 +58,12 @@
}
],
"source": [
"download_base = pathlib.Path('../../abc_download_root')\n",
"download_base = Path('../../abc_download_root')\n",
"abc_cache = AbcProjectCache.from_s3_cache(download_base)\n",
"\n",
"# download_base = Path('../../data/abc_atlas') # Path to the already downloaded data or s3fs-fuse mount.\n",
"# abc_cache = AbcProjectCache.from_local_cache(download_base)\n",
"\n",
"abc_cache.current_manifest"
]
},
Expand Down Expand Up @@ -1917,7 +1923,7 @@
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.11.8"
"version": "3.11.9"
}
},
"nbformat": 4,
Expand Down
10 changes: 8 additions & 2 deletions notebooks/cluster_groups_and_embeddings_tutorial.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -31,6 +31,8 @@
"source": [
"We will interact with the data using the **AbcProjectCache**. This cache object tracks which data has been downloaded and serves the path to the requsted data on disk. For metadata, the cache can also directly serve a up a Pandas Dataframe. See the ``getting_started`` notebook for more details on using the cache including installing it if it has not already been.\n",
"\n",
"The commented section that calls ``from_local_cache`` can be used instead when a download of the data already exists on disk or running on CodeOcean with the attached Allen Brain Cell Atlas or similar mounting of the AWS s3 bucket as a directory through [s3fs-fuse](https://github.com/s3fs-fuse/s3fs-fuse) or similar.\n",
"\n",
"**Change the download_base variable to where you have downloaded the data in your system.**"
]
},
Expand All @@ -51,8 +53,12 @@
}
],
"source": [
"download_base = pathlib.Path('../../abc_download_root')\n",
"download_base = Path('../../abc_download_root')\n",
"abc_cache = AbcProjectCache.from_s3_cache(download_base)\n",
"\n",
"# download_base = Path('../../data/abc_atlas') # Path to the already downloaded data or s3fs-fuse mount.\n",
"# abc_cache = AbcProjectCache.from_local_cache(download_base)\n",
"\n",
"abc_cache.current_manifest"
]
},
Expand Down Expand Up @@ -1379,7 +1385,7 @@
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.11.8"
"version": "3.11.9"
}
},
"nbformat": 4,
Expand Down
10 changes: 8 additions & 2 deletions notebooks/cluster_neighborhood_gallery.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -31,6 +31,8 @@
"source": [
"We will interact with the data using the **AbcProjectCache**. This cache object tracks which data has been downloaded and serves the path to the requsted data on disk. For metadata, the cache can also directly serve a up a Pandas Dataframe. See the ``getting_started`` notebook for more details on using the cache including installing it if it has not already been.\n",
"\n",
"The commented section that calls ``from_local_cache`` can be used instead when a download of the data already exists on disk or running on CodeOcean with the attached Allen Brain Cell Atlas or similar mounting of the AWS s3 bucket as a directory through [s3fs-fuse](https://github.com/s3fs-fuse/s3fs-fuse) or similar.\n",
"\n",
"**Change the download_base variable to where you have downloaded the data in your system.**"
]
},
Expand All @@ -51,8 +53,12 @@
}
],
"source": [
"download_base = pathlib.Path('../../abc_download_root')\n",
"download_base = Path('../../abc_download_root')\n",
"abc_cache = AbcProjectCache.from_s3_cache(download_base)\n",
"\n",
"# download_base = Path('../../data/abc_atlas') # Path to the already downloaded data or s3fs-fuse mount.\n",
"# abc_cache = AbcProjectCache.from_local_cache(download_base)\n",
"\n",
"abc_cache.current_manifest"
]
},
Expand Down Expand Up @@ -878,7 +884,7 @@
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.11.8"
"version": "3.11.9"
}
},
"nbformat": 4,
Expand Down
8 changes: 7 additions & 1 deletion notebooks/general_accessing_10x_snRNASeq_tutorial.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -33,7 +33,9 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"We will interact with the data using the **AbcProjectCache**. This cache object tracks which data has been downloaded and serves the path to the requsted data on disk. For metadata, the cache can also directly serve up a Pandas Dataframe. See the ``getting_started`` notebook for more details on using the cache including installing it if it has not already been.\n",
"We will interact with the data using the **AbcProjectCache**. This cache object tracks which data has been downloaded and serves the path to the requsted data on disk. For metadata, the cache can also directly serve a up a Pandas Dataframe. See the ``getting_started`` notebook for more details on using the cache including installing it if it has not already been.\n",
"\n",
"The commented section that calls ``from_local_cache`` can be used instead when a download of the data already exists on disk or running on CodeOcean with the attached Allen Brain Cell Atlas or similar mounting of the AWS s3 bucket as a directory through [s3fs-fuse](https://github.com/s3fs-fuse/s3fs-fuse) or similar.\n",
"\n",
"**Change the download_base variable to where you have downloaded the data in your system.**"
]
Expand All @@ -57,6 +59,10 @@
"source": [
"download_base = Path('../../abc_download_root')\n",
"abc_cache = AbcProjectCache.from_s3_cache(download_base)\n",
"\n",
"# download_base = Path('../../data/abc_atlas') # Path to the already downloaded data or s3fs-fuse mount.\n",
"# abc_cache = AbcProjectCache.from_local_cache(download_base)\n",
"\n",
"abc_cache.current_manifest"
]
},
Expand Down
8 changes: 7 additions & 1 deletion notebooks/getting_started.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -144,7 +144,9 @@
"\n",
"Below we show how to setup up the cache to download from S3, how to list and switch to a different data release, and additionally how to list the directories available, their size, and the files in that directory.\n",
"\n",
"Setup the **AbcProjectCache** object by specifying a directory and calling ``from_s3_cache`` as shown below. We also print what version of the manifest is being currently loaded by the cache."
"Setup the **AbcProjectCache** object by specifying a directory and calling ``from_s3_cache`` as shown below. We also print what version of the manifest is being currently loaded by the cache.\n",
"\n",
"The commented section that calls ``from_local_cache`` can be used instead when a download of the data already exists on disk or running on CodeOcean with the attached Allen Brain Cell Atlas or similar mounting of the AWS s3 bucket as a directory through [s3fs-fuse](https://github.com/s3fs-fuse/s3fs-fuse) or similar. Using the cache from a s3fs mount means that all files will be streamed over the internet instead of saved to the local disk."
]
},
{
Expand All @@ -166,6 +168,10 @@
"source": [
"download_base = Path('../../abc_download_root') # Path to where you would like to write the downloaded data.\n",
"abc_cache = AbcProjectCache.from_s3_cache(download_base)\n",
"\n",
"# download_base = Path('../../data/abc_atlas') # Path to the already downloaded data or s3fs-fuse mount.\n",
"# abc_cache = AbcProjectCache.from_local_cache(download_base)\n",
"\n",
"abc_cache.current_manifest"
]
},
Expand Down
Loading

0 comments on commit 7247dab

Please sign in to comment.