Reverse Dependencies of gcsfs
The following projects have a declared dependency on gcsfs:
- jouissance — jouissance
- juftin-scripts — Helpful Python scripts by @juftin
- JupyterReviewer — A general tool to create dashboards for manual review
- kedro-datasets — Kedro-Datasets is where you can find all of Kedro's data connectors.
- kedro-kubeflow — Kedro plugin with Kubeflow Pipelines support
- kedro-vertexai — Kedro plugin with GCP Vertex AI support
- kedro-viz — Kedro-Viz helps visualise Kedro data and analytics pipelines
- kensu — no summary
- kensu-pyspark — no summary
- kglab — A simple abstraction layer in Python for building knowledge graphs
- lamindb_setup — Setup & configure LaminDB.
- launchflow — Python-native infrastructure for the cloud: LaunchFlow provides a Python SDK that automatically creates and connects to production-ready infrastructure (such as Postgres, Redis, etc..) in your own cloud account. LaunchFlow completely removes the need for DevOps allowing you to focus on your application logic.
- lazycls — Python Utilities for the Super Lazy
- levanter — Scalable Training for Foundation Models with Named Tensors and JAX
- lib310 — lib310 Python Package
- lib310-lite — lib310 Lite Python Package
- lilac — Organize unstructured data
- lilacai — Organize unstructured data
- llama-index-readers-gcs — llama-index readers gcs integration
- lndb_setup — LaminDB setup.
- luntaiDs — Make Data Scientist life Easier Tool
- malariagen-data — A package for accessing and analysing MalariaGEN data.
- managed-file-system-operator — A Python manager class implementing a variety of file system operations in a spawned process, supporting AWS (and AWS-like systems), HTTP, and the local file system.
- mk-feature-store — Python SDK for Feast
- ml-pipeline-gen — A tool for generating end-to-end pipelines on GCP.
- ml4floods — Machine learning models for end-to-end flood extent segmentation.
- mldock — A docker tool that helps put machine learning in places that empower ml developers
- mlem — Version and deploy your models following GitOps principles
- mlproject — Project Description
- mlrun — Tracking and config of machine learning runs
- mms-pip — A custom MMS Analytics module for Python3 by the Touchpoint Analytics & Data Discovery
- molfeat — molfeat - the hub for all your molecular featurizers
- mozilla-bigquery-etl — Tooling for building derived datasets in BigQuery
- multidimio — Cloud-native, scalable, and user-friendly multi dimensional energy data!
- nam-pt — Neural Additive Models (PyTorch): Intepretable ML with Neural Nets
- nowcasting-dataset — Nowcasting Dataset
- numerai-reports — Reports for the Numerai machine learning competition
- oceantide — Library for ocean tide prediction
- olorenchemengine — Oloren ChemEngine is a library for molecular property prediction, uncertainty quantification and interpretability. It includes 50+ models and molecular representations under a unified API, which achieves state-of-the-art performances on a variety of molecular property prediction tasks. The diversity of models and representations is achieved by integrating all top-performing methods in the literature as well an in-house methods.
- omicidx — The OmicIDX project collects, reprocesses, and then republishes metadata from multiple public genomics repositories. Included are the NCBI SRA, Biosample, and GEO databases. Publication is via the cloud data warehouse platform Bigquery, a set of performant search and retrieval APIs, and a set of json-format files for easy incorporation into other projects.
- openmetadata-ingestion — Ingestion Framework for OpenMetadata
- openssa — OpenSSA: Small Specialist Agents for Industrial AI
- openssa-dev — OpenSSA - 'Small Specialist Agents' for Industrial AI
- openssm-dev — OpenSSM - 'Small Specialist Models' for Industrial AI
- opsml — Python MLOPs quality control tooling for your production ML workflows
- osds — PyTorch Object Storage Dataset
- osds-channel-preview — PyTorch Object Storage Dataset
- pangeo-forge-recipes — Pipeline tools for building and publishing analysis ready datasets.
- pangeo-forge-runner — Commandline tool to manage pangeo-forge feedstocks
- paperless — A papermill implementation to run notebooks inside dataproc serverless
- papermill — Parameterize and run Jupyter and nteract Notebooks
- ParlPy — API Wrapper and Scraper for UK Parliamentary Bills
- petastorm — Petastorm is a library enabling the use of Parquet storage from Tensorflow, Pytorch, and other Python-based ML training frameworks.
- pgcs — Pgcs is an intuitive TUI tool designed to simplify your interaction with Google Cloud Storage. Stay in your coding zone by navigating directories, searching files (with case-insensitive support), and previewing files all from your terminal. Easily save paths to clipboard or download files with straightforward keyboard shortcuts. Experience a seamless Cloud Storage interaction right from your terminal; no more swapping to a browser. Stimulate your productivity with Pgcs.
- pinecone-datasets — Pinecone Datasets lets you easily load datasets into your Pinecone index.
- pins — Publish data sets, models, and other python objects, making it easy to share them across projects and with your colleagues.
- polyaxon — Command Line Interface (CLI) and client to interact with Polyaxon API.
- polyaxon-deploy — Polyaxon deployment and serving tools: streams, sandbox, ML-API, and spaces.
- pv-site-prediction — Photovoltaic per site modeling
- PVNet — PVNet
- pycarol — Carol Python API and Tools
- pycontrails — Python library for modeling aviation climate impacts
- pydantic-cereal — Advanced serialization for Pydantic models
- pydatafabric — SHINSEGAE DataFabric Python Package
- pyiceberg — Apache Iceberg is an open table format for huge analytic datasets
- pyinsta-functions — no summary
- pyplatform-common — Pyplatform-common package provides utility, file management and authentication functions for interacting with APIs and compute services.
- pyplatform-datalake — Pyplatform-datalake package provides functions for Google Cloud Storage and Microsoft Storage services.
- pyplatform-reporting — Pyplatform-reporting package provides function for managing hyper datasources on Tableau server.
- pyreft — REFT: Representation Finetuning for Language Models
- pytest-servers — pytest servers
- python-oauth-token-manager — API for managing stored OAuth credentials.
- python-wrap-gcp — Python helpers to interact with GCP
- pywren-ibm-cloud — Run many jobs over IBM Cloud
- qbeast-sharing — Python Connector for Delta Sharing
- quetz-server — The mamba-org server for conda packages
- rikai — no summary
- rstojnic-tfds-nightly — tensorflow/datasets is a library of datasets ready to use with TensorFlow.
- s3contents — S3 Contents Manager for Jupyter
- schemadiffed — Compare Parquet file schemas across different filesystems
- scivision — Scivision
- segy — The Ultimate Python SEG-Y I/O with Cloud Support and Schemas
- semantic-navigator — An active learning approach to query and search through large archival datasets.
- shapelets-platform — Data Scientist platform
- spice-rack — group of common things we use across different python packages
- spPersist — Spatial transcriptomics with Persistent Homology
- sq-blocks — Blocks provides a simple interface to read, organize, and manipulate structured data in files on local and cloud storage
- squirrel-core — Squirrel is a Python library that enables ML teams to share, load, and transform data in a collaborative, flexible and efficient way.
- ssb-ipython-kernels — Jupyter kernels for working with dapla services
- tap-dbf — Singer tap for .DBF files
- tensorflow-datasets — tensorflow/datasets is a library of datasets ready to use with TensorFlow.
- terrabridge — terrabridge connects your terraform managed resources to your application.
- textual-universal-directorytree — A Textual Directory Tree for all File Systems
- tfds-nightly — tensorflow/datasets is a library of datasets ready to use with TensorFlow.
- the-refinery-toolkit — The Refinery Toolkit (RFTK)
- thirdai — A faster cpu machine learning library
- transparentpath — A class that allows one to use a path in a local file system or a gcs file system (more or less) in almost the same way one would use a pathlib.Path object.
- tux — Tools and Utils. Some tools and utils modified from many other code to fit my needs.
- ufs2arco — Tools for converting Unified Forecast System (UFS) output to Analysis Ready, Cloud Optimized (ARCO) format
- universal_pathlib — pathlib api extended to use fsspec backends