Reverse Dependencies of rapidfuzz
The following projects have a declared dependency on rapidfuzz:
- a-pandas-ex-fuzz — Intuitive way of using fuzz matching in pandas
- a-pandas-ex-fuzzymerge — Merges two DataFrames using fuzzy matching on specified columns
- a-pandas-ex-tesseract-multirow-regex-fuzz — Regex/Fuzz search across multiple rows/Tesseract to pandas.DataFrame
- aa-drifters — Drifter wormhole tracker/manager plugin for Alliance Auth.
- adbkit — Big automation package for ADB
- addrmatcher — Australian Address Matcher to Regions
- agentlib — Framework for development and execution of agents for control and simulation of energy systems.
- agentverse — A versatile framework that streamlines the process of creating custom multi-agent environments for large language models (LLMs).
- aimai-search — no summary
- airclick — airclick 相关python包
- algorin-cli — Acceso a GPT-3 y procesamiento de documentos desde la línea de comandos.
- allusgov — This project attempts to map the organization of the US Federal Government by gathering and consolidating information from various directories.
- aniparser — Provides a parser to parse file names and return data based on the names
- ansible-risk-insight — My package description
- aptos-verify — A small example package
- ArchivesSnake — ArchivesSpace Client Library
- ascript — airclick 相关python包
- Assistant-Ostap — Your personal assistant Ostap
- auto-learn-gpt — autoML for training and inference Deep Learning model
- avwx-engine — Aviation weather report parsing library
- babygruut — A tokenizer, text cleaner, and phonemizer for many human languages.
- bblib — beambusters library to refine the detector center for crystallography data processing.
- bdcctools — Biodiversity Data Cleaning and Curation Tools
- bent — BENT: Biomedical Entity Annotator
- bib-dedupe — Identify and merge duplicates in bibliographic records
- bigeye-sdk — Bigeye SDK offers developer tools and clients to interact with Bigeye programmatically.
- bionty-base — Bionty.
- blendsql — Orchestrate SQLite logic and LLM reasoning within a unified dialect.
- boaviztapi — An API to access Boavizta's methodologies and footprint reference data
- brewt — a brewing tool for CivMC/CivCraft
- calendruparser — Parser for a list of official and unofficial holidays from calend.ru
- carsomenlp — Carsome NLP package
- casparser-isin — ISIN database for casparser
- cdp-backend — Data storage utilities and processing pipelines to run on CDP server deployments.
- cdptools — Tools to interact with and deploy CouncilDataProject instances
- chemex — An analysis program for chemical exchange detected by NMR
- cir-duplicate-detector — PDQ hash and URL duplicate detector. Developed by Sam Sweere from BigData Repulic as part of their Social Good Initiaive.
- citoplasm — CITOplasm is a Python library for writing LLM code in a declarative way.
- cleo — Cleo allows you to create beautiful and testable command-line interfaces.
- cloudknot — Cloudknot: a python library designed to run your existing python code on AWS Batch
- cltk — The Classical Language Toolkit
- codecarbon — no summary
- colrev — CoLRev: An open-source environment for collaborative reviews
- comicapi — Comic archive (cbr/cbz/cbt) and metadata utilities. Extracted from the comictagger project.
- comics-ocr — ComicsOCR is a Python package created for easily distributing OCR models trained for golden age of comics.
- comictagger — A cross-platform GUI/CLI app for writing metadata to comic archives
- compiloor — no summary
- countASAP — A software for converting ASAPseq FASTQs to count matrices
- CountESS — CountESS
- cron-times — Timetable for your cronjobs
- cso-classifier — A light-weight Python app for classifying scientific documents with the topics from the Computer Science Ontology (https://cso.kmi.open.ac.uk/home).
- dagster — Dagster is an orchestration platform for the development, production, and observation of data assets.
- DAJIN2 — One-step genotyping tools for targeted long-read sequencing
- data-gradients — DataGradients
- data-harmonization-ai-dp — Create data quality rules and apply them to datasets.
- dataherb — Get clean datasets from DataHerb to boost your data science and data analysis projects
- DataProfiler — What is in your data? Detect schema, statistics and entities in almost any file.
- dbgpt-hub — DB-GPT-Hub: Text-to-SQL parsing with LLMs
- deduce — Deduce: de-identification method for Dutch medical text
- dedupe-FuzzyWuzzy — Deduplication using RapidFuzz library.
- desktop-env — The package provides a desktop environment for setting and evaluating desktop automation tasks.
- DevRewind — no summary
- dicomselect — no summary
- digital-eval — Evaluate Mass Digitalization Data
- dinglehopper — The OCR evaluation tool
- distill-trainer — Knowledge distillation toolkit
- docprompt — Documents and large language models.
- dp-ai-data-harmonization — Create data quality rules and apply them to datasets.
- dracula-cli — A beautiful command line interface for draculatheme.org and dracula in general
- dt4dds — dt4dds is a Python package providing a customizable, digital representation of the widely-used DNA data storage workflow involving array synthesis, PCR, Aging, and Sequencing-By-Synthesis. By modelling each part of such user-defined workflows with fully customizable experimental parameters, dt4dds enables data-driven experimental design and rational design of redundancy. dt4dds also includes a pipeline for comprehensively analyzing errors in sequencing data, both from experiments and simulation.
- dvha — Create a database of DVHs, GUI with wxPython, plots with Bokeh
- e2xgradingtools — A package for creating autograder tests in Jupyter notebooks
- ecowitt2mqtt — A small web server to send data from Ecowitt devices to an MQTT Broker
- edgar — Scrape data from SEC's EDGAR
- edu-segmentation — To improve EDU segmentation performance using Segbot. As Segbot has an encoder-decoder model architecture, we can replace bidirectional GRU encoder with generative pretraining models such as BART and T5. Evaluate the new model using the RST dataset by using few-shot based settings (e.g. 100 examples) to train the model, instead of using the full dataset.
- edwh — Education Warehouse maintenance tools
- eggcell — Add a short description here!
- emm — Entity Matching Model package
- ENGR131-2024 — ENGR131_2024 package
- EntityNormalizer — Library for normalizing entities based on a dictionary
- EorzeaEnv — Final Fantasy XIV weather & time tools.
- errant — The ERRor ANnotation Toolkit (ERRANT). Automatically extract and classify edits in parallel sentences.
- errant-prep — The ERRor ANnotation Toolkit (ERRANT). Automatically extract and classify edits in parallel sentences.
- eso-names — no summary
- fadoudou2 — Awesome OCR toolkits based on PaddlePaddle (8.6M ultra-lightweight pre-trained model, support training and deployment among server, mobile, embeded and IoT devices)
- farm-haystack — LLM framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data.
- fifacodes — FIFA member associations codes query and search.
- filip — [FI]WARE [Li]brary for [P]ython
- fotools — Tool for searching and organizing files
- french-cities — Toolbox on french cities: set vintage, find departments, find cities...
- fusion-stat — Scrape football data from multiple sources simultaneously.
- FuzzTypes — FuzzTypes is a Pydantic extension for annotating autocorrecting fields
- FuzzUp — A Fuzzy Matching Approach for Clustering Strings
- fuzzy-types — rapdifuzz-based utilities and data structures
- fuzzyfiles — Binary fuzzy matching in all file types [fzf (pre-filter)/rapidfuzz (finds the best result)]
- fuzzymatcher — Fuzzy match two pandas dataframes based on one or more common fields
- fuzzypandaswuzzy — Fuzzy Comparison Utilities for DataFrame Columns
- fuzzyset2 — A simple python fuzzyset implementation.
- fuzzysyn — no summary
- fuzzyydictyy — Dict with fuzzy key matching