Reverse Dependencies of librosa
The following projects have a declared dependency on librosa:
- testgailbot002 — GailBot API
- testgailbotapi — GailBot Test API
- testgailbotapi001 — GailBot Test API
- Tetra-Model-Zoo — Models optimized for export to run on device.
- tfds-nightly — tensorflow/datasets is a library of datasets ready to use with TensorFlow.
- tfds-nightly-gradient — tensorflow/datasets is a library of datasets ready to use with TensorFlow.
- tflibrosa — Re-implementation of some librosa function for tensorflow. Reproduction from torchlibrosa.
- tflite-model-maker — TFLite Model Maker: a model customization library for on-device applications.
- tflite-model-maker-nightly — TFLite Model Maker: a model customization library for on-device applications.
- tglcourse — work-in-progress course
- thefiarlib — thefiarlib
- tifex-py — TODO
- tifresi — Time Frequency Spectrogram Inversion
- tinygrad — You like pytorch? You like micrograd? You love tinygrad! <3
- tinygrad-tools — You like pytorch? You like micrograd? You love tinygrad! <3
- tmh — TMH Speech package
- tokensynth — tokensynth
- tomtom — Funções para tese de doutorado. Em breve mais.
- tonic — Neuromorphic datasets and transformations.
- torch-ecg — A Deep Learning Framework for ECG Processing Tasks Based on PyTorch
- torch-mfcc — A librosa's STFT/FBANK/MFCC implement based on Torch
- torch-stft — An STFT/iSTFT for PyTorch
- torchaudio-augmentations — Audio augmentations library for PyTorch, for audio in the time-domain.
- torchlibrosa — PyTorch implemention of part of librosa functions.
- torchmetrics — PyTorch native Metrics
- torchopenl3 — Deep audio and image embeddings, based on Look, Listen, and Learn approach Pytorch
- torchsynth — A modular synthesizer in pytorch, GPU-optional and differentiable
- transcription-diff — Speech to transcription comparison
- transformers — State-of-the-art Machine Learning for JAX, PyTorch and TensorFlow
- tsaug — A package for time series augmentation
- tts — Deep learning for Text to Speech by Coqui.
- tts-uk — High-fidelity speech synthesis for Ukrainian using modern neural networks.
- tts-with-rvc — TTS with RVC pipeline
- tts-with-rvc-onnx — TTS with RVC pipeline (ONNX Version)
- TTS2 — Deep learning for Text to Speech by Coqui.
- ttscube — Text-to-speech synthesis engine, based on end-to-end GAN training. Features: multilanguage, multispeaker, realtime CPU synthesis
- ttskit — text to speech toolkit
- ttsmms — Text-to-speech with The Massively Multilingual Speech (MMS) project
- ttspersian — A Python library for Persian text-to-speech using Microsoft Azure service.
- tuning-fork — A clip/sample auto tuner
- tvsm-extractor — Extract TVSM dataset
- twang — Machine learning tools for guitarists
- twilio-phone-calls — Python Implementation for Handling Twilio Phone Calls
- udls — Base class and presets for fast dataset creation inside IRCAM
- ultimate-rvc — Ultimate RVC
- ultravox-vllm — no summary
- univoc — A PyTorch implementation of Towards Achieving Robust Universal Neural Vocoding.
- useful-moonshine — Speech Recognition for Live Transcription and Voice Commands
- uss — Universal source separation (USS) with weakly labelled data.
- vanpy — VANPY - Voice Analysis framework in Python
- vectorhub — One liner to encode data into vectors with state-of-the-art models using tensorflow, pytorch and other open source libraries. Word2Vec, Image2Vec, BERT, etc
- vectorhub-nightly — One liner to encode data into vectors with state-of-the-art models using tensorflow, pytorch and other open source libraries. Word2Vec, Image2Vec, BERT, etc
- versatile-audio-upscaler — Versatile AI-driven audio upscaler to enhance the quality of any audio.
- versatile-audio-upscaler-fixed — Versatile AI-driven audio upscaler to enhance the quality of any audio -- Now supporting Numpy +1.26
- vibe-analyser — A vibration analysis and data acquisition suite for the rpi
- vid2aud — A python module to extract audio from a video
- vid2cleantxt — A command-line tool to easily transcribe speech-based video files into clean text. also in Colab.
- vidtoolz-beats — Get beats from a mp3 song
- viet-tts — VietTTS: An Open-Source Vietnamese Text to Speech
- ViewCube — Datacube visualization and sonification
- visbeat — Code for 'Visual Rhythm and Beat' SIGGRAPH 2018
- visbeat3 — Python3 Implementation for 'Visual Rhythm and Beat' SIGGRAPH 2018
- vision-unlearning — Vision Unlearning: a tool for Machine Unlearning in Computer Vision
- vital-sqi — Signal quality control pipeline for electrocardiogram and photoplethysmogram
- vllm-emissary — A high-throughput and memory-efficient inference and serving engine for LLMs
- vllm-xft — A high-throughput and memory-efficient inference and serving engine for LLMs
- vocal — A vocal source separation
- vocal-remover — Vocal remover.
- vocalpy — A core package for acoustic communication research in Python
- vocex — Voice Frame-Level and Utterance-Level Attribute Extraction
- Voice-Cloning — Introducing Voice_Cloning: A Python Package for Speech Synthesis and Voice Cloning!
- voice-gender-recognition — A voice gender recognition package using machine learning
- voice-toolbox — Convenient wrappers for audio signal processing in Python
- voice100-runtime — Voice100 Runtime is a TTS/ASR sample app that uses ONNX Runtime, WORLD and Voice100 neural TTS/ASR models on Python. Inference of Voice100 is low cost as its models are tiny and only depend on CNN without recursion.
- voiceauth — AI Voice Detection System
- voiceauthCore — A deepfake audio detection tool
- voicefixer — This package is written for the restoration of degraded speech
- Voicelab — Fully Automated Reproducible Acoustical Analysis
- voicerestore-fork — This is a an unofficial fork of the https://github.com/skirdey/voicerestore repository for pip packaging.
- voices — 适用于 diffsinger 的多功能工具集
- vox-box — Vox box
- voxws — Few Shot Language Agnostic Keyword Spotting (FSLAKWS) System
- VuVoPy — Voice features
- waddle-ai — A podcast preprocessing library for aligning, normalizing, and transcribing audio files.
- wandas — Wandas is an open source library for efficient signal analysis in Python
- warpq — WARP-Q: Quality Prediction For Generative Neural Speech Codecs
- wav-autoencoder — WavAutoencoder: A Self-Supervised Framework for Learning Audio Representations
- wav2clip — Wav2CLIP: Learning Robust Audio Representations From CLIP.
- Wav2lip-integration — no summary
- Wav2Lipy — Wrapper Package for LipGan Project
- Wav2TextGrid — A python forced alignment package
- waveglow — Waveglow library
- waveglow-cli — Command-line interface (CLI) to train WaveGlow using .wav files.
- wavescapes — Python library to build wavescapes, plots used in musicology.
- waveser — Used to process audio data.
- wavesongs — A python package for birdsongs creation and data extraction.
- wavetabler — Wavetabler: A tool for generating wavetables from audio files
- waveuse — Used to process audio data.
- wavmark — AI-Based Audio Watermarking Tool
- wavx — 音频分析和处理工具库