Reverse Dependencies of librosa
The following projects have a declared dependency on librosa:
- GAMuT — Granular Audio Musaicing Toolkit for Python
- genaibook — Utilities for 'Hands-On Generative AI with Transformers and Diffusion Models' (upcoming)
- geniusrise-audio — audio bolts for geniusrise
- gft — GFT (general fine-tuning) A Little Language for Deepnets: 1-line programs for fine-tuning, inference and more
- gft-cpu — GFT (general fine-tuning) A Little Language for Deepnets: 1-line programs for fine-tuning, inference and more
- gnes — GNES is Generic Neural Elastic Search, a cloud-native semantic search system based on deep neural network.
- gpt-sovits-python — Python wrapper for fast inference with GPT-SoVITS
- graphite-datasets — tensorflow/datasets is a library of datasets ready to use with TensorFlow.
- grinding-lib — Demo Python Library for Siemens Grinding project
- guitarsounds — A python package to analyze and visualize harmonic sounds
- h2ogpt — no summary
- hallooworldgk — A basic hello package
- hay-say-common — Constants and methods that are shared between the Hay Say UI and various Docker containers it communicates with.
- hear-savi — An HEAR API for SAVI AudioCNN
- hearbaseline — Holistic Evaluation of Audio Representations (HEAR) 2021 -- Baseline Model
- hezar — Hezar: The all-in-one AI library for Persian, supporting a wide variety of tasks and modalities!
- HMBasr — Automatic Speech Recognition
- hmc-mir — Collection of tools developed by HMCs MIR Lab
- horoscopy — Python module for speech signal processing
- howl — A wake word detection toolkit
- huggingsound — HuggingSound: A toolkit for speech-related tasks based on HuggingFace's tools.
- humanlikehearing — Psychometric testing on Automatic Speech Recognition systems
- hyperion-ml — Toolkit for speaker recognition
- iatorch — PyTorch Wrapper for Inspection AI
- iautils — Image & Audio Common Utils
- infer-rvc-python — Python wrapper for fast inference with rvc
- instruwav — Generate sounds using a base note
- insynth — Domain-specific generation of test inputs for robustness testing of ML models
- iracema — Audio Content Analysis for Research on Musical Expressiveness and Individuality
- iSparrowRecord — Audio Recording Facilities for the isparrow package
- Jabberjay — 🦜 Synthetic Voice Detection
- jac-speech — no summary
- jack-audio — A Python package for stationary audio noise reduction.
- jackAudio — A Python package for stationary audio noise reduction.
- jarvis-akul2010 — A library built to make it extremely easy to build a simple voice assistant.
- jaseci-ai-kit — no summary
- Jems-Video — Create 🔥 videos with Stable Diffusion by exploring the latent space and morphing between text prompts.
- jina — Jina (v%s) is a cloud-native semantic search engine powered by deep neural networks.It provides a universal solution of large-scale index and query for media contents.
- jotts — JoTTS is a German text-to-speech engine.
- jrvc — Libraries for RVC inference
- jump-reward-inference — A package for fast real-time music joint rhythmic parameters tracking including beats, downbeats, tempo and meter using the BeatNet AI, a super compact 1D state space and the jump back reward technique
- kapre — Kapre: Keras Audio Preprocessors. Keras layers for audio pre-processing in deep learning
- ketos — MERIDIAN Python package for deep-learning based acoustic detector and classifiers
- klay-beam — Toolkit for massively parallel audio processing via Apache Beam
- klio-audio — Library for audio-related Klio transforms and helpers
- konnyaku-gpt — AI-powered multimodal subtitle generator.
- kudio — Audio Toolbox™ KUDIO
- laion-clap — Contrastive Language-Audio Pretraining Model from LAION
- langchain_1111_Dev_cerebrum — Building applications with LLMs through composability
- langchain-by-johnsnowlabs — Building applications with LLMs through composability
- langchain-xfyun — 在LangChain中流畅地使用讯飞星火大模型
- langchaincoexpert — Building applications with LLMs through composability
- langchainn — Building applications with LLMs through composability
- lcp-video — LCP video analysis
- lhvqt — Frontend filterbank learning module with HVQT initialization capabilities
- libf0 — A Python Library for Fundamental Frequency Estimation in Music Recordings
- libfmp — Python module for fundamentals of music processing
- libmv — a library to create music videos
- libquantum — Library for implementing standardized time-frequency representations.
- libsoni — A Python Toolbox for Sonifying Music Annotations and Feature Representations
- libtsm — Python Package for Time-Scale Modification and Pitch-Shifting
- lightning-flash — Your PyTorch AI Factory - Flash enables you to easily configure and run complex AI recipes.
- lightwood — Lightwood is Legos for Machine Learning.
- llmchatbot — LLM-based Chatbot
- LPCTorch — LPC Utility for Pytorch Library.
- lungdata — no summary
- lvc — Unofficial pip package for zero-shot voice conversion
- MAAP — no summary
- macls — Audio Classification toolkit on Pytorch
- maestro-music — A simple command line tool to play songs (or any audio files, really).
- mafe — Music Audio Feature Extractor
- magenta — Use machine learning to create art and music
- magenta-gpu — Use machine learning to create art and music
- malaya-speech — Speech-Toolkit for bahasa Malaysia, powered by Tensorflow and PyTorch.
- malayalam-asr-benchmarking — A study to benchmark whisper based ASRs in Malayalam
- masked_prosody_model — no summary
- matcha-tts — 🍵 Matcha-TTS: A fast TTS architecture with conditional flow matching
- mayavoz — Deep learning toolkit for speech enhancement
- measure-spkr — An app for measuring impulse-frequency response of speaker
- mel-cepstral-distance — CLI and library to compute the Mel-Cepstral Distance of two WAV files based on the paper 'Mel-Cepstral Distance Measure for Objective Speech Quality Assessment' by Robert F. Kubichek.
- mexca — Emotion expression capture from multiple modalities.
- microfaune-ai — Module package used for the Microfaune project
- mimikit — Python package for generating audio with neural networks
- mindtorch — MindTorch is a toolkit for support the PyTorch model running on Ascend.
- minimalml — A python package for out-of-the-box ML solutions
- mir-bootleg-score — Built for MIR Lab. Tools for converting png images into bootleg score features.
- mirdata — Common loaders for MIR datasets.
- mixsim — An open-source dataset for multiple purposes, such as speaker localization/tracking, dereverberation, enhancement, separation, and recognition.
- ml4a — A toolkit for making art with machine learning, including an API for popular deep learning models, recipes for combining them, and a suite of educational examples
- mlcroissant — MLCommons datasets format.
- MLProto — Modular Neural Network Protyping for Stock Market Prediction
- mmaction2 — OpenMMLab Video Understanding Toolbox and Benchmark
- mmkit-features — A multimodal architecture to build multimodal knowledge graphs with flexible multimodal feature extraction and dynamic multimodal concept generation.
- MMSA-FET — A Tool for extracting multimodal features from videos.
- mockingbirdforuse — no summary
- mockingbirdonlyforuse — A Raw Use Version for MockingBird
- model-creator-bird-sing-v2 — Autoencoder singing
- model-evaluation-777 — Evaluate speech enhancemnt model performance
- modelscope — ModelScope: bring the notion of Model-as-a-Service to life.
- Modulation — time series signal analysis