Reverse Dependencies of librosa
The following projects have a declared dependency on librosa:
- ml4a — A toolkit for making art with machine learning, including an API for popular deep learning models, recipes for combining them, and a suite of educational examples
- mlcroissant — MLCommons datasets format.
- MLProto — Modular Neural Network Protyping for Stock Market Prediction
- MLTask-utils — a collection of commonly used tools by MLTask
- mmaction2 — OpenMMLab Video Understanding Toolbox and Benchmark
- mmkit-features — A multimodal architecture to build multimodal knowledge graphs with flexible multimodal feature extraction and dynamic multimodal concept generation.
- MMSA-FET — A Tool for extracting multimodal features from videos.
- mockingbirdforuse — no summary
- mockingbirdonlyforuse — A Raw Use Version for MockingBird
- model-creator-bird-sing-v2 — Autoencoder singing
- model-evaluation-777 — Evaluate speech enhancemnt model performance
- modelscope — ModelScope: bring the notion of Model-as-a-Service to life.
- Modulation — time series signal analysis
- Montreal-Forced-Aligner — Montreal Forced Aligner is a package for aligning speech corpora using Kaldi functionality.
- movie-utils — 视频文件处理的辅助工具
- movis — A video editing library
- mpai-cae-arp — The MPAI CAE-ARP software API
- mpk — Media Programming Kit Python bindings
- mplc — A distributed-learning package for the study of multi-partner learning approaches and contributivity measurement methods
- MPSENet — Python package of MP-SENet from Explicit Estimation of Magnitude and Phase Spectra in Parallel for High-Quality Speech Enhancement.
- ms-funcodec — FunCodec: A Fundamental, Reproducible and Integrable Open-source Toolkit for Neural Speech Codec
- msa-toolbox — MSA Toolbox
- msaf — Python module to discover the structure of music files
- msaf-test — Python module to discover the structure of music files
- msclap — CLAP (Contrastive Language-Audio Pretraining) is a model that learns acoustic concepts from natural language supervision and enables “Zero-Shot” inference. The model has been extensively evaluated in 26 audio downstream tasks achieving SoTA in several of them including classification, retrieval, and captioning.
- mseep-mcp-music-analysis — \A MCP server with music tools\
- mseep-nash-mcp — Nash MCP
- mser — Speech Emotion Recognition toolkit on Pytorch
- mtsa — Multiple Time Series Analysis
- muko — 加速实现AIGC、自动办公的中文编程工具
- multimodel-ai — A Python module for efficient multi-model AI inference with memory management
- muprocdurham — Python package for the Music Processing lecture and practicals at Durham University
- MusFeat — no summary
- mushan — Personal toolkit.
- music_util — Collection of utilities for musicians
- music22 — A tool for musicological analysis from audio files. Now it is focused on modal music analysis : Scale analysis, tonic detection
- musicflower — no summary
- MusicInsights — no summary
- musicnet — no summary
- musicnn-keras — Pronounced as "musician", musicnn is a set of pre-trained deep convolutional neural networks for music audio tagging. Musicnn_keras is a tf.keras implementation of musicnn
- musicnotefinder — A library for analyzing audio files and detecting musical notes.
- musicntwrk — music as data, data as music
- muvimaker — A small package to generate moving pictures from sound
- mvits — VITS toolkit on Pytorch
- mvmake — A script to cut a directory of clips to music.
- mw-adapter-transformers — A friendly fork of HuggingFace's Transformers, adding Adapters to PyTorch language models
- nemo-asr — Collection of Neural Modules for Speech Recognition
- nemo-toolkit — NeMo - a toolkit for Conversational AI
- nemo-tts — Collection of Neural Modules for Speech Synthesis
- nendo — The Nendo AI Audio Tool Suite
- neon-iris — Interactive Relay for Intelligence Systems
- neon-utils — Utilities for NeonAI
- neural-homomorphic-vocoder — Pytorch implementation of neural homomorphic vocoder
- neverlib — A successful sign for python setup
- new-fave — New Vowel Extraction Suite
- nexaai-metal — Nexa AI SDK
- nexai — Nexa AI SDK
- nijtaio — Helper module to streamline access to Nijta's API
- nisqa — NISQA - Non-Intrusive Speech Quality and TTS Naturalness Assessment
- nnAudio — A fast GPU audio processing toolbox with 1D convolutional neural network
- nonebot-plugin-checkbpm — 基于Librosa的音频文件BPM计算器 通过上传到群文件方式计算音频文件的bpm值(beat per minute)
- nonebot-plugin-diffsinger — 用DiffSinger让bot唱歌
- note-seq — Use machine learning to create art and music
- npc-engine — Deep learning inference and NLP toolkit for game development.
- nusacrowd — no summary
- nussl — A flexible sound source separation library.
- NViXTTS — Deep learning for Vietnamese Text to Speech
- nwave — Multithread batch resampling and waveform transforms
- o2-speechless — Speechless repo for sales call analysis
- oarc — OARC Python Package
- oceanai — OCEAN-AI
- oceansoundscape — A python package for analyzing ocean acoustic data.
- odysee — High-performance quantum-inspired multimodal memory system with adaptive routing and distributed processing capabilities
- ofautils — AIO Triton Utilities
- omnizart — Omniscient Mozart, being able to transcribe everything in the music.
- openav — OpenAV
- openmmla — OpenMMLA: A unified multimedia platform integrating audio and vision modules for analysis and computer aided instruction.
- openmmla-audio — Audio module for the OpenMMLA platform
- openscenesense — A video analysis toolkit using OpenAI and Openrouter vision models
- openscenesense-ollama — Offline video analysis using Ollama models and local Whisper
- opensoundscape — Open source, scalable acoustic classification for ecology and conservation
- openvoice-cli — Use OpenVoice 2 stage via console or python scripts
- openwillis-voice — digital health measurement
- oplangchain — langchain for OpenPlugin
- optimum-furiosa — Optimum Furiosa is the interface between the 🤗 Transformers library and Furiosa NPUs such as Furiosa Warboy. It provides a set of tools enabling easy model loading and inference for different downstream tasks for Furiosa NPU.
- optimum-graphcore — Optimum Library is an extension of the Hugging Face Transformers library, providing a framework to integrate third-party libraries from Hardware Partners and interface with their specific functionality.
- optimum-neuron — Optimum Neuron is the interface between the Hugging Face Transformers and Diffusers libraries and AWS Trainium and Inferentia accelerators. It provides a set of tools enabling easy model loading, training and inference on single and multiple neuron core settings for different downstream tasks.
- osagent — osagent
- otozu — A library for converting between audio files and spectrograms
- otxmmaction2 — OpenMMLab Video Understanding Toolbox and Benchmark
- paddle-parakeet — Speech synthesis tools and models based on Paddlepaddle
- paddleaudio — Speech audio tools based on Paddlepaddle
- paddlelibrosa — Paddle implemention of part of librosa functions.
- paddlespeech — Speech tools and models based on Paddlepaddle
- paderbox — Collection of utilities in the department of communications engineering of the UPB
- pafst — Library That Preprocessing Audio For TTS/STT.
- pafts — Library That Preprocessing Audio For TTS.
- panns-AT-inference — panns_AT_inference: audio tagging inference toolbox
- panns-inference — panns_inference: audio tagging and sound event detection inference toolbox
- paule — paule implements the Predictive Articulatory speech synthesis model Utilizing Lexical Embeddings (PAULE), which is a control model for the articulatory speech synthesizer VocalTractLab (VTL).