Reverse Dependencies of librosa
The following projects have a declared dependency on librosa:
- Montreal-Forced-Aligner — Montreal Forced Aligner is a package for aligning speech corpora using Kaldi functionality.
- movie-utils — 视频文件处理的辅助工具
- movis — A video editing library
- mpai-cae-arp — The MPAI CAE-ARP software API
- mpk — Media Programming Kit Python bindings
- mplc — A distributed-learning package for the study of multi-partner learning approaches and contributivity measurement methods
- msa-toolbox — MSA Toolbox
- msaf — Python module to discover the structure of music files
- msaf-test — Python module to discover the structure of music files
- msclap — CLAP (Contrastive Language-Audio Pretraining) is a model that learns acoustic concepts from natural language supervision and enables “Zero-Shot” inference. The model has been extensively evaluated in 26 audio downstream tasks achieving SoTA in several of them including classification, retrieval, and captioning.
- mser — Speech Emotion Recognition toolkit on Pytorch
- mtsa — Multiple Time Series Analysis
- muko — 加速实现AIGC、自动办公的中文编程工具
- muprocdurham — Python package for the Music Processing lecture and practicals at Durham University
- MusFeat — no summary
- mushan — Personal toolkit.
- music_util — Collection of utilities for musicians
- music22 — A tool for musicological analysis from audio files. Now it is focused on modal music analysis : Scale analysis, tonic detection
- musicflower — no summary
- MusicInsights — no summary
- musicnet — no summary
- musicnn-keras — Pronounced as "musician", musicnn is a set of pre-trained deep convolutional neural networks for music audio tagging. Musicnn_keras is a tf.keras implementation of musicnn
- musicntwrk — music as data, data as music
- muvimaker — A small package to generate moving pictures from sound
- mvector — Voice Print Recognition toolkit on Pytorch
- mvits — VITS toolkit on Pytorch
- mvmake — A script to cut a directory of clips to music.
- mw-adapter-transformers — A friendly fork of HuggingFace's Transformers, adding Adapters to PyTorch language models
- nemo-asr — Collection of Neural Modules for Speech Recognition
- nemo-toolkit — NeMo - a toolkit for Conversational AI
- nemo-tts — Collection of Neural Modules for Speech Synthesis
- nendo — The Nendo AI Audio Tool Suite
- neon-iris — Interactive Relay for Intelligence Systems
- neon-utils — Utilities for NeonAI
- neural-homomorphic-vocoder — Pytorch implementation of neural homomorphic vocoder
- nijtaio — Helper module to streamline access to Nijta's API
- nisqa — NISQA - Non-Intrusive Speech Quality and TTS Naturalness Assessment
- nnAudio — A fast GPU audio processing toolbox with 1D convolutional neural network
- noisereduce — Noise reduction using Spectral Gating in Python
- nonebot-plugin-diffsinger — 用DiffSinger让bot唱歌
- note-seq — Use machine learning to create art and music
- npc-engine — Deep learning inference and NLP toolkit for game development.
- nusacrowd — no summary
- nussl — A flexible sound source separation library.
- nwave — Multithread batch resampling and waveform transforms
- oceanai — OCEAN-AI
- oceansoundscape — A python package for analyzing ocean acoustic data.
- omnizart — Omniscient Mozart, being able to transcribe everything in the music.
- openav — OpenAV
- opensoundscape — Open source, scalable acoustic classification for ecology and conservation
- openvoice-cli — Use OpenVoice 2 stage via console or python scripts
- oplangchain — langchain for OpenPlugin
- optimum-furiosa — Optimum Furiosa is the interface between the 🤗 Transformers library and Furiosa NPUs such as Furiosa Warboy. It provides a set of tools enabling easy model loading and inference for different downstream tasks for Furiosa NPU.
- optimum-graphcore — Optimum Library is an extension of the Hugging Face Transformers library, providing a framework to integrate third-party libraries from Hardware Partners and interface with their specific functionality.
- optimum-nvidia — Optimum Nvidia is the interface between the Hugging Face Transformers and NVIDIA GPUs. "
- otxmmaction2 — OpenMMLab Video Understanding Toolbox and Benchmark
- paddle-parakeet — Speech synthesis tools and models based on Paddlepaddle
- paddleaudio — Speech audio tools based on Paddlepaddle
- paddlelibrosa — Paddle implemention of part of librosa functions.
- paddlespeech — Speech tools and models based on Paddlepaddle
- paderbox — Collection of utilities in the department of communications engineering of the UPB
- panns-AT-inference — panns_AT_inference: audio tagging inference toolbox
- panns-inference — panns_inference: audio tagging and sound event detection inference toolbox
- paule — paule implements the Predictive Articulatory speech synthesis model Utilizing Lexical Embeddings (PAULE), which is a control model for the articulatory speech synthesizer VocalTractLab (VTL).
- penn — Pitch Estimating Neural Networks (PENN)
- piano-transcription-inference — Piano transcription inference toolbox
- pianoputer — Use your computer keyboard as a "piano"
- pipepal — PipePal is a Python package that simplifies building pipelines for speech and voice analysis.
- pitch-detectors — collection of pitch detection algorithms with unified interface
- pitchsqueezer — Robust pitch tracker for speech, using synchrosqueezing and spectral autocorrelation
- pliers — Multimodal feature extraction in Python
- plixkws — Plug-and-Play Multilingual Few-shot Spoken Words Recognition
- plot-wav — no summary
- Plotting-funcs — Auxiliary functions for plotting purposes mainly.
- polyglotdb — no summary
- ppacls — Audio Classification toolkit on PaddlePaddle
- ppgan — Awesome GAN toolkits based on PaddlePaddle
- ppgs — Phonetic posteriorgrams
- ppser — Speech Emotion Recognition toolkit on PaddlePaddle
- ppvector — Voice Print Recognition toolkit on PaddlePaddle
- ppvits — VITS toolkit on PaddlePaddle
- praudio — Complex preprocessing of entire audio datasets with 1 command
- precountify — A tool for pre-countifying
- prowav — The package for preprocessing wave data
- psychopy-whisper — Extension for transcription using OpenAI Whisper.
- pumpp — A practically universal music pre-processor
- py-data-juicer — A One-Stop Data Processing System for Large Language Models.
- pyabelab — Library that are likely to be used frequently in ABELAB.
- Pyara — Library for audio classification
- pyaudioaugment — no summary
- pyaudioclassification — Dead simple audio classification
- pyAudioKits — Powerful Python audio workflow support based on librosa and other libraries
- pyclarity — Tools for the Clarity Challenge
- pyfinch — A python package for analyzing neural & bioacoustics signals from songbirds
- pyfoal — Python forced aligner
- PyHa-test — A python package for automatically detecting species and comparing to ground truth
- pylights — Module used to change the color and brightness of lights to the beat of an udio file
- pymcd — Calculate Mel-Cepstral Distortion (MCD)
- pymixing — A simple daw in python.
- pymusickit — A Python package for music analysis. Keyfinder Forked from "https://github.com/jackmcarthur/musical-key-finder"