#typescript #ai #audio #diffusion #music #nextjs #stable_diffusion #threejs
https://github.com/hmartiro/riffusion-app
https://github.com/hmartiro/riffusion-app
GitHub
GitHub - riffusion/riffusion-app-hobby: Stable diffusion for real-time music generation (web app)
Stable diffusion for real-time music generation (web app) - riffusion/riffusion-app-hobby
#python #ai #audio #diffusers #diffusion #music #stable_diffusion
https://github.com/hmartiro/riffusion-inference
https://github.com/hmartiro/riffusion-inference
GitHub
GitHub - riffusion/riffusion-hobby: Stable diffusion for real-time music generation
Stable diffusion for real-time music generation. Contribute to riffusion/riffusion-hobby development by creating an account on GitHub.
#typescript #android #audio #ios #media_control #music_library #music_player #react_native #windows
https://github.com/doublesymmetry/react-native-track-player
https://github.com/doublesymmetry/react-native-track-player
GitHub
GitHub - doublesymmetry/react-native-track-player: A fully fledged audio module created for music apps. Provides audio playback…
A fully fledged audio module created for music apps. Provides audio playback, external media controls, background mode and more! - doublesymmetry/react-native-track-player
#python #artificial_intelligence #attention_mechanisms #audio_synthesis #deep_learning #transformers
https://github.com/lucidrains/audiolm-pytorch
https://github.com/lucidrains/audiolm-pytorch
GitHub
GitHub - lucidrains/audiolm-pytorch: Implementation of AudioLM, a SOTA Language Modeling Approach to Audio Generation out of Google…
Implementation of AudioLM, a SOTA Language Modeling Approach to Audio Generation out of Google Research, in Pytorch - lucidrains/audiolm-pytorch
#python #audio #deep_learning #noise_suppression #pytorch #rust #speech #speech_enhancement
https://github.com/Rikorose/DeepFilterNet
https://github.com/Rikorose/DeepFilterNet
GitHub
GitHub - Rikorose/DeepFilterNet: Noise supression using deep filtering
Noise supression using deep filtering. Contribute to Rikorose/DeepFilterNet development by creating an account on GitHub.
❤1
#typescript #audio_player #deep_learning #electron #emby #jellyfin #librosa #media_server #music #music_player #music_server #navidrome #nsmusics #player #subsonic #vue3
NSMusicS is a free, open-source music platform designed to work on various operating systems like Windows, Linux, and macOS. It supports multiple streaming servers and offers features such as music library management and internationalization. The platform includes desktop, mobile, and server components, ensuring compatibility and data interoperability across different devices. This means users can enjoy a seamless music experience without relying on proprietary software, making it user-friendly and accessible globally.
https://github.com/Super-Badmen-Viper/NSMusicS
NSMusicS is a free, open-source music platform designed to work on various operating systems like Windows, Linux, and macOS. It supports multiple streaming servers and offers features such as music library management and internationalization. The platform includes desktop, mobile, and server components, ensuring compatibility and data interoperability across different devices. This means users can enjoy a seamless music experience without relying on proprietary software, making it user-friendly and accessible globally.
https://github.com/Super-Badmen-Viper/NSMusicS
GitHub
GitHub - Super-Badmen-Viper/NSMusicS: NSMusicS NineSong Cloud-Native Music Server/ Full platform Client,support navidrome, jellyfin…
NSMusicS NineSong Cloud-Native Music Server/ Full platform Client,support navidrome, jellyfin, emby - Super-Badmen-Viper/NSMusicS
#cplusplus #android #audio_processing #c_plus_plus #calculator #computer_vision #deep_learning #framework #graph_based #graph_framework #inference #machine_learning #mediapipe #mobile_development #perception #pipeline_framework #stream_processing #video_processing
MediaPipe is a tool that helps you add smart machine learning features to your apps and devices. It works on mobile, web, desktop, and other devices. You can use pre-made solutions for tasks like vision, text, and audio processing, or customize the models to fit your needs. MediaPipe also offers tools like Model Maker and Studio to help you create and test your solutions easily. This makes it easier to delight your customers with innovative features without needing deep machine learning expertise.
https://github.com/google-ai-edge/mediapipe
MediaPipe is a tool that helps you add smart machine learning features to your apps and devices. It works on mobile, web, desktop, and other devices. You can use pre-made solutions for tasks like vision, text, and audio processing, or customize the models to fit your needs. MediaPipe also offers tools like Model Maker and Studio to help you create and test your solutions easily. This makes it easier to delight your customers with innovative features without needing deep machine learning expertise.
https://github.com/google-ai-edge/mediapipe
GitHub
GitHub - google-ai-edge/mediapipe: Cross-platform, customizable ML solutions for live and streaming media.
Cross-platform, customizable ML solutions for live and streaming media. - google-ai-edge/mediapipe
#typescript #audio_fingerprinting #browser #browser_fingerprint #browser_fingerprinting #detection #fingerprint #fingerprinting #fingerprintjs #fraud #fraud_detection #identification #javascript #visitor_identifier
FingerprintJS is a tool that helps identify website visitors by collecting and hashing browser attributes. It works even in incognito or private browsing modes and when browser data is cleared. This means you can track visitors more reliably than with cookies or local storage. However, it has limited accuracy (40-60%) because it relies on browser attributes that might be the same for different users. For higher accuracy (99.5%), you can use Fingerprint Identification, a commercial version that processes data on the server and includes more detailed information like IP addresses and visit patterns. This makes it better for preventing fraud and improving user experiences.
https://github.com/fingerprintjs/fingerprintjs
FingerprintJS is a tool that helps identify website visitors by collecting and hashing browser attributes. It works even in incognito or private browsing modes and when browser data is cleared. This means you can track visitors more reliably than with cookies or local storage. However, it has limited accuracy (40-60%) because it relies on browser attributes that might be the same for different users. For higher accuracy (99.5%), you can use Fingerprint Identification, a commercial version that processes data on the server and includes more detailed information like IP addresses and visit patterns. This makes it better for preventing fraud and improving user experiences.
https://github.com/fingerprintjs/fingerprintjs
GitHub
GitHub - fingerprintjs/fingerprintjs: The most advanced open-source browser fingerprinting library
The most advanced open-source browser fingerprinting library - fingerprintjs/fingerprintjs
#python #audio #docker_image #dsp #equalizer #filter #limiter #mastering #matchering #matching #music #numpy #python #python_library #python3 #scipy #sound #spectrum #vst
Matchering 2.0 is a tool that helps make your music sound like your favorite songs. Here’s how it works: you give it two audio files - one you want to master (your song) and another you want it to sound like (a reference song). Matchering then adjusts your song to match the volume, frequency response, peak amplitude, and stereo width of the reference song. This means you can quickly make your music sound professional and consistent, just like popular tracks. You can use it without installation through online services or install it on your computer for more control. It's especially useful for music producers and audio engineers who want to enhance their tracks easily.
https://github.com/sergree/matchering
Matchering 2.0 is a tool that helps make your music sound like your favorite songs. Here’s how it works: you give it two audio files - one you want to master (your song) and another you want it to sound like (a reference song). Matchering then adjusts your song to match the volume, frequency response, peak amplitude, and stereo width of the reference song. This means you can quickly make your music sound professional and consistent, just like popular tracks. You can use it without installation through online services or install it on your computer for more control. It's especially useful for music producers and audio engineers who want to enhance their tracks easily.
https://github.com/sergree/matchering
GitHub
GitHub - sergree/matchering: 🎚️ Open Source Audio Matching and Mastering
🎚️ Open Source Audio Matching and Mastering. Contribute to sergree/matchering development by creating an account on GitHub.
❤1
#cplusplus #audio #c #c_plus_plus #dash #hevc #hls #live #live_streaming #low_latency #media_server #multimedia #prometheus_exporter #rtmp #server_side #srt #streaming #video #video_conferencing #video_streaming #webrtc
SRS (Simple Realtime Server) is a powerful and efficient video server that supports multiple streaming protocols like RTMP, WebRTC, HLS, and more. It works on various operating systems (Linux, Windows, macOS) and hardware architectures. You can easily set it up using Docker and stream videos using tools like FFmpeg or OBS. SRS is free and open-source, licensed under MIT, making it a great choice for developers to build high-quality streaming platforms. It also offers extensive documentation and community support, making it easier to get started and troubleshoot issues.
https://github.com/ossrs/srs
SRS (Simple Realtime Server) is a powerful and efficient video server that supports multiple streaming protocols like RTMP, WebRTC, HLS, and more. It works on various operating systems (Linux, Windows, macOS) and hardware architectures. You can easily set it up using Docker and stream videos using tools like FFmpeg or OBS. SRS is free and open-source, licensed under MIT, making it a great choice for developers to build high-quality streaming platforms. It also offers extensive documentation and community support, making it easier to get started and troubleshoot issues.
https://github.com/ossrs/srs
GitHub
GitHub - ossrs/srs: SRS is a simple, high-efficiency, real-time media server supporting RTMP, WebRTC, HLS, HTTP-FLV, HTTP-TS, SRT…
SRS is a simple, high-efficiency, real-time media server supporting RTMP, WebRTC, HLS, HTTP-FLV, HTTP-TS, SRT, MPEG-DASH, and GB28181, with codec support for H.264, H.265, AV1, VP9, AAC, Opus, and ...
🔥1
#python #asr #audio #audio_processing #deep_learning #huggingface #language_model #pytorch #speaker_diarization #speaker_recognition #speaker_verification #speech_enhancement #speech_processing #speech_recognition #speech_separation #speech_to_text #speech_toolkit #speechrecognition #spoken_language_understanding #transformers #voice_recognition
SpeechBrain is an open-source toolkit that helps you quickly develop Conversational AI technologies, such as speech assistants, chatbots, and language models. It uses PyTorch and offers many pre-trained models and tutorials to make it easy to get started. You can train models for various tasks like speech recognition, speaker recognition, and text processing with just a few lines of code. SpeechBrain also supports GPU training, dynamic batching, and integration with HuggingFace models, making it powerful and efficient. This toolkit is beneficial because it simplifies the development process, provides extensive documentation and tutorials, and is highly customizable, making it ideal for research, prototyping, and educational purposes.
https://github.com/speechbrain/speechbrain
SpeechBrain is an open-source toolkit that helps you quickly develop Conversational AI technologies, such as speech assistants, chatbots, and language models. It uses PyTorch and offers many pre-trained models and tutorials to make it easy to get started. You can train models for various tasks like speech recognition, speaker recognition, and text processing with just a few lines of code. SpeechBrain also supports GPU training, dynamic batching, and integration with HuggingFace models, making it powerful and efficient. This toolkit is beneficial because it simplifies the development process, provides extensive documentation and tutorials, and is highly customizable, making it ideal for research, prototyping, and educational purposes.
https://github.com/speechbrain/speechbrain
GitHub
GitHub - speechbrain/speechbrain: A PyTorch-based Speech Toolkit
A PyTorch-based Speech Toolkit. Contribute to speechbrain/speechbrain development by creating an account on GitHub.
#python #audio_generation #audio_synthesis #audioldm #audit #fastspeech2 #hifi_gan #music_generation #naturalspeech2 #singing_voice_conversion #speech_synthesis #text_to_audio #text_to_speech #vall_e #vits #voice_conversion
Amphion is a toolkit for generating audio, music, and speech. It helps researchers and engineers, especially beginners, by providing tools for various tasks like turning text into speech (TTS), singing voice conversion (SVC), and text to audio (TTA). Amphion includes visualizations to help understand how these models work, which is very useful for learning. It also offers different vocoders to produce high-quality audio and evaluation metrics to ensure the generated audio is good. This toolkit is free to use under the MIT License and can be installed easily using Python or Docker. Using Amphion, you can create high-quality audio and music with advanced features, making it a powerful tool for both research and practical applications.
https://github.com/open-mmlab/Amphion
Amphion is a toolkit for generating audio, music, and speech. It helps researchers and engineers, especially beginners, by providing tools for various tasks like turning text into speech (TTS), singing voice conversion (SVC), and text to audio (TTA). Amphion includes visualizations to help understand how these models work, which is very useful for learning. It also offers different vocoders to produce high-quality audio and evaluation metrics to ensure the generated audio is good. This toolkit is free to use under the MIT License and can be installed easily using Python or Docker. Using Amphion, you can create high-quality audio and music with advanced features, making it a powerful tool for both research and practical applications.
https://github.com/open-mmlab/Amphion
GitHub
GitHub - open-mmlab/Amphion: Amphion (/æmˈfaɪən/) is a toolkit for Audio, Music, and Speech Generation. Its purpose is to support…
Amphion (/æmˈfaɪən/) is a toolkit for Audio, Music, and Speech Generation. Its purpose is to support reproducible research and help junior researchers and engineers get started in the field of audi...
👍1
#cplusplus #ai #api #audio_generation #distributed #gemma #gpt4all #image_generation #kubernetes #llama #llama3 #llm #mamba #mistral #musicgen #p2p #rerank #rwkv #stable_diffusion #text_generation #tts
LocalAI is a free, open-source alternative to OpenAI that you can run on your own computer or server. It allows you to generate text, images, and audio locally without needing a GPU. You can use it with various models and it supports multiple functionalities like text-to-audio, audio-to-text, and image generation. LocalAI is easy to set up using an installer script or Docker, and it has a user-friendly web interface. This tool is beneficial because it saves you money by not requiring cloud services and gives you full control over your data privacy. Plus, it's community-driven, so there are many resources and integrations available to help you get started and customize it to your needs.
https://github.com/mudler/LocalAI
LocalAI is a free, open-source alternative to OpenAI that you can run on your own computer or server. It allows you to generate text, images, and audio locally without needing a GPU. You can use it with various models and it supports multiple functionalities like text-to-audio, audio-to-text, and image generation. LocalAI is easy to set up using an installer script or Docker, and it has a user-friendly web interface. This tool is beneficial because it saves you money by not requiring cloud services and gives you full control over your data privacy. Plus, it's community-driven, so there are many resources and integrations available to help you get started and customize it to your needs.
https://github.com/mudler/LocalAI
GitHub
GitHub - mudler/LocalAI: :robot: The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and local-first. Drop…
:robot: The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and local-first. Drop-in replacement for OpenAI, running on consumer-grade hardware. No GPU required. Runs gguf,...
#ruby #audio #hotwire #music #music_player #music_streaming #musicplayer #rails #ruby #self_hosted
Black Candy is a self-hosted music streaming server that lets you manage your music collection easily. You can try it out with a demo account or install it using Docker, which makes setup simple. Once installed, you can access your music library through a web interface or mobile apps available on App Store and F-Droid. Black Candy also supports advanced features like using PostgreSQL for the database, configuring Nginx for better file delivery, and integrating with Discogs API for artist and album images. This tool helps you keep all your music organized and accessible from anywhere, making it a convenient personal music center.
https://github.com/blackcandy-org/blackcandy
Black Candy is a self-hosted music streaming server that lets you manage your music collection easily. You can try it out with a demo account or install it using Docker, which makes setup simple. Once installed, you can access your music library through a web interface or mobile apps available on App Store and F-Droid. Black Candy also supports advanced features like using PostgreSQL for the database, configuring Nginx for better file delivery, and integrating with Discogs API for artist and album images. This tool helps you keep all your music organized and accessible from anywhere, making it a convenient personal music center.
https://github.com/blackcandy-org/blackcandy
GitHub
GitHub - blackcandy-org/blackcandy: A self hosted music streaming server
A self hosted music streaming server. Contribute to blackcandy-org/blackcandy development by creating an account on GitHub.
#cplusplus #aax #au #audio #audiounit #auv3 #c_plus_plus #cpp #framework #juce #plugin #vst #vst3
JUCE is a free, open-source framework that helps you create desktop and mobile applications, including audio plug-ins. It works on many platforms like macOS, iOS, Windows, and Linux. You can use JUCE with tools like CMake or the Projucer to manage your projects easily. This makes it simple to start new projects, view tutorials, and run examples. The benefit to you is that JUCE allows you to develop cross-platform applications quickly and efficiently, saving time and effort. It also provides detailed documentation and tutorials to help you get started.
https://github.com/juce-framework/JUCE
JUCE is a free, open-source framework that helps you create desktop and mobile applications, including audio plug-ins. It works on many platforms like macOS, iOS, Windows, and Linux. You can use JUCE with tools like CMake or the Projucer to manage your projects easily. This makes it simple to start new projects, view tutorials, and run examples. The benefit to you is that JUCE allows you to develop cross-platform applications quickly and efficiently, saving time and effort. It also provides detailed documentation and tutorials to help you get started.
https://github.com/juce-framework/JUCE
GitHub
GitHub - juce-framework/JUCE: JUCE is an open-source cross-platform C++ application framework for desktop and mobile applications…
JUCE is an open-source cross-platform C++ application framework for desktop and mobile applications, including VST, VST3, AU, AUv3, LV2 and AAX audio plug-ins. - juce-framework/JUCE
#python #apple_silicon #audio_processing #mlx #multimodal #speech_recognition #speech_synthesis #speech_to_text #text_to_speech #transformers
MLX-Audio is a powerful tool for converting text into speech and speech into new audio. It works well on Apple Silicon devices, like M-series chips, making it fast and efficient. You can choose from different languages and voices, and even adjust how fast the speech is. It also includes a web interface where you can see audio in 3D and play your own files. This tool is helpful for making audiobooks, interactive media, and personal projects because it's easy to use and provides high-quality audio quickly.
https://github.com/Blaizzy/mlx-audio
MLX-Audio is a powerful tool for converting text into speech and speech into new audio. It works well on Apple Silicon devices, like M-series chips, making it fast and efficient. You can choose from different languages and voices, and even adjust how fast the speech is. It also includes a web interface where you can see audio in 3D and play your own files. This tool is helpful for making audiobooks, interactive media, and personal projects because it's easy to use and provides high-quality audio quickly.
https://github.com/Blaizzy/mlx-audio
GitHub
GitHub - Blaizzy/mlx-audio: A text-to-speech (TTS), speech-to-text (STT) and speech-to-speech (STS) library built on Apple's MLX…
A text-to-speech (TTS), speech-to-text (STT) and speech-to-speech (STS) library built on Apple's MLX framework, providing efficient speech analysis on Apple Silicon. - Blaizzy/mlx-audio
#c_lang #alsa #audio_visualizer #freebsd #glsl_shaders #linux #macos #ncurses #pipewire #portaudio #pulseaudio #sdl2 #sndio #windows
Cava is a free, open-source audio visualizer that works on Linux, FreeBSD, macOS, and Windows, letting you see music as moving bars right in your terminal or desktop window[1][3]. It’s easy to install and use, supports many audio systems, and lets you customize colors and settings. The main benefit is that it makes listening to music more fun and visually engaging, especially for people who enjoy seeing sound represented in real time, and it works on almost any computer or device[1][3][4].
https://github.com/karlstav/cava
Cava is a free, open-source audio visualizer that works on Linux, FreeBSD, macOS, and Windows, letting you see music as moving bars right in your terminal or desktop window[1][3]. It’s easy to install and use, supports many audio systems, and lets you customize colors and settings. The main benefit is that it makes listening to music more fun and visually engaging, especially for people who enjoy seeing sound represented in real time, and it works on almost any computer or device[1][3][4].
https://github.com/karlstav/cava
GitHub
GitHub - karlstav/cava: Cross-platform Audio Visualizer
Cross-platform Audio Visualizer. Contribute to karlstav/cava development by creating an account on GitHub.
#cmake #audio #ios #linux #macos #plugins #sdk #vst3 #win32
VST 3 is an improved version of the VST audio plug-in interface. It offers several benefits, including better performance by only processing audio when needed, dynamic input/output configurations, and precise automation. Users can also enjoy a more organized interface and support for advanced audio features like 3D sound. These improvements make it easier for developers to create plugins and for users to work with them in digital audio workstations (DAWs), enhancing overall audio production efficiency.
https://github.com/steinbergmedia/vst3sdk
VST 3 is an improved version of the VST audio plug-in interface. It offers several benefits, including better performance by only processing audio when needed, dynamic input/output configurations, and precise automation. Users can also enjoy a more organized interface and support for advanced audio features like 3D sound. These improvements make it easier for developers to create plugins and for users to work with them in digital audio workstations (DAWs), enhancing overall audio production efficiency.
https://github.com/steinbergmedia/vst3sdk
GitHub
GitHub - steinbergmedia/vst3sdk: VST 3 Plug-In SDK
VST 3 Plug-In SDK. Contribute to steinbergmedia/vst3sdk development by creating an account on GitHub.
#python #audio_generation #diffusion #image_generation #inference #model_serving #multimodal #pytorch #transformer #video_generation
vLLM-Omni is a free, open-source tool that makes serving AI models for text, images, videos, and audio fast, easy, and cheap. It builds on vLLM for top speed using smart memory tricks, overlapping tasks, and flexible resource sharing across GPUs. You get 2x higher throughput, 35% less delay, and simple setup with Hugging Face models via OpenAI API—perfect for building quick multi-modal apps like chatbots or media generators without high costs.
https://github.com/vllm-project/vllm-omni
vLLM-Omni is a free, open-source tool that makes serving AI models for text, images, videos, and audio fast, easy, and cheap. It builds on vLLM for top speed using smart memory tricks, overlapping tasks, and flexible resource sharing across GPUs. You get 2x higher throughput, 35% less delay, and simple setup with Hugging Face models via OpenAI API—perfect for building quick multi-modal apps like chatbots or media generators without high costs.
https://github.com/vllm-project/vllm-omni
GitHub
GitHub - vllm-project/vllm-omni: A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models - vllm-project/vllm-omni