GitHub Trends
10.1K subscribers
15.3K links
See what the GitHub community is most excited about today.

A bot automatically fetches new repositories from https://github.com/trending and sends them to the channel.

Author and maintainer: https://github.com/katursis
Download Telegram
#python #llama #transformer #tts #valle #vits #vqgan #vqvae

Fish Speech is a powerful tool that converts text into speech in many languages, including English, Japanese, Korean, Chinese, and more. You can use it by inputting a short vocal sample to generate high-quality speech. It supports multiple languages without needing phonemes and is highly accurate with low error rates. The tool is fast, with real-time processing on various devices, and has a user-friendly web and GUI interface. You can try the demo online or set it up locally. It's released under a CC BY-NC-SA 4.0 license, which means you can use and modify it freely, but you must give credit and share any changes under the same license. This tool helps you create realistic speech quickly and easily, making it useful for various applications like voice cloning and multilingual communication.

https://github.com/fishaudio/fish-speech
#typescript #agent_monitoring #analytics #evaluation #gpt #langchain #large_language_models #llama_index #llm #llm_cost #llm_evaluation #llm_observability #llmops #monitoring #open_source #openai #playground #prompt_engineering #prompt_management #ycombinator

Helicone is an all-in-one, open-source platform for developing and managing Large Language Models (LLMs). It allows you to integrate with various LLM providers like OpenAI, Anthropic, and more with just one line of code. You can observe and debug your model's performance, analyze metrics such as cost and latency, and fine-tune your models easily. The platform also offers a playground to test and iterate on prompts and sessions, and it supports prompt management and automatic evaluations. Helicone is enterprise-ready, compliant with SOC 2 and GDPR, and offers a generous free tier of 100k requests per month. This makes it easier to manage and optimize your LLM projects efficiently.

https://github.com/Helicone/helicone
1
#python #agent #ai #chatbot #chatgpt #docker #function_calling #gemini #gpt #llama #llm #ollama #openai #python #qq #qqbot #qqchannel #telegram

AstrBot is a powerful chatbot and development framework that supports multiple messaging platforms like QQ, WeChat, Telegram, and more. It integrates with large language models (LLMs) such as OpenAI, Google Gemini, and others, allowing for multi-round conversations, personality settings, and multimodal capabilities like image understanding and speech-to-text. The bot has a user-friendly plugin system, a visual management panel, and high stability due to its modular design. This makes it easy to deploy and manage, with various deployment options including Docker, Windows, and Replit. Using AstrBot benefits users by providing a versatile and highly customizable chatbot solution that can be easily extended with new features through plugins.

https://github.com/Soulter/AstrBot
#cplusplus #ggml #llama

The `llama.cpp` project allows you to run large language models (LLMs) like LLaMA and others with high performance on various hardware, including local machines and cloud services. Here are the key benefits It works on Apple Silicon, x86 architectures, NVIDIA, AMD, and Moore Threads GPUs, as well as CPUs, ensuring you can use it on a wide range of devices.
- **Optimized Performance** You can build and run the project locally, install it via package managers, use Docker images, or download pre-built binaries.
- **Extensive Model Support** It includes tools like `llama-cli` for simple text completion, `llama-server` for setting up an HTTP server, and `llama-perplexity` for measuring model quality.

This makes `llama.cpp` a powerful and flexible tool for anyone looking to work with LLMs efficiently.

https://github.com/ggerganov/llama.cpp
#python #deepseek #deepseek_r1 #fine_tuning #finetuning #gemma #gemma2 #llama #llama3 #llm #llms #lora #mistral #phi3 #qlora #unsloth

Using Unsloth.ai, you can finetune AI models like Llama, Mistral, and others up to 2x faster and with 70% less memory. The process is beginner-friendly; you just need to add your dataset, click "Run All" in the provided notebooks, and you'll get a faster, finetuned model that can be exported or uploaded to platforms like Hugging Face. This saves time and resources, making it easier to work with large AI models without needing powerful hardware. Additionally, Unsloth supports various features like 4-bit quantization, long context windows, and integration with tools from Hugging Face, making it a powerful tool for AI model development.

https://github.com/unslothai/unsloth
#python #cuda #deepseek #deepseek_llm #deepseek_v3 #inference #llama #llama2 #llama3 #llama3_1 #llava #llm #llm_serving #moe #pytorch #transformer #vlm

SGLang is a tool that makes working with large language models and vision language models much faster and more manageable. It has a fast backend runtime that optimizes model performance with features like prefix caching, continuous batching, and quantization. The frontend language is flexible and easy to use, allowing for complex tasks like chained generation calls and multi-modal inputs. SGLang supports many different models and has an active community behind it. This means you can get your models running quickly and efficiently, saving time and resources. Additionally, the extensive documentation and community support make it easier to get started and resolve any issues.

https://github.com/sgl-project/sglang
#typescript #ai #artificial_intelligence #browser #browser_automation #gpt #gpt_4 #langchain #llama #llm #openai #playwright #puppeteer #scraper

LLM Scraper is a tool that helps you get structured data from any webpage using large language models (LLMs). It supports different AI providers like OpenAI and Ollama, and it uses the Playwright framework to work with web pages. You can define what data you want to extract using schemas, which makes sure everything is organized correctly. This tool also allows you to generate code automatically for scraping tasks, making it easier to reuse scripts. The benefit is that you can easily collect data from websites in a structured way, which is helpful for projects that need specific information from the internet.

https://github.com/mishushakov/llm-scraper
#jupyter_notebook #ai #langchain #llama_index #llm #llms #opeani #python #rag #tutorials

This project is about improving Retrieval-Augmented Generation (RAG) systems, which combine information retrieval with AI to generate more accurate and relevant responses. By sponsoring this project through GitHub Sponsors, you help support the development of these advanced techniques. Your sponsorship fuels innovation in RAG technologies, allowing for better maintenance and expansion of this valuable resource. This benefits users by providing them with cutting-edge tools and insights that enhance their work with AI systems.

https://github.com/NirDiamant/RAG_Techniques
👍1
#swift #ai #aichat #chatbot #chatgpt #deepseek #deepseek_r1 #gemma #gemma3 #gguf #llama #llama3 #llm #macos #qwen #qwen2 #qwq #qwq_32b #rag #swift #swiftui

Sidekick is a local-first AI application for Macs that helps you find information from your files, folders, and websites without needing the internet. It's private, so your data stays secure on your device. You can ask questions like "Did the Aztecs use captured Spanish weapons?" and get answers with references. Sidekick also supports image generation, LaTeX rendering, and more. This makes it useful for research and work because it keeps your data safe and provides quick access to relevant information.

https://github.com/johnbean393/Sidekick
#cplusplus #accelerator #llama #llm #low_level_programming #metal #mistral #mixtral #ml #resnet #stable_diffusion #tenstorrent

Tenstorrent's TT-Metal is a powerful tool for developing AI models. It allows users to create custom kernels for their hardware, which can improve performance by reducing memory usage. This is especially useful for large language models (LLMs) like Llama and Mixtral. The TT-Metal system supports efficient data movement and computation, making it beneficial for users who need to run complex AI tasks quickly and effectively. By optimizing how data is stored and processed, TT-Metal helps users achieve better results with less effort.

https://github.com/tenstorrent/tt-metal
#typescript #electron #llama #llms #lora #mlx #rlhf #transformers

Transformer Lab is a free, open-source tool that lets you easily work with large language models on your own computer, offering one-click downloads for popular models like Llama3 and Mistral, fine-tuning across different hardware (including Apple Silicon and GPUs), and features like chatting, training, and evaluating models through a simple interface—saving you from complex setups like CUDA or Python version issues[1][2][5].

https://github.com/transformerlab/transformerlab-app
#java #anthropic #chatgpt #chroma #embeddings #gemini #gpt #huggingface #java #langchain #llama #milvus #ollama #onnx #openai #openai_api #pgvector #pinecone #vector_database #weaviate

LangChain4j helps you add powerful AI to your Java applications by making it easy to use Large Language Models (LLMs). It provides a simple way to switch between different LLMs and embedding stores without needing to learn each one's specific API. This means you can easily experiment with different models and tools, making your development process faster and more flexible. LangChain4j also offers many examples and tools to help you build complex AI applications quickly, such as chatbots and retrieval systems. This simplifies the integration of AI into your projects, allowing you to focus on creating better applications.

https://github.com/langchain4j/langchain4j
#rust #ai #ai_engineering #anthropic #artificial_intelligence #deep_learning #genai #generative_ai #gpt #large_language_models #llama #llm #llmops #llms #machine_learning #ml #ml_engineering #mlops #openai #python #rust

TensorZero is a free, open-source tool that helps you build and improve large language model (LLM) applications by using real-world data and feedback. It gives you one simple API to connect with all major LLM providers, collects data from your app’s use, and lets you easily test and improve prompts, models, and strategies. You can see how your LLMs perform, compare different options, and make them smarter, faster, and cheaper over time—all while keeping your data private and under your control. This means you get better results with less effort and cost, and your apps keep improving as you use them[1][2][3].

https://github.com/tensorzero/tensorzero
#typescript #ai_gateway #gateway #generative_ai #hacktoberfest #langchain #llama_index #llmops #llms #openai #prompt_engineering #router

The AI Gateway by Portkey lets you connect to over 1600 AI models quickly and securely through one simple API, making it easy to integrate any language, vision, or audio AI model in under two minutes. It ensures fast responses with less than 1ms latency, automatic retries, load balancing, and fallback options to keep your AI apps reliable and scalable. It also offers strong security with role-based access, guardrails, and compliance with standards like SOC2 and GDPR. You can save costs with smart caching and optimize usage without changing your code. This helps you build powerful, cost-effective, and secure AI applications faster and with less hassle.

https://github.com/Portkey-AI/gateway
#python #ai #faiss #gpt_oss #langchain #llama_index #llm #localstorage #offline_first #ollama #privacy #python #rag #retrieval_augmented_generation #vector_database #vector_search #vectors

LEANN is a tiny, powerful vector database that lets you turn your laptop into a personal AI assistant capable of searching millions of documents using 97% less storage than traditional systems without losing accuracy. It works by storing a compact graph and computing embeddings only when needed, saving huge space and keeping your data private on your device. You can search your files, emails, browser history, chat logs, live data from platforms like Slack and Twitter, and even codebases—all locally without cloud costs. This means fast, private, and efficient AI-powered search and retrieval on your own laptop.

https://github.com/yichuan-w/LEANN
#go #gemma3 #go #gpt_oss #granite4 #llama #llama3 #llm #on_device_ai #phi3 #qwen3 #qwen3vl #sdk #stable_diffusion #vlm

NexaSDK runs AI models locally on CPUs, GPUs, and NPUs with a single command, supports GGUF/MLX/.nexa formats, and offers NPU-first Android and macOS support for fast, multimodal (text, image, audio) inference, plus an OpenAI‑compatible API for easy integration. This gives you low-latency, private on-device AI across laptops, phones, and embedded systems, reduces cloud costs and data exposure, and lets you deploy and test new models immediately on target hardware for faster development and better user experience.

https://github.com/NexaAI/nexa-sdk