GitHub Trends
10.1K subscribers
15.3K links
See what the GitHub community is most excited about today.

A bot automatically fetches new repositories from https://github.com/trending and sends them to the channel.

Author and maintainer: https://github.com/katursis
Download Telegram
#python #deep_learning #intel #machine_learning #neural_network #pytorch #quantization

Intel Extension for PyTorch boosts the speed of PyTorch on Intel hardware, including both CPUs and GPUs, by using special features like AVX-512, AMX, and XMX for faster calculations[5][2][4]. It supports many popular large language models (LLMs) such as Llama, Qwen, Phi, and DeepSeek, offering optimizations for different data types and easy GPU acceleration. This means you can run advanced AI models much faster and more efficiently on your Intel computer, with simple setup and support for both ready-made and custom models.

https://github.com/intel/intel-extension-for-pytorch
#jupyter_notebook #ai #artificial_intelligence #chatgpt #deep_learning #from_scratch #gpt #language_model #large_language_models #llm #machine_learning #python #pytorch #transformer

You can learn how to build your own large language model (LLM) like GPT from scratch with clear, step-by-step guidance, including coding, training, and fine-tuning, all explained with examples and diagrams. This approach mirrors how big models like ChatGPT are made but is designed to run on a regular laptop without special hardware. You also get access to code for loading pretrained models and fine-tuning them for tasks like text classification or instruction following. This helps you deeply understand how LLMs work inside and lets you create your own functional AI assistant, gaining practical skills in AI development[1][2][3][4].

https://github.com/rasbt/LLMs-from-scratch
#other #automl #chatgpt #data_analysis #data_science #data_visualization #data_visualizations #deep_learning #gpt #gpt_3 #jax #keras #machine_learning #ml #nlp #python #pytorch #scikit_learn #tensorflow #transformer

This is a comprehensive, regularly updated list of 920 top open-source Python machine learning libraries, organized into 34 categories like frameworks, data visualization, NLP, image processing, and more. Each project is ranked by quality using GitHub and package manager metrics, helping you find the best tools for your needs. Popular libraries like TensorFlow, PyTorch, scikit-learn, and Hugging Face transformers are included, along with specialized ones for time series, reinforcement learning, and model interpretability. This resource saves you time by guiding you to high-quality, actively maintained libraries for building, optimizing, and deploying machine learning models efficiently.

https://github.com/ml-tooling/best-of-ml-python
#jupyter_notebook #darknet #pytorch #scaled_yolov4 #yolor #yolov3 #yolov4 #yolov7

YOLOv7 is a powerful tool for detecting objects in images and videos. It is fast, accurate, and can work well on devices with limited power, making it useful for real-time applications like self-driving cars and surveillance systems. YOLOv7 uses advanced techniques like Feature Pyramid Networks to detect objects of different sizes and can handle complex scenes with overlapping objects. This makes it beneficial for users who need quick and precise object detection in various environments.

https://github.com/WongKinYiu/yolov7
#typescript #ai #chatgpt #docsgpt #hacktoberfest #information_retrieval #language_model #llm #machine_learning #natural_language_processing #python #pytorch #rag #react #semantic_search #transformers #web_app

DocsGPT is an open-source AI tool that helps you quickly find accurate answers from many types of documents and web sources without errors. It supports formats like PDF, DOCX, images, and integrates with websites, APIs, and chat platforms like Discord and Telegram. You can deploy it privately for security, customize it to fit your brand, and connect it to tools for advanced actions. This means you save time searching for information, get reliable answers with sources, and improve productivity whether you’re a developer, support team, or business user. It’s easy to set up and scales well for many users[2][3][4].

https://github.com/arc53/DocsGPT
1
#python #deep_learning #diffusion #flax #flux #hacktoberfest #image_generation #image2image #image2video #jax #latent_diffusion_models #pytorch #score_based_generative_modeling #stable_diffusion #stable_diffusion_diffusers #text2image #text2video #video2video

The Hugging Face Diffusers library is a powerful and easy-to-use tool for generating images, audio, and 3D molecular structures using advanced diffusion models. It offers ready-to-use pretrained models and flexible components like pipelines, schedulers, and model building blocks, allowing you to quickly create or customize your own diffusion-based projects. Installation is simple via pip or conda, and you can generate high-quality outputs with just a few lines of code. This library benefits you by making cutting-edge AI generation accessible, customizable, and efficient, whether you want to run models or train your own[1][2][5].

https://github.com/huggingface/diffusers
#jupyter_notebook #chatgpt #finance #fingpt #fintech #large_language_models #machine_learning #nlp #prompt_engineering #pytorch #reinforcement_learning #robo_advisor #sentiment_analysis #technical_analysis

FinGPT is an open-source AI tool designed specifically for finance, helping you analyze financial news, predict stock prices, and get personalized investment advice quickly and affordably. Unlike costly models like BloombergGPT, FinGPT can be updated frequently with new data at a low cost, making it more accessible and timely. It uses advanced techniques like reinforcement learning from human feedback to tailor advice to your preferences, such as risk tolerance. You can use FinGPT for tasks like sentiment analysis, robo-advising, fraud detection, and portfolio optimization, helping you make smarter financial decisions with up-to-date insights.

https://github.com/AI4Finance-Foundation/FinGPT
#python #deep_learning #inference #llm #nlp #pytorch #transformer

Nano-vLLM is a small, fast, and easy-to-understand tool for running large language models offline. It matches the speed of bigger systems like vLLM but uses only about 1,200 lines of clean Python code, making it simple to read and modify. It includes smart features like prefix caching and tensor parallelism to boost performance. You can install it easily and run models like Qwen3-0.6B on your own GPU. This tool is great if you want fast, efficient AI inference without complex setups, ideal for learning, research, or small deployments on limited hardware.

https://github.com/GeeeekExplorer/nano-vllm
#jupyter_notebook #deep_learning #pytorch

You can learn PyTorch effectively in 20 days with a friendly, well-structured guide designed for those who already know some machine learning basics and have used Keras, TensorFlow, or PyTorch before. The book breaks down PyTorch concepts from easy to hard, with clear examples and practical code you can use right away. It includes a daily plan requiring 30 minutes to 2 hours, covering modeling, core concepts, APIs, and even advanced topics like GPU training and recommendation systems. This approach makes mastering PyTorch easier and faster, helping you build strong skills for deep learning projects and real applications.

https://github.com/lyhue1991/eat_pytorch_in_20_days
#python #audio_generation #diffusion #image_generation #inference #model_serving #multimodal #pytorch #transformer #video_generation

vLLM-Omni is a free, open-source tool that makes serving AI models for text, images, videos, and audio fast, easy, and cheap. It builds on vLLM for top speed using smart memory tricks, overlapping tasks, and flexible resource sharing across GPUs. You get 2x higher throughput, 35% less delay, and simple setup with Hugging Face models via OpenAI API—perfect for building quick multi-modal apps like chatbots or media generators without high costs.

https://github.com/vllm-project/vllm-omni