GitHub Trends
10.1K subscribers
15.3K links
See what the GitHub community is most excited about today.

A bot automatically fetches new repositories from https://github.com/trending and sends them to the channel.

Author and maintainer: https://github.com/katursis
Download Telegram
#jupyter_notebook #ai #artificial_intelligence #chatgpt #deep_learning #from_scratch #gpt #language_model #large_language_models #llm #machine_learning #python #pytorch #transformer

You can learn how to build your own large language model (LLM) like GPT from scratch with clear, step-by-step guidance, including coding, training, and fine-tuning, all explained with examples and diagrams. This approach mirrors how big models like ChatGPT are made but is designed to run on a regular laptop without special hardware. You also get access to code for loading pretrained models and fine-tuning them for tasks like text classification or instruction following. This helps you deeply understand how LLMs work inside and lets you create your own functional AI assistant, gaining practical skills in AI development[1][2][3][4].

https://github.com/rasbt/LLMs-from-scratch
#other #automl #chatgpt #data_analysis #data_science #data_visualization #data_visualizations #deep_learning #gpt #gpt_3 #jax #keras #machine_learning #ml #nlp #python #pytorch #scikit_learn #tensorflow #transformer

This is a comprehensive, regularly updated list of 920 top open-source Python machine learning libraries, organized into 34 categories like frameworks, data visualization, NLP, image processing, and more. Each project is ranked by quality using GitHub and package manager metrics, helping you find the best tools for your needs. Popular libraries like TensorFlow, PyTorch, scikit-learn, and Hugging Face transformers are included, along with specialized ones for time series, reinforcement learning, and model interpretability. This resource saves you time by guiding you to high-quality, actively maintained libraries for building, optimizing, and deploying machine learning models efficiently.

https://github.com/ml-tooling/best-of-ml-python
#python #data_mining #data_science #deep_learning #deep_reinforcement_learning #genetic_algorithm #machine_learning #machine_learning_from_scratch

This project offers Python code for many basic machine learning models and algorithms built from scratch, focusing on clear, understandable implementations rather than speed or optimization. You can learn how these algorithms work inside by running examples like polynomial regression, convolutional neural networks, clustering, and genetic algorithms. This hands-on approach helps you deeply understand machine learning concepts and build your own custom models. Using Python makes it easier because of its simple, readable code and flexibility, letting you quickly test and modify algorithms. This can improve your skills and confidence in machine learning development.

https://github.com/eriklindernoren/ML-From-Scratch
#other #artificial_intelligence #artificial_intelligence_projects #awesome #computer_vision #computer_vision_project #data_science #deep_learning #deep_learning_project #machine_learning #machine_learning_projects #nlp #nlp_projects #python

You can access a huge, constantly updated list of over 500 artificial intelligence projects with ready-to-use code covering machine learning, deep learning, computer vision, and natural language processing. This collection includes projects for beginners and advanced users, with links to tutorials, datasets, and real-world applications like chatbots, healthcare, and time series forecasting. Using this resource helps you learn AI by doing practical projects, speeding up your coding skills, and building a strong portfolio for jobs or research. It saves you time searching for quality projects and gives you tested, working code to study and modify.

https://github.com/ashishpatel26/500-AI-Machine-learning-Deep-learning-Computer-vision-NLP-Projects-with-code
#python #deep_learning #diffusion #flax #flux #hacktoberfest #image_generation #image2image #image2video #jax #latent_diffusion_models #pytorch #score_based_generative_modeling #stable_diffusion #stable_diffusion_diffusers #text2image #text2video #video2video

The Hugging Face Diffusers library is a powerful and easy-to-use tool for generating images, audio, and 3D molecular structures using advanced diffusion models. It offers ready-to-use pretrained models and flexible components like pipelines, schedulers, and model building blocks, allowing you to quickly create or customize your own diffusion-based projects. Installation is simple via pip or conda, and you can generate high-quality outputs with just a few lines of code. This library benefits you by making cutting-edge AI generation accessible, customizable, and efficient, whether you want to run models or train your own[1][2][5].

https://github.com/huggingface/diffusers
#python #artificial_intelligence #cloud_ml #computer_systems #courseware #deep_learning #edge_machine_learning #embedded_ml #machine_learning #machine_learning_systems #mobile_ml #textbook #tinyml

You can learn how to build real-world AI systems from start to finish with an open-source textbook originally from Harvard University. It teaches you not just how to train AI models but how to design scalable systems, manage data pipelines, deploy models in production, monitor them continuously, and optimize for devices like phones or IoT gadgets. This helps you become an engineer who can create efficient, reliable, and sustainable AI systems that work well in practice. The book offers hands-on labs, community support, and free online access, making it easier to gain practical skills in machine learning systems engineering.

https://github.com/harvard-edge/cs249r_book
#python #deep_learning #inference #llm #nlp #pytorch #transformer

Nano-vLLM is a small, fast, and easy-to-understand tool for running large language models offline. It matches the speed of bigger systems like vLLM but uses only about 1,200 lines of clean Python code, making it simple to read and modify. It includes smart features like prefix caching and tensor parallelism to boost performance. You can install it easily and run models like Qwen3-0.6B on your own GPU. This tool is great if you want fast, efficient AI inference without complex setups, ideal for learning, research, or small deployments on limited hardware.

https://github.com/GeeeekExplorer/nano-vllm
#jupyter_notebook #deep_learning #pytorch

You can learn PyTorch effectively in 20 days with a friendly, well-structured guide designed for those who already know some machine learning basics and have used Keras, TensorFlow, or PyTorch before. The book breaks down PyTorch concepts from easy to hard, with clear examples and practical code you can use right away. It includes a daily plan requiring 30 minutes to 2 hours, covering modeling, core concepts, APIs, and even advanced topics like GPU training and recommendation systems. This approach makes mastering PyTorch easier and faster, helping you build strong skills for deep learning projects and real applications.

https://github.com/lyhue1991/eat_pytorch_in_20_days
#cplusplus #arm #baidu #deep_learning #embedded #fpga #mali #mdl #mobile #mobile_deep_learning #neural_network

Paddle Lite is a lightweight, high-performance deep learning inference framework designed to run AI models efficiently on mobile, embedded, and edge devices. It supports multiple platforms like Android, iOS, Linux, Windows, and macOS, and languages including C++, Java, and Python. You can easily convert models from other frameworks to PaddlePaddle format, optimize them for faster and smaller deployment, and run them with ready-made examples. This helps you deploy AI applications quickly on various devices with low memory use and fast speed, making it ideal for real-time, resource-limited environments. It also supports many hardware accelerators for better performance.

https://github.com/PaddlePaddle/Paddle-Lite
#python #brain_inspired_ai #deep_learning #large_language_models #reasoning

The Hierarchical Reasoning Model (HRM) is a new type of AI that reasons more like a human brain, using a fast part for quick details and a slow part for big-picture planning. It solves hard logic tasks like Sudoku, mazes, and IQ-style puzzles very well, even though it is tiny (only 27 million parameters) and learns from very little data (just 1,000 examples). Unlike most large language models, it does not need long chains of written reasoning steps or huge amounts of training, which makes it much faster, cheaper, and more efficient. For the user, this means powerful reasoning in a small, fast system that can run on ordinary hardware and still beat much larger models on tough problems.

https://github.com/sapientinc/HRM
1