GitHub Trends
10.1K subscribers
15.3K links
See what the GitHub community is most excited about today.

A bot automatically fetches new repositories from https://github.com/trending and sends them to the channel.

Author and maintainer: https://github.com/katursis
Download Telegram
#go #gemma #gemma2 #go #golang #llama #llama2 #llama3 #llava #llm #llms #mistral #ollama #phi3

Ollama is a tool that lets you use large language models on your own computer. You can download and install it for macOS, Windows, or Linux. It supports various models like Llama 3.2, Phi 3, and others, which you can run locally using simple commands. For example, to run the Llama 3.2 model, you just need to type `ollama run llama3.2`.

The benefit to you is that you can use powerful language models without relying on cloud services, ensuring your data stays private and secure. You can also customize the models with specific prompts and settings to fit your needs. Additionally, there are many community integrations and libraries available to extend its functionality in various applications.

https://github.com/ollama/ollama
#jupyter_notebook #ai #finetuning #langchain #llama #llama2 #llm #machine_learning #python #pytorch #vllm

The `llama-recipes` repository helps you get started with Meta's Llama models, including Llama 3.2 Text and Vision. It provides example scripts and notebooks for various use cases, such as fine-tuning the models and building applications. You can use these models locally, in the cloud, or on-premises. The repository includes guides for installing the necessary tools, converting models to Hugging Face format, and using features like multimodal inference and responsible AI practices. This makes it easier for you to quickly set up and use the Llama models for your projects, saving time and effort.

https://github.com/meta-llama/llama-recipes
1
#typescript #electron #gpt #llama2 #llamacpp #localai #self_hosted

Jan is a local AI assistant that runs entirely on your device, giving you full control and privacy. It supports various hardware like NVIDIA GPUs, Apple M-series, and Linux systems. You can download and use popular language models like Llama, Gemma, and Mistral without needing an internet connection. Jan also allows you to customize it with extensions and connect to remote AI APIs. The benefit is that you can use AI without sending your data online, keeping your information private and secure.

https://github.com/janhq/jan
#python #cuda #deepseek #deepseek_llm #deepseek_v3 #inference #llama #llama2 #llama3 #llama3_1 #llava #llm #llm_serving #moe #pytorch #transformer #vlm

SGLang is a tool that makes working with large language models and vision language models much faster and more manageable. It has a fast backend runtime that optimizes model performance with features like prefix caching, continuous batching, and quantization. The frontend language is flexible and easy to use, allowing for complex tasks like chained generation calls and multi-modal inputs. SGLang supports many different models and has an active community behind it. This means you can get your models running quickly and efficiently, saving time and resources. Additionally, the extensive documentation and community support make it easier to get started and resolve any issues.

https://github.com/sgl-project/sglang