GitHub Trends
10.1K subscribers
15.3K links
See what the GitHub community is most excited about today.

A bot automatically fetches new repositories from https://github.com/trending and sends them to the channel.

Author and maintainer: https://github.com/katursis
Download Telegram
#go #gemma #gemma2 #go #golang #llama #llama2 #llama3 #llava #llm #llms #mistral #ollama #phi3

Ollama is a tool that lets you use large language models on your own computer. You can download and install it for macOS, Windows, or Linux. It supports various models like Llama 3.2, Phi 3, and others, which you can run locally using simple commands. For example, to run the Llama 3.2 model, you just need to type `ollama run llama3.2`.

The benefit to you is that you can use powerful language models without relying on cloud services, ensuring your data stays private and secure. You can also customize the models with specific prompts and settings to fit your needs. Additionally, there are many community integrations and libraries available to extend its functionality in various applications.

https://github.com/ollama/ollama
#cplusplus #ai #api #audio_generation #distributed #gemma #gpt4all #image_generation #kubernetes #llama #llama3 #llm #mamba #mistral #musicgen #p2p #rerank #rwkv #stable_diffusion #text_generation #tts

LocalAI is a free, open-source alternative to OpenAI that you can run on your own computer or server. It allows you to generate text, images, and audio locally without needing a GPU. You can use it with various models and it supports multiple functionalities like text-to-audio, audio-to-text, and image generation. LocalAI is easy to set up using an installer script or Docker, and it has a user-friendly web interface. This tool is beneficial because it saves you money by not requiring cloud services and gives you full control over your data privacy. Plus, it's community-driven, so there are many resources and integrations available to help you get started and customize it to your needs.

https://github.com/mudler/LocalAI
#python #deepseek #deepseek_r1 #fine_tuning #finetuning #gemma #gemma2 #llama #llama3 #llm #llms #lora #mistral #phi3 #qlora #unsloth

Using Unsloth.ai, you can finetune AI models like Llama, Mistral, and others up to 2x faster and with 70% less memory. The process is beginner-friendly; you just need to add your dataset, click "Run All" in the provided notebooks, and you'll get a faster, finetuned model that can be exported or uploaded to platforms like Hugging Face. This saves time and resources, making it easier to work with large AI models without needing powerful hardware. Additionally, Unsloth supports various features like 4-bit quantization, long context windows, and integration with tools from Hugging Face, making it a powerful tool for AI model development.

https://github.com/unslothai/unsloth
#swift #ai #aichat #chatbot #chatgpt #deepseek #deepseek_r1 #gemma #gemma3 #gguf #llama #llama3 #llm #macos #qwen #qwen2 #qwq #qwq_32b #rag #swift #swiftui

Sidekick is a local-first AI application for Macs that helps you find information from your files, folders, and websites without needing the internet. It's private, so your data stays secure on your device. You can ask questions like "Did the Aztecs use captured Spanish weapons?" and get answers with references. Sidekick also supports image generation, LaTeX rendering, and more. This makes it useful for research and work because it keeps your data safe and provides quick access to relevant information.

https://github.com/johnbean393/Sidekick
#jupyter_notebook #chatglm #chatglm3 #gemma_2b_it #glm_4 #internlm2 #llama3 #llm #lora #minicpm #q_wen #qwen #qwen1_5 #qwen2

This guide helps beginners set up and use open-source large language models (LLMs) on Linux or cloud platforms like AutoDL, with step-by-step instructions for environment setup, model deployment, and fine-tuning for models such as LLaMA, ChatGLM, and InternLM[2][4][5]. It covers everything from basic installation to advanced techniques like LoRA and distributed fine-tuning, and supports integration with tools like LangChain and online demo deployment. The main benefit is making powerful AI models accessible and easy to use for students, researchers, and anyone interested in experimenting with or customizing LLMs for their own projects[2][4][5].

https://github.com/datawhalechina/self-llm