GitHub Trends
10.1K subscribers
15.3K links
See what the GitHub community is most excited about today.

A bot automatically fetches new repositories from https://github.com/trending and sends them to the channel.

Author and maintainer: https://github.com/katursis
Download Telegram
#cplusplus #aarch64 #arm #arm64 #avx2 #avx512 #c_plus_plus #clang #clang_cl #cpp11 #gcc_compiler #json #json_parser #json_pointer #loongarch #neon #simd #sse42 #vs2019 #x64

The simdjson library is very fast and efficient for parsing JSON files. It uses special computer instructions called SIMD to parse JSON up to 4 times faster than other popular parsers. Here are the key benefits Parses JSON much quicker than other libraries.
- **Easy to Use** Ensures full JSON and UTF-8 validation without losing any data.
- **Automatic Optimization** Designed to avoid unexpected errors and surprises.

Using simdjson can significantly speed up your application's performance when dealing with large amounts of JSON data.

https://github.com/simdjson/simdjson
#cplusplus #arm #convolution #deep_learning #embedded_devices #llm #machine_learning #ml #mnn #transformer #vulkan #winograd_algorithm

MNN is a lightweight and efficient deep learning framework that helps run AI models on mobile devices and other small devices. It supports many types of AI models and can handle tasks like image recognition and language processing quickly and locally on your device. This means you can use AI features without needing to send data to the cloud, which improves privacy and speed. MNN is used in many apps, including those from Alibaba, and supports various platforms like Android and iOS. It also helps reduce the size of AI models, making them faster and more efficient.

https://github.com/alibaba/MNN
#cplusplus #arm #baidu #deep_learning #embedded #fpga #mali #mdl #mobile #mobile_deep_learning #neural_network

Paddle Lite is a lightweight, high-performance deep learning inference framework designed to run AI models efficiently on mobile, embedded, and edge devices. It supports multiple platforms like Android, iOS, Linux, Windows, and macOS, and languages including C++, Java, and Python. You can easily convert models from other frameworks to PaddlePaddle format, optimize them for faster and smaller deployment, and run them with ready-made examples. This helps you deploy AI applications quickly on various devices with low memory use and fast speed, making it ideal for real-time, resource-limited environments. It also supports many hardware accelerators for better performance.

https://github.com/PaddlePaddle/Paddle-Lite