#cuda #training #inference #transformer #bart #beam_search #sampling #bert #multilingual_nmt #gpt_2 #diverse_decoding
https://github.com/bytedance/lightseq
https://github.com/bytedance/lightseq
GitHub
GitHub - bytedance/lightseq: LightSeq: A High Performance Library for Sequence Processing and Generation
LightSeq: A High Performance Library for Sequence Processing and Generation - bytedance/lightseq
#cplusplus #3d #3d_perception #arm #computer_graphics #cpp #cuda #gpu #gui #machine_learning #mesh_processing #odometry #opengl #pointcloud #python #pytorch #reconstruction #registration #rendering #tensorflow #visualization
https://github.com/isl-org/Open3D
https://github.com/isl-org/Open3D
GitHub
GitHub - isl-org/Open3D: Open3D: A Modern Library for 3D Data Processing
Open3D: A Modern Library for 3D Data Processing. Contribute to isl-org/Open3D development by creating an account on GitHub.
#cplusplus #cuda #deep_learning #deep_neural_networks #distributed #machine_learning #ml #neural_network
https://github.com/Oneflow-Inc/oneflow
https://github.com/Oneflow-Inc/oneflow
GitHub
GitHub - Oneflow-Inc/oneflow: OneFlow is a deep learning framework designed to be user-friendly, scalable and efficient.
OneFlow is a deep learning framework designed to be user-friendly, scalable and efficient. - Oneflow-Inc/oneflow
#cplusplus #cuda #deep_learning #gpu #mlp #nerf #neural_network #real_time #rendering
https://github.com/NVlabs/tiny-cuda-nn
https://github.com/NVlabs/tiny-cuda-nn
GitHub
GitHub - NVlabs/tiny-cuda-nn: Lightning fast C++/CUDA neural network framework
Lightning fast C++/CUDA neural network framework. Contribute to NVlabs/tiny-cuda-nn development by creating an account on GitHub.
#cuda #3d_reconstruction #computer_graphics #computer_vision #function_approximation #machine_learning #nerf #neural_network #real_time #real_time_rendering #realtime #signed_distance_functions
https://github.com/NVlabs/instant-ngp
https://github.com/NVlabs/instant-ngp
GitHub
GitHub - NVlabs/instant-ngp: Instant neural graphics primitives: lightning fast NeRF and more
Instant neural graphics primitives: lightning fast NeRF and more - NVlabs/instant-ngp
#python #cublas #cuda #cudnn #cupy #curand #cusolver #cusparse #cusparselt #cutensor #gpu #nccl #numpy #nvrtc #nvtx #rocm #scipy #tensor
https://github.com/cupy/cupy
https://github.com/cupy/cupy
GitHub
GitHub - cupy/cupy: NumPy & SciPy for GPU
NumPy & SciPy for GPU. Contribute to cupy/cupy development by creating an account on GitHub.
#jupyter_notebook #3d_reconstruction #cuda #instant_ngp #nerf #pytorch #pytorch_lightning
https://github.com/kwea123/ngp_pl
https://github.com/kwea123/ngp_pl
GitHub
GitHub - kwea123/ngp_pl: Instant-ngp in pytorch+cuda trained with pytorch-lightning (high quality with high speed, with only few…
Instant-ngp in pytorch+cuda trained with pytorch-lightning (high quality with high speed, with only few lines of legible code) - kwea123/ngp_pl
#python #command_line_tool #console #cuda #curses #gpu #gpu_monitoring #htop #monitoring #monitoring_tool #nvidia #nvidia_smi #nvml #process_monitoring #resource_monitor #top
https://github.com/XuehaiPan/nvitop
https://github.com/XuehaiPan/nvitop
GitHub
GitHub - XuehaiPan/nvitop: An interactive NVIDIA-GPU process viewer and beyond, the one-stop solution for GPU process management.
An interactive NVIDIA-GPU process viewer and beyond, the one-stop solution for GPU process management. - XuehaiPan/nvitop
#cplusplus #compiler #cuda #jax #machine_learning #mlir #pytorch #runtime #spirv #tensorflow #vulkan
IREE is a tool that helps run Machine Learning (ML) models on different devices, from big data centers to small mobile and edge devices. It uses a special way to convert ML models into a uniform format, making it easier to deploy them anywhere. This tool is still in the early stages but is being actively improved. Using IREE can help you scale your ML models efficiently across various platforms, making it beneficial for developers who need to deploy models in different environments.
https://github.com/iree-org/iree
IREE is a tool that helps run Machine Learning (ML) models on different devices, from big data centers to small mobile and edge devices. It uses a special way to convert ML models into a uniform format, making it easier to deploy them anywhere. This tool is still in the early stages but is being actively improved. Using IREE can help you scale your ML models efficiently across various platforms, making it beneficial for developers who need to deploy models in different environments.
https://github.com/iree-org/iree
GitHub
GitHub - iree-org/iree: A retargetable MLIR-based machine learning compiler and runtime toolkit.
A retargetable MLIR-based machine learning compiler and runtime toolkit. - iree-org/iree
#python #amd #cuda #gpt #inference #inferentia #llama #llm #llm_serving #llmops #mlops #model_serving #pytorch #rocm #tpu #trainium #transformer #xpu
vLLM is a library that makes it easy, fast, and cheap to use large language models (LLMs). It is designed to be fast with features like efficient memory management, continuous batching, and optimized CUDA kernels. vLLM supports many popular models and can run on various hardware including NVIDIA GPUs, AMD CPUs and GPUs, and more. It also offers seamless integration with Hugging Face models and supports different decoding algorithms. This makes it flexible and easy to use for anyone needing to serve LLMs, whether for research or other applications. You can install vLLM easily with `pip install vllm` and find detailed documentation on their website.
https://github.com/vllm-project/vllm
vLLM is a library that makes it easy, fast, and cheap to use large language models (LLMs). It is designed to be fast with features like efficient memory management, continuous batching, and optimized CUDA kernels. vLLM supports many popular models and can run on various hardware including NVIDIA GPUs, AMD CPUs and GPUs, and more. It also offers seamless integration with Hugging Face models and supports different decoding algorithms. This makes it flexible and easy to use for anyone needing to serve LLMs, whether for research or other applications. You can install vLLM easily with `pip install vllm` and find detailed documentation on their website.
https://github.com/vllm-project/vllm
GitHub
GitHub - vllm-project/vllm: A high-throughput and memory-efficient inference and serving engine for LLMs
A high-throughput and memory-efficient inference and serving engine for LLMs - vllm-project/vllm
❤1