#cuda #training #inference #transformer #bart #beam_search #sampling #bert #multilingual_nmt #gpt_2 #diverse_decoding
https://github.com/bytedance/lightseq
https://github.com/bytedance/lightseq
GitHub
GitHub - bytedance/lightseq: LightSeq: A High Performance Library for Sequence Processing and Generation
LightSeq: A High Performance Library for Sequence Processing and Generation - bytedance/lightseq
#cplusplus #3d #3d_perception #arm #computer_graphics #cpp #cuda #gpu #gui #machine_learning #mesh_processing #odometry #opengl #pointcloud #python #pytorch #reconstruction #registration #rendering #tensorflow #visualization
https://github.com/isl-org/Open3D
https://github.com/isl-org/Open3D
GitHub
GitHub - isl-org/Open3D: Open3D: A Modern Library for 3D Data Processing
Open3D: A Modern Library for 3D Data Processing. Contribute to isl-org/Open3D development by creating an account on GitHub.
#cplusplus #cuda #deep_learning #deep_neural_networks #distributed #machine_learning #ml #neural_network
https://github.com/Oneflow-Inc/oneflow
https://github.com/Oneflow-Inc/oneflow
GitHub
GitHub - Oneflow-Inc/oneflow: OneFlow is a deep learning framework designed to be user-friendly, scalable and efficient.
OneFlow is a deep learning framework designed to be user-friendly, scalable and efficient. - Oneflow-Inc/oneflow
#cplusplus #cuda #deep_learning #gpu #mlp #nerf #neural_network #real_time #rendering
https://github.com/NVlabs/tiny-cuda-nn
https://github.com/NVlabs/tiny-cuda-nn
GitHub
GitHub - NVlabs/tiny-cuda-nn: Lightning fast C++/CUDA neural network framework
Lightning fast C++/CUDA neural network framework. Contribute to NVlabs/tiny-cuda-nn development by creating an account on GitHub.
#cuda #3d_reconstruction #computer_graphics #computer_vision #function_approximation #machine_learning #nerf #neural_network #real_time #real_time_rendering #realtime #signed_distance_functions
https://github.com/NVlabs/instant-ngp
https://github.com/NVlabs/instant-ngp
GitHub
GitHub - NVlabs/instant-ngp: Instant neural graphics primitives: lightning fast NeRF and more
Instant neural graphics primitives: lightning fast NeRF and more - NVlabs/instant-ngp
#python #cublas #cuda #cudnn #cupy #curand #cusolver #cusparse #cusparselt #cutensor #gpu #nccl #numpy #nvrtc #nvtx #rocm #scipy #tensor
https://github.com/cupy/cupy
https://github.com/cupy/cupy
GitHub
GitHub - cupy/cupy: NumPy & SciPy for GPU
NumPy & SciPy for GPU. Contribute to cupy/cupy development by creating an account on GitHub.
#jupyter_notebook #3d_reconstruction #cuda #instant_ngp #nerf #pytorch #pytorch_lightning
https://github.com/kwea123/ngp_pl
https://github.com/kwea123/ngp_pl
GitHub
GitHub - kwea123/ngp_pl: Instant-ngp in pytorch+cuda trained with pytorch-lightning (high quality with high speed, with only few…
Instant-ngp in pytorch+cuda trained with pytorch-lightning (high quality with high speed, with only few lines of legible code) - kwea123/ngp_pl
#python #command_line_tool #console #cuda #curses #gpu #gpu_monitoring #htop #monitoring #monitoring_tool #nvidia #nvidia_smi #nvml #process_monitoring #resource_monitor #top
https://github.com/XuehaiPan/nvitop
https://github.com/XuehaiPan/nvitop
GitHub
GitHub - XuehaiPan/nvitop: An interactive NVIDIA-GPU process viewer and beyond, the one-stop solution for GPU process management.
An interactive NVIDIA-GPU process viewer and beyond, the one-stop solution for GPU process management. - XuehaiPan/nvitop
#cplusplus #compiler #cuda #jax #machine_learning #mlir #pytorch #runtime #spirv #tensorflow #vulkan
IREE is a tool that helps run Machine Learning (ML) models on different devices, from big data centers to small mobile and edge devices. It uses a special way to convert ML models into a uniform format, making it easier to deploy them anywhere. This tool is still in the early stages but is being actively improved. Using IREE can help you scale your ML models efficiently across various platforms, making it beneficial for developers who need to deploy models in different environments.
https://github.com/iree-org/iree
IREE is a tool that helps run Machine Learning (ML) models on different devices, from big data centers to small mobile and edge devices. It uses a special way to convert ML models into a uniform format, making it easier to deploy them anywhere. This tool is still in the early stages but is being actively improved. Using IREE can help you scale your ML models efficiently across various platforms, making it beneficial for developers who need to deploy models in different environments.
https://github.com/iree-org/iree
GitHub
GitHub - iree-org/iree: A retargetable MLIR-based machine learning compiler and runtime toolkit.
A retargetable MLIR-based machine learning compiler and runtime toolkit. - iree-org/iree