#python #class_activation_maps #deep_learning #grad_cam #interpretability #interpretable_ai #interpretable_deep_learning #pytorch #score_cam #vision_transformers #visualizations
https://github.com/jacobgil/pytorch-grad-cam
https://github.com/jacobgil/pytorch-grad-cam
GitHub
GitHub - jacobgil/pytorch-grad-cam: Advanced AI Explainability for computer vision. Support for CNNs, Vision Transformers, Classification…
Advanced AI Explainability for computer vision. Support for CNNs, Vision Transformers, Classification, Object detection, Segmentation, Image similarity and more. - jacobgil/pytorch-grad-cam
#other #attention_mechanism #attention_mechanisms #awesome_list #computer_vision #deep_learning #detr #papers #self_attention #transformer #transformer_architecture #transformer_awesome #transformer_cv #transformer_models #transformer_with_cv #transformers #vision_transformer #visual_transformer #vit
https://github.com/cmhungsteve/Awesome-Transformer-Attention
https://github.com/cmhungsteve/Awesome-Transformer-Attention
GitHub
GitHub - cmhungsteve/Awesome-Transformer-Attention: An ultimately comprehensive paper list of Vision Transformer/Attention, including…
An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites - cmhungsteve/Awesome-Transformer-Attention
#python #classification #computer_vision #object_detection #pytorch #self_supervised_learning #transformers #vision_transformer
https://github.com/alibaba/EasyCV
https://github.com/alibaba/EasyCV
GitHub
GitHub - alibaba/EasyCV: An all-in-one toolkit for computer vision
An all-in-one toolkit for computer vision. Contribute to alibaba/EasyCV development by creating an account on GitHub.
#python #deep_learning #deep_learning_library #image_captioning #multimodal_datasets #multimodal_deep_learning #salesforce #vision_and_language #vision_framework #vision_language_pretraining #vision_language_transformer #visual_question_anwsering
https://github.com/salesforce/LAVIS
https://github.com/salesforce/LAVIS
GitHub
GitHub - salesforce/LAVIS: LAVIS - A One-stop Library for Language-Vision Intelligence
LAVIS - A One-stop Library for Language-Vision Intelligence - salesforce/LAVIS
#python #computer_vision #convolutional_networks #embedding_vectors #embeddings #feature_extraction #feature_vector #image_processing #image_retrieval #machine_learning #milvus #pipeline #towhee #transformer #unstructured_data #video_processing #vision_transformer #vit
https://github.com/towhee-io/towhee
https://github.com/towhee-io/towhee
GitHub
GitHub - towhee-io/towhee: Towhee is a framework that is dedicated to making neural data processing pipelines simple and fast.
Towhee is a framework that is dedicated to making neural data processing pipelines simple and fast. - towhee-io/towhee
#python #chatgpt #clip #deep_learning #gpt #hacktoberfest #hnsw #information_retrieval #knn #large_language_models #machine_learning #machinelearning #multi_modal #natural_language_processing #search_engine #semantic_search #tensor_search #transformers #vector_search #vision_language #visual_search
https://github.com/marqo-ai/marqo
https://github.com/marqo-ai/marqo
GitHub
GitHub - marqo-ai/marqo: Unified embedding generation and search engine. Also available on cloud - cloud.marqo.ai
Unified embedding generation and search engine. Also available on cloud - cloud.marqo.ai - marqo-ai/marqo
#cplusplus #artificial_intelligence #computer_vision #document #document_analysis #document_intelligence #document_recognition #document_understanding #documentai #end_to_end_ocr #multimodal #multimodal_deep_learning #ocr #scene_text_detection #scene_text_detection_recognition #scene_text_recognition #text_detection #text_recognition #vision_language #vision_language_model #vision_language_transformer
https://github.com/AlibabaResearch/AdvancedLiterateMachinery
https://github.com/AlibabaResearch/AdvancedLiterateMachinery
GitHub
GitHub - AlibabaResearch/AdvancedLiterateMachinery: A collection of original, innovative ideas and algorithms towards Advanced…
A collection of original, innovative ideas and algorithms towards Advanced Literate Machinery. This project is maintained by the OCR Team in the Language Technology Lab, Tongyi Lab, Alibaba Group. ...
👍1
#python #augmix #convnext #distributed_training #dual_path_networks #efficientnet #image_classification #imagenet #maxvit #mixnet #mobile_deep_learning #mobilenet_v2 #mobilenetv3 #nfnets #normalization_free_training #pretrained_models #pretrained_weights #pytorch #randaugment #resnet #vision_transformer_models
PyTorch Image Models (`timm`) is a comprehensive library that includes a wide range of state-of-the-art image models, layers, utilities, optimizers, and training scripts. Here are the key benefits `timm` offers over 300 pre-trained models from various families like Vision Transformers, ResNets, EfficientNets, and more, allowing you to choose the best model for your task.
- **Pre-trained Weights** You can easily extract features at different levels of the network using `features_only=True` and `out_indices`, making it versatile for various applications.
- **Optimizers and Schedulers** It provides several augmentation techniques like AutoAugment, RandAugment, and regularization methods like DropPath and DropBlock to enhance model performance.
- **Reference Training Scripts**: Included are high-performance training, validation, and inference scripts that support multiple GPUs and mixed-precision training.
Overall, `timm` simplifies the process of working with deep learning models for image tasks by providing a unified interface and extensive tools for training and evaluation.
https://github.com/huggingface/pytorch-image-models
PyTorch Image Models (`timm`) is a comprehensive library that includes a wide range of state-of-the-art image models, layers, utilities, optimizers, and training scripts. Here are the key benefits `timm` offers over 300 pre-trained models from various families like Vision Transformers, ResNets, EfficientNets, and more, allowing you to choose the best model for your task.
- **Pre-trained Weights** You can easily extract features at different levels of the network using `features_only=True` and `out_indices`, making it versatile for various applications.
- **Optimizers and Schedulers** It provides several augmentation techniques like AutoAugment, RandAugment, and regularization methods like DropPath and DropBlock to enhance model performance.
- **Reference Training Scripts**: Included are high-performance training, validation, and inference scripts that support multiple GPUs and mixed-precision training.
Overall, `timm` simplifies the process of working with deep learning models for image tasks by providing a unified interface and extensive tools for training and evaluation.
https://github.com/huggingface/pytorch-image-models
GitHub
GitHub - huggingface/pytorch-image-models: The largest collection of PyTorch image encoders / backbones. Including train, eval…
The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (V...
#python #chinese #clip #computer_vision #contrastive_loss #coreml_models #deep_learning #image_text_retrieval #multi_modal #multi_modal_learning #nlp #pretrained_models #pytorch #transformers #vision_and_language_pre_training #vision_language
This project is about a Chinese version of the CLIP (Contrastive Language-Image Pretraining) model, trained on a large dataset of Chinese text and images. Here’s what you need to know This model helps you quickly perform tasks like calculating text and image features, cross-modal retrieval (finding images based on text or vice versa), and zero-shot image classification (classifying images without any labeled examples).
- **Ease of Use** The model has been tested on various datasets and shows strong performance in zero-shot image classification and cross-modal retrieval tasks.
- **Resources**: The project includes pre-trained models, training and testing codes, and detailed tutorials on how to use the model for different tasks.
Overall, this project makes it easy to work with Chinese text and images using advanced AI techniques, saving you time and effort.
https://github.com/OFA-Sys/Chinese-CLIP
This project is about a Chinese version of the CLIP (Contrastive Language-Image Pretraining) model, trained on a large dataset of Chinese text and images. Here’s what you need to know This model helps you quickly perform tasks like calculating text and image features, cross-modal retrieval (finding images based on text or vice versa), and zero-shot image classification (classifying images without any labeled examples).
- **Ease of Use** The model has been tested on various datasets and shows strong performance in zero-shot image classification and cross-modal retrieval tasks.
- **Resources**: The project includes pre-trained models, training and testing codes, and detailed tutorials on how to use the model for different tasks.
Overall, this project makes it easy to work with Chinese text and images using advanced AI techniques, saving you time and effort.
https://github.com/OFA-Sys/Chinese-CLIP
GitHub
GitHub - OFA-Sys/Chinese-CLIP: Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation. - OFA-Sys/Chinese-CLIP
#swift #ai #android #barcode #camera #instagram #ios #javascript #jsi #library #native #qr #qrcode #react #react_native #react_native_camera #scanner #snapchat #typescript #vision #worklet
VisionCamera is a powerful camera library for React Native that offers many useful features. You can capture photos and videos, scan QR codes and barcodes, use multiple cameras, and adjust resolutions and frame rates. It also supports advanced features like facial recognition, object detection, and real-time video chats through frame processors. Additionally, you can draw shapes, text, and filters on the camera view, and it includes smooth zooming, fast pause and resume, HDR and night modes, and a custom video pipeline. Installing it is easy with npm, and there are detailed guides and examples to help you get started. Using VisionCamera can enhance your app's camera capabilities significantly.
https://github.com/mrousavy/react-native-vision-camera
VisionCamera is a powerful camera library for React Native that offers many useful features. You can capture photos and videos, scan QR codes and barcodes, use multiple cameras, and adjust resolutions and frame rates. It also supports advanced features like facial recognition, object detection, and real-time video chats through frame processors. Additionally, you can draw shapes, text, and filters on the camera view, and it includes smooth zooming, fast pause and resume, HDR and night modes, and a custom video pipeline. Installing it is easy with npm, and there are detailed guides and examples to help you get started. Using VisionCamera can enhance your app's camera capabilities significantly.
https://github.com/mrousavy/react-native-vision-camera
GitHub
GitHub - mrousavy/react-native-vision-camera: 📸 A powerful, high-performance React Native Camera library.
📸 A powerful, high-performance React Native Camera library. - mrousavy/react-native-vision-camera
❤1👍1
#python #api #automation #browser #browser_automation #computer #gpt #llm #playwright #python #rpa #vision #workflow
Skyvern is a tool that automates browser-based workflows using Large Language Models (LLMs) and computer vision. It can interact with websites without needing custom scripts, making it resistant to website layout changes. Here’s how it benefits you Skyvern can handle tasks on websites it has never seen before, filling out forms, extracting data, and even handling 2FA authentication.
- **Flexibility** Unlike traditional automation methods, Skyvern is less likely to break when website layouts change.
- **Ease of Use**: You can create tasks and workflows through a simple API or a user-friendly UI, without needing to write complex code.
Overall, Skyvern simplifies and stabilizes the automation of web-based tasks, making it easier to manage and scale your workflows.
https://github.com/Skyvern-AI/skyvern
Skyvern is a tool that automates browser-based workflows using Large Language Models (LLMs) and computer vision. It can interact with websites without needing custom scripts, making it resistant to website layout changes. Here’s how it benefits you Skyvern can handle tasks on websites it has never seen before, filling out forms, extracting data, and even handling 2FA authentication.
- **Flexibility** Unlike traditional automation methods, Skyvern is less likely to break when website layouts change.
- **Ease of Use**: You can create tasks and workflows through a simple API or a user-friendly UI, without needing to write complex code.
Overall, Skyvern simplifies and stabilizes the automation of web-based tasks, making it easier to manage and scale your workflows.
https://github.com/Skyvern-AI/skyvern
GitHub
GitHub - Skyvern-AI/skyvern: Automate browser based workflows with AI
Automate browser based workflows with AI. Contribute to Skyvern-AI/skyvern development by creating an account on GitHub.
#rust #ai #computer_vision #llm #machine_learning #ml #multimodal #vision
ScreenPipe is an AI assistant that records your screen and voice 24/7, giving you all the context you need. It's like having a personal recorder that helps you remember everything. You can use it as a desktop app, command line tool, or even integrate it into other applications. The benefit is that you'll never miss important details again, and you can prepare for the future where data is crucial. Plus, it's open-source, so you can customize it to your needs. Downloading ScreenPipe can help you stay organized and prepared in the age of super intelligence.
https://github.com/mediar-ai/screenpipe
ScreenPipe is an AI assistant that records your screen and voice 24/7, giving you all the context you need. It's like having a personal recorder that helps you remember everything. You can use it as a desktop app, command line tool, or even integrate it into other applications. The benefit is that you'll never miss important details again, and you can prepare for the future where data is crucial. Plus, it's open-source, so you can customize it to your needs. Downloading ScreenPipe can help you stay organized and prepared in the age of super intelligence.
https://github.com/mediar-ai/screenpipe
GitHub
GitHub - mediar-ai/screenpipe: AI app store powered by 24/7 desktop history. open source | 100% local | dev friendly | 24/7 screen…
AI app store powered by 24/7 desktop history. open source | 100% local | dev friendly | 24/7 screen, mic recording - mediar-ai/screenpipe
❤1
#python #auto_regressive_model #autoregressive_models #diffusion_models #generative_ai #generative_model #gpt #gpt_2 #image_generation #large_language_models #neurips #transformers #vision_transformer
VAR (Visual Autoregressive Modeling) is a new way to generate images that improves upon existing methods. It uses a "next-scale prediction" approach, which means it generates images from coarse to fine details, unlike the traditional method of predicting pixel by pixel. This makes VAR models better than diffusion models for the first time. You can try VAR on a demo website and generate images interactively, which is fun and easy. VAR also follows power-law scaling laws, making it efficient and scalable. The benefit to you is that you can create high-quality images quickly and easily, and even explore technical details through provided scripts and models.
https://github.com/FoundationVision/VAR
VAR (Visual Autoregressive Modeling) is a new way to generate images that improves upon existing methods. It uses a "next-scale prediction" approach, which means it generates images from coarse to fine details, unlike the traditional method of predicting pixel by pixel. This makes VAR models better than diffusion models for the first time. You can try VAR on a demo website and generate images interactively, which is fun and easy. VAR also follows power-law scaling laws, making it efficient and scalable. The benefit to you is that you can create high-quality images quickly and easily, and even explore technical details through provided scripts and models.
https://github.com/FoundationVision/VAR
GitHub
GitHub - FoundationVision/VAR: [NeurIPS 2024 Best Paper Award][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official…
[NeurIPS 2024 Best Paper Award][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Predi...
👍1😁1
#python #agent #ai #asr #cpp #gemini #golang #gpt_4 #gpt_4o #llm #low_latency #multimodal #nextjs14 #openai #python #rag #real_time #realtime #tts #vision #voice_assistant
The TEN Agent is a powerful tool that helps you create and manage AI agents with various capabilities like real-time vision, screen detection, and integration with services like Google Gemini Multimodal Live API, Weather Check, and Web Search. To use it, you need to set up your environment with Docker, Node.js, and specific API keys. You can follow simple steps to configure and start your agent locally. The benefits include easy integration of advanced AI features, a supportive community through Discord and GitHub discussions, and the ability to customize and extend your agents with ready-to-use extensions. This makes it easier to develop and deploy sophisticated AI applications quickly.
https://github.com/TEN-framework/TEN-Agent
The TEN Agent is a powerful tool that helps you create and manage AI agents with various capabilities like real-time vision, screen detection, and integration with services like Google Gemini Multimodal Live API, Weather Check, and Web Search. To use it, you need to set up your environment with Docker, Node.js, and specific API keys. You can follow simple steps to configure and start your agent locally. The benefits include easy integration of advanced AI features, a supportive community through Discord and GitHub discussions, and the ability to customize and extend your agents with ready-to-use extensions. This makes it easier to develop and deploy sophisticated AI applications quickly.
https://github.com/TEN-framework/TEN-Agent
GitHub
GitHub - TEN-framework/ten-framework: Open-source framework for conversational voice AI agents
Open-source framework for conversational voice AI agents - TEN-framework/ten-framework
#python #foundation_models #vision_language_model #vision_language_pretraining
DeepSeek-VL is a powerful, open-source Vision-Language (VL) Model that helps you understand and interact with both images and text. It can process various types of data like logical diagrams, web pages, scientific literature, and natural images. You can use it for different applications, such as describing images, recognizing formulas, and more. The model is available in different sizes and variants, making it flexible for various needs. You can download and use the models freely, even for commercial purposes, under the specified licenses. This tool makes it easier to integrate vision and language understanding into your projects.
https://github.com/deepseek-ai/DeepSeek-VL
DeepSeek-VL is a powerful, open-source Vision-Language (VL) Model that helps you understand and interact with both images and text. It can process various types of data like logical diagrams, web pages, scientific literature, and natural images. You can use it for different applications, such as describing images, recognizing formulas, and more. The model is available in different sizes and variants, making it flexible for various needs. You can download and use the models freely, even for commercial purposes, under the specified licenses. This tool makes it easier to integrate vision and language understanding into your projects.
https://github.com/deepseek-ai/DeepSeek-VL
GitHub
GitHub - deepseek-ai/DeepSeek-VL: DeepSeek-VL: Towards Real-World Vision-Language Understanding
DeepSeek-VL: Towards Real-World Vision-Language Understanding - deepseek-ai/DeepSeek-VL
👍1