#python #deep_learning #deep_learning_library #image_captioning #multimodal_datasets #multimodal_deep_learning #salesforce #vision_and_language #vision_framework #vision_language_pretraining #vision_language_transformer #visual_question_anwsering
https://github.com/salesforce/LAVIS
https://github.com/salesforce/LAVIS
GitHub
GitHub - salesforce/LAVIS: LAVIS - A One-stop Library for Language-Vision Intelligence
LAVIS - A One-stop Library for Language-Vision Intelligence - salesforce/LAVIS
#python #chatgpt #clip #deep_learning #gpt #hacktoberfest #hnsw #information_retrieval #knn #large_language_models #machine_learning #machinelearning #multi_modal #natural_language_processing #search_engine #semantic_search #tensor_search #transformers #vector_search #vision_language #visual_search
https://github.com/marqo-ai/marqo
https://github.com/marqo-ai/marqo
GitHub
GitHub - marqo-ai/marqo: Unified embedding generation and search engine. Also available on cloud - cloud.marqo.ai
Unified embedding generation and search engine. Also available on cloud - cloud.marqo.ai - marqo-ai/marqo
#cplusplus #artificial_intelligence #computer_vision #document #document_analysis #document_intelligence #document_recognition #document_understanding #documentai #end_to_end_ocr #multimodal #multimodal_deep_learning #ocr #scene_text_detection #scene_text_detection_recognition #scene_text_recognition #text_detection #text_recognition #vision_language #vision_language_model #vision_language_transformer
https://github.com/AlibabaResearch/AdvancedLiterateMachinery
https://github.com/AlibabaResearch/AdvancedLiterateMachinery
GitHub
GitHub - AlibabaResearch/AdvancedLiterateMachinery: A collection of original, innovative ideas and algorithms towards Advanced…
A collection of original, innovative ideas and algorithms towards Advanced Literate Machinery. This project is maintained by the OCR Team in the Language Technology Lab, Tongyi Lab, Alibaba Group. ...
👍1
#python #chinese #clip #computer_vision #contrastive_loss #coreml_models #deep_learning #image_text_retrieval #multi_modal #multi_modal_learning #nlp #pretrained_models #pytorch #transformers #vision_and_language_pre_training #vision_language
This project is about a Chinese version of the CLIP (Contrastive Language-Image Pretraining) model, trained on a large dataset of Chinese text and images. Here’s what you need to know This model helps you quickly perform tasks like calculating text and image features, cross-modal retrieval (finding images based on text or vice versa), and zero-shot image classification (classifying images without any labeled examples).
- **Ease of Use** The model has been tested on various datasets and shows strong performance in zero-shot image classification and cross-modal retrieval tasks.
- **Resources**: The project includes pre-trained models, training and testing codes, and detailed tutorials on how to use the model for different tasks.
Overall, this project makes it easy to work with Chinese text and images using advanced AI techniques, saving you time and effort.
https://github.com/OFA-Sys/Chinese-CLIP
This project is about a Chinese version of the CLIP (Contrastive Language-Image Pretraining) model, trained on a large dataset of Chinese text and images. Here’s what you need to know This model helps you quickly perform tasks like calculating text and image features, cross-modal retrieval (finding images based on text or vice versa), and zero-shot image classification (classifying images without any labeled examples).
- **Ease of Use** The model has been tested on various datasets and shows strong performance in zero-shot image classification and cross-modal retrieval tasks.
- **Resources**: The project includes pre-trained models, training and testing codes, and detailed tutorials on how to use the model for different tasks.
Overall, this project makes it easy to work with Chinese text and images using advanced AI techniques, saving you time and effort.
https://github.com/OFA-Sys/Chinese-CLIP
GitHub
GitHub - OFA-Sys/Chinese-CLIP: Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation. - OFA-Sys/Chinese-CLIP
#python #foundation_models #vision_language_model #vision_language_pretraining
DeepSeek-VL is a powerful, open-source Vision-Language (VL) Model that helps you understand and interact with both images and text. It can process various types of data like logical diagrams, web pages, scientific literature, and natural images. You can use it for different applications, such as describing images, recognizing formulas, and more. The model is available in different sizes and variants, making it flexible for various needs. You can download and use the models freely, even for commercial purposes, under the specified licenses. This tool makes it easier to integrate vision and language understanding into your projects.
https://github.com/deepseek-ai/DeepSeek-VL
DeepSeek-VL is a powerful, open-source Vision-Language (VL) Model that helps you understand and interact with both images and text. It can process various types of data like logical diagrams, web pages, scientific literature, and natural images. You can use it for different applications, such as describing images, recognizing formulas, and more. The model is available in different sizes and variants, making it flexible for various needs. You can download and use the models freely, even for commercial purposes, under the specified licenses. This tool makes it easier to integrate vision and language understanding into your projects.
https://github.com/deepseek-ai/DeepSeek-VL
GitHub
GitHub - deepseek-ai/DeepSeek-VL: DeepSeek-VL: Towards Real-World Vision-Language Understanding
DeepSeek-VL: Towards Real-World Vision-Language Understanding - deepseek-ai/DeepSeek-VL
👍1
#python #any_to_any #foundation_models #llm #multimodal #unified_model #vision_language_pretraining
The Janus-Series models, including Janus, Janus-Pro, and JanusFlow, are advanced AI tools that combine multimodal understanding and generation capabilities. These models can process both text and images, allowing for tasks like answering questions based on images and generating images from text descriptions. Janus-Pro is an improved version with better performance due to optimized training strategies and larger model sizes. JanusFlow integrates autoregressive language models with rectified flow for efficient image generation. The benefit to the user is the ability to perform complex multimodal tasks with high accuracy and flexibility, making these models useful for a wide range of applications in research and industry.
https://github.com/deepseek-ai/Janus
The Janus-Series models, including Janus, Janus-Pro, and JanusFlow, are advanced AI tools that combine multimodal understanding and generation capabilities. These models can process both text and images, allowing for tasks like answering questions based on images and generating images from text descriptions. Janus-Pro is an improved version with better performance due to optimized training strategies and larger model sizes. JanusFlow integrates autoregressive language models with rectified flow for efficient image generation. The benefit to the user is the ability to perform complex multimodal tasks with high accuracy and flexibility, making these models useful for a wide range of applications in research and industry.
https://github.com/deepseek-ai/Janus
GitHub
GitHub - deepseek-ai/Janus: Janus-Series: Unified Multimodal Understanding and Generation Models
Janus-Series: Unified Multimodal Understanding and Generation Models - deepseek-ai/Janus
❤1
#python #agent #context_engineering #electron #embedding_models #memory #proactive_ai #python #python3 #rag #react #vector_database #vision_language_model
MineContext is a special AI tool that helps you work more efficiently. It collects information from your computer screen and other sources, then uses this data to give you useful insights, summaries, and reminders. This helps you stay organized and focused on important tasks. MineContext is also very private because it stores all your data on your local device, not in the cloud. It's like having a personal assistant that helps you manage your digital life better.
https://github.com/volcengine/MineContext
MineContext is a special AI tool that helps you work more efficiently. It collects information from your computer screen and other sources, then uses this data to give you useful insights, summaries, and reminders. This helps you stay organized and focused on important tasks. MineContext is also very private because it stores all your data on your local device, not in the cloud. It's like having a personal assistant that helps you manage your digital life better.
https://github.com/volcengine/MineContext
GitHub
GitHub - volcengine/MineContext: MineContext is your proactive context-aware AI partner(Context-Engineering+ChatGPT Pulse)
MineContext is your proactive context-aware AI partner(Context-Engineering+ChatGPT Pulse) - volcengine/MineContext