#go #cloud_native #cluster_management #distributed_cloud #edge_cloud #hybrid_cloud #kubernetes #multicloud #orchestration #out_of_the_box #public_cloud
https://github.com/clusternet/clusternet
https://github.com/clusternet/clusternet
GitHub
GitHub - clusternet/clusternet: [CNCF Sandbox Project] Managing your Kubernetes clusters (including public, private, edge, etc.)…
[CNCF Sandbox Project] Managing your Kubernetes clusters (including public, private, edge, etc.) as easily as visiting the Internet - clusternet/clusternet
#java #data #data_engineering #data_orchestration #data_orchestrator #data_pipeline #dataflow #elt #etl #kestra #orchestration #pipeline #scheduler #workflow #workflow_automation #workflow_engine
https://github.com/kestra-io/kestra
https://github.com/kestra-io/kestra
GitHub
GitHub - kestra-io/kestra: Orchestrate everything - from scripts to data, infra, AI, and business - as code, with UI and AI Copilot.…
Orchestrate everything - from scripts to data, infra, AI, and business - as code, with UI and AI Copilot. Simple. Fast. Scalable. - kestra-io/kestra
#java #cloud #cloudstack #iaas #infrastructure #kubernetes #kvm #libvirt #orchestration #python #virtual_machine #virtualization #vmware #vsphere #xcp_ng #xenserver
https://github.com/apache/cloudstack
https://github.com/apache/cloudstack
GitHub
GitHub - apache/cloudstack: Apache CloudStack is an opensource Infrastructure as a Service (IaaS) cloud computing platform
Apache CloudStack is an opensource Infrastructure as a Service (IaaS) cloud computing platform - apache/cloudstack
#typescript #agent #ai #anthropic #backend_as_a_service #chatbot #gemini #genai #gpt #gpt_4 #llama3 #llm #llmops #nextjs #openai #orchestration #python #rag #workflow #workflows
Dify is an open-source platform for developing AI applications, especially those using Large Language Models (LLMs). It offers a user-friendly interface to build and test AI workflows, integrate various LLMs, and manage models. Key features include a visual workflow builder, comprehensive model support (including GPT, Mistral, and more), a prompt IDE for crafting and testing prompts, RAG pipeline capabilities for document ingestion and retrieval, and agent capabilities with pre-built tools like Google Search and DALL·E.
Using Dify, you can quickly move from prototyping to production with features like observability to monitor application performance and backend-as-a-service for easy integration into your business logic. You can deploy Dify via their cloud service or self-host it in your environment. This makes it highly versatile and beneficial for developers looking to leverage AI efficiently in their projects.
https://github.com/langgenius/dify
Dify is an open-source platform for developing AI applications, especially those using Large Language Models (LLMs). It offers a user-friendly interface to build and test AI workflows, integrate various LLMs, and manage models. Key features include a visual workflow builder, comprehensive model support (including GPT, Mistral, and more), a prompt IDE for crafting and testing prompts, RAG pipeline capabilities for document ingestion and retrieval, and agent capabilities with pre-built tools like Google Search and DALL·E.
Using Dify, you can quickly move from prototyping to production with features like observability to monitor application performance and backend-as-a-service for easy integration into your business logic. You can deploy Dify via their cloud service or self-host it in your environment. This makes it highly versatile and beneficial for developers looking to leverage AI efficiently in their projects.
https://github.com/langgenius/dify
GitHub
GitHub - langgenius/dify: Production-ready platform for agentic workflow development.
Production-ready platform for agentic workflow development. - langgenius/dify
👍1
#go #docker #docker_compose #go #golang #orchestration
Docker Compose is a tool that helps you run multiple containers for your application easily. You define how your containers should be set up using a Compose file, and then you can start your entire application with just one command: `docker compose up`. This makes it simple to manage and run complex applications. You can get Docker Compose through Docker Desktop on Windows and macOS, or download it manually for Linux. Using Docker Compose saves time and effort by automating the setup and execution of your multi-container applications.
https://github.com/docker/compose
Docker Compose is a tool that helps you run multiple containers for your application easily. You define how your containers should be set up using a Compose file, and then you can start your entire application with just one command: `docker compose up`. This makes it simple to manage and run complex applications. You can get Docker Compose through Docker Desktop on Windows and macOS, or download it manually for Linux. Using Docker Compose saves time and effort by automating the setup and execution of your multi-container applications.
https://github.com/docker/compose
GitHub
GitHub - docker/compose: Define and run multi-container applications with Docker
Define and run multi-container applications with Docker - docker/compose
#java #airflow #azkaban #cloud_native #data_pipelines #job_scheduler #orchestration #powerful_data_pipelines #task_scheduler #workflow #workflow_orchestration #workflow_schedule
Apache DolphinScheduler is a powerful tool for managing data workflows. It makes it easy to create and manage complex tasks with a user-friendly interface and low-code options. You can deploy it in several ways, including standalone, cluster, Docker, and Kubernetes, making it flexible for different environments. It's highly reliable, scalable, and performs much faster than other platforms, supporting millions of tasks daily. The tool also offers features like versioning, state control of workflows, multi-tenancy support, and permission control. This helps you manage your data pipelines efficiently and reliably, saving time and effort.
https://github.com/apache/dolphinscheduler
Apache DolphinScheduler is a powerful tool for managing data workflows. It makes it easy to create and manage complex tasks with a user-friendly interface and low-code options. You can deploy it in several ways, including standalone, cluster, Docker, and Kubernetes, making it flexible for different environments. It's highly reliable, scalable, and performs much faster than other platforms, supporting millions of tasks daily. The tool also offers features like versioning, state control of workflows, multi-tenancy support, and permission control. This helps you manage your data pipelines efficiently and reliably, saving time and effort.
https://github.com/apache/dolphinscheduler
GitHub
GitHub - apache/dolphinscheduler: Apache DolphinScheduler is the modern data orchestration platform. Agile to create high performance…
Apache DolphinScheduler is the modern data orchestration platform. Agile to create high performance workflow with low-code - apache/dolphinscheduler
#python #analytics #dagster #data_engineering #data_integration #data_orchestrator #data_pipelines #data_science #etl #metadata #mlops #orchestration #python #scheduler #workflow #workflow_automation
Dagster is a tool that helps you manage and automate your data workflows. You can define your data assets, like tables or machine learning models, using Python functions. Dagster then runs these functions at the right time and keeps your data up-to-date. It offers features like integrated lineage and observability, making it easier to track and manage your data. This tool is useful for every stage of data development, from local testing to production, and it integrates well with other popular data tools. Using Dagster, you can build reusable components, spot data quality issues early, and scale your data pipelines efficiently. This makes your work more productive and helps maintain control over complex data systems.
https://github.com/dagster-io/dagster
Dagster is a tool that helps you manage and automate your data workflows. You can define your data assets, like tables or machine learning models, using Python functions. Dagster then runs these functions at the right time and keeps your data up-to-date. It offers features like integrated lineage and observability, making it easier to track and manage your data. This tool is useful for every stage of data development, from local testing to production, and it integrates well with other popular data tools. Using Dagster, you can build reusable components, spot data quality issues early, and scale your data pipelines efficiently. This makes your work more productive and helps maintain control over complex data systems.
https://github.com/dagster-io/dagster
GitHub
GitHub - dagster-io/dagster: An orchestration platform for the development, production, and observation of data assets.
An orchestration platform for the development, production, and observation of data assets. - dagster-io/dagster
👍1
#python #airflow #apache #apache_airflow #automation #dag #data_engineering #data_integration #data_orchestrator #data_pipelines #data_science #elt #etl #machine_learning #mlops #orchestration #python #scheduler #workflow #workflow_engine #workflow_orchestration
Apache Airflow is a tool that helps you manage and automate workflows. You can write your workflows as code, making them easier to maintain, version, test, and collaborate on. Airflow lets you schedule tasks and monitor their progress through a user-friendly interface. It supports dynamic pipeline generation, is highly extensible, and scalable, allowing you to define your own operators and executors.
Using Airflow benefits you by making your workflows more organized, efficient, and reliable. It simplifies the process of managing complex tasks and provides clear visualizations of your workflow's performance, helping you identify and troubleshoot issues quickly. This makes it easier to manage data processing and other automated tasks effectively.
https://github.com/apache/airflow
Apache Airflow is a tool that helps you manage and automate workflows. You can write your workflows as code, making them easier to maintain, version, test, and collaborate on. Airflow lets you schedule tasks and monitor their progress through a user-friendly interface. It supports dynamic pipeline generation, is highly extensible, and scalable, allowing you to define your own operators and executors.
Using Airflow benefits you by making your workflows more organized, efficient, and reliable. It simplifies the process of managing complex tasks and provides clear visualizations of your workflow's performance, helping you identify and troubleshoot issues quickly. This makes it easier to manage data processing and other automated tasks effectively.
https://github.com/apache/airflow
GitHub
GitHub - apache/airflow: Apache Airflow - A platform to programmatically author, schedule, and monitor workflows
Apache Airflow - A platform to programmatically author, schedule, and monitor workflows - apache/airflow
👍1
#python #automation #data #data_engineering #data_ops #data_science #infrastructure #ml_ops #observability #orchestration #pipeline #prefect #python #workflow #workflow_engine
Prefect is a tool that helps you automate and manage data workflows in Python. It makes it easy to turn your scripts into reliable and flexible workflows that can handle unexpected changes. With Prefect, you can schedule tasks, retry failed operations, and monitor your workflows. You can install it using `pip install -U prefect` and start creating workflows with just a few lines of code. This helps data teams work more efficiently, reduce errors, and save time. You can also use Prefect Cloud for more advanced features and support.
https://github.com/PrefectHQ/prefect
Prefect is a tool that helps you automate and manage data workflows in Python. It makes it easy to turn your scripts into reliable and flexible workflows that can handle unexpected changes. With Prefect, you can schedule tasks, retry failed operations, and monitor your workflows. You can install it using `pip install -U prefect` and start creating workflows with just a few lines of code. This helps data teams work more efficiently, reduce errors, and save time. You can also use Prefect Cloud for more advanced features and support.
https://github.com/PrefectHQ/prefect
GitHub
GitHub - PrefectHQ/prefect: Prefect is a workflow orchestration framework for building resilient data pipelines in Python.
Prefect is a workflow orchestration framework for building resilient data pipelines in Python. - PrefectHQ/prefect
#python #cloud_native #cncf #deep_learning #docker #fastapi #framework #generative_ai #grpc #jaeger #kubernetes #llmops #machine_learning #microservice #mlops #multimodal #neural_search #opentelemetry #orchestration #pipeline #prometheus
Jina-serve is a tool that helps you build and deploy AI services easily. It supports major machine learning frameworks and allows you to scale your services from local development to production quickly. You can use it to create AI services that communicate via gRPC, HTTP, and WebSockets. It has features like built-in Docker integration, one-click cloud deployment, and support for Kubernetes and Docker Compose, making it easy to manage and scale your AI applications. This makes it simpler for you to focus on the core logic of your AI projects without worrying about the technical details of deployment and scaling.
https://github.com/jina-ai/serve
Jina-serve is a tool that helps you build and deploy AI services easily. It supports major machine learning frameworks and allows you to scale your services from local development to production quickly. You can use it to create AI services that communicate via gRPC, HTTP, and WebSockets. It has features like built-in Docker integration, one-click cloud deployment, and support for Kubernetes and Docker Compose, making it easy to manage and scale your AI applications. This makes it simpler for you to focus on the core logic of your AI projects without worrying about the technical details of deployment and scaling.
https://github.com/jina-ai/serve
GitHub
GitHub - jina-ai/serve: ☁️ Build multimodal AI applications with cloud-native stack
☁️ Build multimodal AI applications with cloud-native stack - jina-ai/serve
#rust #beginner_friendly #featured #finance #hacktoberfest #high_performance #open_source #orchestration #payments #postgresql #redis #restful_api #rust #sdk #works_with_react
Hyperswitch is an open-source payments platform that helps businesses manage payments easily. It uses a single API to access various payment methods and features, making it simple to integrate different payment flows like cards, wallets, and bank transfers. The platform includes a backend for seamless payment processing, an SDK for unified payment experiences across web and mobile, and a Control Center for managing payments without coding. You can try Hyperswitch by setting it up locally or deploying it on cloud services like AWS, GCP, or Azure. This platform is designed to be reliable, secure, and customizable, allowing businesses to own and tailor their payment stack according to their needs.
https://github.com/juspay/hyperswitch
Hyperswitch is an open-source payments platform that helps businesses manage payments easily. It uses a single API to access various payment methods and features, making it simple to integrate different payment flows like cards, wallets, and bank transfers. The platform includes a backend for seamless payment processing, an SDK for unified payment experiences across web and mobile, and a Control Center for managing payments without coding. You can try Hyperswitch by setting it up locally or deploying it on cloud services like AWS, GCP, or Azure. This platform is designed to be reliable, secure, and customizable, allowing businesses to own and tailor their payment stack according to their needs.
https://github.com/juspay/hyperswitch
GitHub
GitHub - juspay/hyperswitch: An open source payments switch written in Rust to make payments fast, reliable and affordable
An open source payments switch written in Rust to make payments fast, reliable and affordable - juspay/hyperswitch