GitHub Trends
10.1K subscribers
15.3K links
See what the GitHub community is most excited about today.

A bot automatically fetches new repositories from https://github.com/trending and sends them to the channel.

Author and maintainer: https://github.com/katursis
Download Telegram
#python #api #async #crawler #douyin #douyin_api #douyin_scraper #douyin_tiktok_api #douyin_tiktok_download #fastapi #no_watermark #online_parsing #python #pywebio #scraper #spider #tiktok #tiktok_api #tiktok_scraper #tiktok_signature #web_scraping

This tool, called "Douyin_TikTok_Download_API," allows you to easily download videos and data from Douyin (Chinese TikTok), TikTok, and Bilibili without watermarks. Here’s how it benefits you The tool is ready to use out of the box and supports API calls, online batch parsing, and downloading.
- **Versatile** Built with PyWebIO, FastAPI, and HTTPX for fast asynchronous data scraping.
- **Customizable** Provides comprehensive API documentation for easy integration into your projects.
- **iOS Integration**: Supports iOS shortcuts for quick video downloads without watermarks.

Overall, this tool simplifies the process of downloading content from these platforms while offering flexibility and high performance.

https://github.com/Evil0ctal/Douyin_TikTok_Download_API
1
#javascript #batch_processing #batch_script #code_free #crawler #data_collection #frontend #gui #html #input_parameters #layman #parameters #robotics #rpa #scraper #spider #visual #visualization #visualprogramming #web #www

EasySpider is a free, code-free web crawler software that helps you collect data from websites easily. You can use it without writing any code, just by selecting the content you want to operate on a web page and following the prompts. It can also be run from the command line, making it easy to integrate into other systems. This software is very user-friendly and allows you to collect data from websites quickly, even if you don't know how to code. It also supports various proxy services and captcha solutions, making data collection more efficient. Using EasySpider can save you a lot of time and effort in collecting web data.

https://github.com/NaiboWang/EasySpider
#typescript #api #automation #browser #browser_automation #low_code #maxun #no_code #playwright #robotic_process_automation #rpa #scraper #self_hosted #spreadsheet #web_automation #web_scraper #web_scraping #website_to_api

Maxun is a simple way to extract data from websites without coding. You can train a robot in just 2 minutes to automatically scrape web data for you. Here’s how it helps: You can create robots that capture lists, text, or screenshots from websites. These robots can run on their own, even handling pagination and scrolling, and can be scheduled to run at specific times. Maxun also integrates with tools like Google Sheets and offers a cloud version that manages anti-bot detection and proxy rotation, making it easy to extract data at scale. This saves you time and effort, allowing you to focus on other tasks while the robots do the work for you.

https://github.com/getmaxun/maxun
#python #agent #application_resume #artificial_intelligence #automate #automation #bot #chatgpt #chrome #gpt #human_resources #job #jobs #jobsearch #jobseeker #opeai #python #resume #scraper #scraping #selenium

AIHawk is an AI-powered job search assistant that helps you automate job applications, get personalized recommendations, and land your dream job faster. Here’s how it benefits you AIHawk automates the tedious process of filling out job application forms, saving you time and effort.
- **Personalized Recommendations** It generates customized resumes and cover letters to match the requirements of each job you apply for.
- **Community Support** The project is open-source and actively maintained by a team of contributors, ensuring it keeps getting better with user input.

Overall, AIHawk streamlines the job search process, making it more efficient and increasing your chances of finding the right job quickly.

https://github.com/feder-cr/Jobs_Applier_AI_Agent
2
#typescript #ai #ai_scraping #crawler #data #html_to_markdown #llm #markdown #rag #scraper #scraping #web_crawler #webscraping

Firecrawl is a tool that helps you get clean data from any website. Here’s how it benefits you Firecrawl can scrape and crawl websites, converting the content into formats like markdown, structured data, or HTML, making it ready for use in AI applications.
- **Advanced Capabilities** You can customize the scraping process by excluding certain tags, crawling behind authentication walls, and setting the maximum crawl depth.
- **Batch Processing** Firecrawl integrates with various SDKs and frameworks like Python, Node, Go, Rust, and more, making it easy to use in different projects.
- **Cloud and Self-Host Options**: You can use the hosted version or self-host it, depending on your needs.

Overall, Firecrawl simplifies the process of extracting data from websites, saving you time and effort.

https://github.com/mendableai/firecrawl
3
#typescript #ai #artificial_intelligence #browser #browser_automation #gpt #gpt_4 #langchain #llama #llm #openai #playwright #puppeteer #scraper

LLM Scraper is a tool that helps you get structured data from any webpage using large language models (LLMs). It supports different AI providers like OpenAI and Ollama, and it uses the Playwright framework to work with web pages. You can define what data you want to extract using schemas, which makes sure everything is organized correctly. This tool also allows you to generate code automatically for scraping tasks, making it easier to reuse scripts. The benefit is that you can easily collect data from websites in a structured way, which is helpful for projects that need specific information from the internet.

https://github.com/mishushakov/llm-scraper