A list of Python Bots used to extract data from several websites. Data extraction is for products on e-commerce (ecommerce) websites. Data fetched is such as the product images, title, price range, price, type of data etc. List of Website: https://www.qoovee.com/en/ https://merxu.com/en/ https://daraz.com/ https://www.nihaojewelry.com/ https://www.ecplaza.net/mask--product https://www.exportportal.com/ https://www.mallory.com/ https://www.townandcountryhardware.com/ https://www.like123.com/en/ https://www.ishopping.pk/ http://global.gmarket.co.kr/ https://shoptheglobe.co/ https://www.rannthai.com/ https://www.industrybuying.com/ https://www.ralali.com/ https://globaltradeplaza.com/ https://www.wholesalebox.in/ https://madeinindonesia.com/ https://dubaiyellowpagesonline.com/ https://www.qualitymill.com/ https://www.grainger.com/ https://www.abraa.com/ Python Libraries: Selenium, Beautiful Soup, Pandas, Scrapy, Requests, Urllib & Credentials etc. ___________________________________ This Project was based on a Python Internship Summer 2021 Dated: June-July 2021
A list of Python Bots used to extract data from several websites
Overview
Danbooru scraper with python
Danbooru Version: 0.0.1 License under: MIT License Dependencies Python: = 3.9.7 beautifulsoup4 cloudscraper Example of use Danbooru from danbooru imp
a small library for extracting rich content from urls
A small library for extracting rich content from urls. what does it do? micawber supplies a few methods for retrieving rich metadata about a variety o
Scrape data on SpaceX: Capsules, Rockets, Cores, Roadsters, SpaceX Info
SpaceX Sofware I developed software to scrape data on SpaceX: Capsules, Rockets, Cores, Roadsters, SpaceX Info to use the software you need Python a
PaperRobot: a paper crawler that can quickly download numerous papers, facilitating paper studying and management
PaperRobot PaperRobot 是一个论文抓取工具,可以快速批量下载大量论文,方便后期进行持续的论文管理与学习。 PaperRobot通过多个接口抓取论文,目前抓取成功率维持在90%以上。通过配置Config文件,可以抓取任意计算机领域相关会议的论文。 Installation Down
Web scraper build using python.
Web Scraper This project is made in pyhthon. It took some info. from website list then add them into data.json file. The dependencies used are: reques
用python爬取江苏几大高校的就业网站,并提供3种方式通知给用户,分别是通过微信发送、命令行直接输出、windows气泡通知。
crawler_for_university 用python爬取江苏几大高校的就业网站,并提供3种方式通知给用户,分别是通过微信发送、命令行直接输出、windows气泡通知。 环境依赖 wxpy,requests,bs4等库 功能描述 该项目基于python,通过爬虫爬各高校的就业信息网,爬取招聘信
Parse feeds in Python
feedparser - Parse Atom and RSS feeds in Python. Copyright 2010-2020 Kurt McKee 1.5k Dec 30, 2022
Searching info from Google using Python Scrapy
Python-Search-Engine-Scrapy || Python-爬虫-索引/利用爬虫获取谷歌信息**/ Searching info from Google using Python Scrapy /* 利用 PYTHON 爬虫获取天气信息,以及城市信息和资料**/ translatio
Automated data scraper for Thailand COVID-19 data
The Researcher COVID data Automated data scraper for Thailand COVID-19 data Accessing the Data 1st Dose Provincial Vaccination Data 2nd Dose Provincia
一款利用Python来自动获取QQ音乐上某个歌手所有歌曲歌词的爬虫软件
QQ音乐歌词爬虫 一款利用Python来自动获取QQ音乐上某个歌手所有歌曲歌词的爬虫软件,默认去除了所有演唱会(Live)版本的歌曲。 使用方法 直接运行python run.py即可,然后输入你想获取的歌手名字,然后静静等待片刻。 output目录下保存生成的歌词和歌名文件。以周杰伦为例,会生成两
Web scrapping tool written in python3, using regex, to get CVEs, Source and URLs.
searchcve Web scrapping tool written in python3, using regex, to get CVEs, Source and URLs. Generates a CSV file in the current directory. Uses the NI
FilmMikirAPI - A simple rest-api which is used for scrapping on the Kincir website using the Python and Flask package
FilmMikirAPI - A simple rest-api which is used for scrapping on the Kincir website using the Python and Flask package
UdemyBot - A Simple Udemy Free Courses Scrapper
UdemyBot - A Simple Udemy Free Courses Scrapper
A Spider for BiliBili comments with a simple API server.
BiliComment A spider for BiliBili comment. Spider Usage Put config.json into config directory, and then python . ./config/config.json. A example confi
A high-level distributed crawling framework.
Cola: high-level distributed crawling framework Overview Cola is a high-level distributed crawling framework, used to crawl pages and extract structur
A social networking service scraper in Python
snscrape snscrape is a scraper for social networking services (SNS). It scrapes things like user profiles, hashtags, or searches and returns the disco
Grab the changelog from releases on Github
release-notes-scraper This simple script can be used to grab the release notes for projects from github that do not keep a CHANGELOG, but publish thei
Create crawler get some new products with maximum discount in banimode website
crawler-banimode create crawler and get some new products with maximum discount in banimode website. این پروژه کوچک جهت یادگیری و کار با ابزار سلنیوم
Google Developer Profile Badge Scraper
Google Developer Profile Badge Scraper GDev Profile Badge Scraper is a Google Developer Profile Web Scraper which scrapes for specific badges in a use
Crawl BookCorpus
These are scripts to reproduce BookCorpus by yourself.