薅薅乐 - JD 测试脚本

Overview

薅薅乐

安裝

使用docker

  • docker一键安装: docker run -d --name jd classmatelin/hhl:latest.

使用

  • 进入容器: docker exec -it jd bash

  • 获取JD_COOKIES: python get_jd_cookies.py, 扫描登录成功后控制台会打印JD_COOKIES.

  • vim /scripts/conf/config.yaml, 填入上一步获取的JD_COOKIES。

    debug: true
    
    # JD_COOKIES配置, 一行一个, -符号是必须的。
    jd_cookies: 
      - pt_pin=jd_78b;pt_key=AAJgyqEMOsFQr5a0ucVzJepxU;
    
    # 此处省略更多配置
  • 配置好JD_COOKIES, 随便运行一个脚本检查配置, 如: python update_share_code.py.

  • 内置定时任务, 配好JD_COOKIES自动会执行脚本。

  • 消息通知支持TG机器人, 需要在config.yaml文件中填写 tg_bot_token和tg_user_id配置.

脚本列表

脚本名称 脚本描述 完成进度
jd_big_winner.py 京东极速版->大赢家翻翻乐 100%
update_share_code.py 更新助力码 100%
jd_bean_change.py 资产变动通知 100%
get_jd_cookies.py 获取京东的COOKIES 100%
jd_cute_pet.py 京东APP->东东萌宠 100%
jd_factory.py 京东APP->东东工厂 100%
jd_farm.py 京东APP->东东农场 100%
jd_lucky_turntable.py 幸运大转盘 100%
jd_pet_dog 京东APP-> 宠汪汪 10%
jd_planting_bean.py 京东APP->种豆得豆 100%
jd_ranking_list 京东APP->排行榜 100%
jd_shark_bean 京东APP->摇金豆 100%
jd_sign_collection 京东签到合集 100%
jr_daily_task_goose.py 京东金融->天天提鹅 100%
jr_pet_pig 京东金融->养猪猪 100%
jr_money_tree 京东金融->摇钱树 100%
jd_bean_lottery.py 京东APP->签到领京豆->摇金豆->京豆夺宝 100%
jd_earn_bean.py 微信小程序-赚金豆 100%
jd_cash.py 京东APP-领现金 100%
jd_burning_summer.py 京东APP-燃动夏季 100%
jd_wishing_pool.py 京东APP-京东众筹-许愿池 100%
jd_good_morning.py 京东APP-领金豆->早起福利 100%
jd_gold_creator.py 京东APP-京东排行榜->金榜创造营 100%
jd_beauty.py 京东APP->美丽研究院 0%
jx_factory.py 京喜APP->京喜工厂 0%
jx_farm.py 京喜APP->京喜农场 0%
jd_esports_manager.py 京东APP->电竞经理 0%

配置说明

  • 请查看/scripts/conf/config.yaml中的配置项。

特别声明:

  • 本仓库发布的jd_scripts项目中涉及的任何解锁和解密分析脚本,仅用于测试和学习研究,禁止用于商业用途,不能保证其合法性,准确性,完整性和有效性,请根据情况自行判断.

  • 本项目内所有资源文件,禁止任何公众号、自媒体进行任何形式的转载、发布。

  • ClassmateLin对任何脚本问题概不负责,包括但不限于由任何脚本错误导致的任何损失或损害.

  • 间接使用脚本的任何用户,包括但不限于建立VPS或在某些行为违反国家/地区法律或相关法规的情况下进行传播, ClassmateLin对于由此引起的任何隐私泄漏或其他后果概不负责.

  • 请勿将jd_scripts项目的任何内容用于商业或非法目的,否则后果自负.

  • 如果任何单位或个人认为该项目的脚本可能涉嫌侵犯其权利,则应及时通知并提供身份证明,所有权证明,本人将在收到认证文件后删除相关脚本.

  • 任何以任何方式查看此项目的人或直接或间接使用该jd_scripts项目的任何脚本的使用者都应仔细阅读此声明。ClassmateLin保留随时更改或补充此免责声明的权利。一旦使用并复制了任何相关脚本或jd_scripts项目的规则,则视为您已接受此免责声明.

您必须在下载后的24小时内从计算机或手机中完全删除以上内容.
您使用或者复制了本仓库且本人制作的任何脚本,则视为已接受此声明,请仔细阅读

Owner
ClassmateLin
ClassmateLin
A powerful annex BUBT, BUBT Soft, and BUBT website scraping script.

Annex Bubt Scraping Script I think this is the first public repository that provides free annex-BUBT, BUBT-Soft, and BUBT website scraping API script

Md Imam Hossain 4 Dec 03, 2022
Scrapes the Sun Life of Canada Philippines web site for historical prices of their investment funds and then saves them as CSV files.

slocpi-scraper Sun Life of Canada Philippines Inc Investment Funds Scraper Install dependencies pip install -r requirements.txt Usage General format:

Daryl Yu 2 Jan 07, 2022
Deep Web Miner Python | Spyder Crawler

Webcrawler written in Python. This crawler does dig in till the 3 level of inside addressed and mine the respective data accordingly

Karan Arora 17 Jan 24, 2022
An Web Scraping API for MDL(My Drama List) for Python.

PyMDL An API for MyDramaList(MDL) based on webscraping for python. Description An API for MDL to make your life easier in retriving and working on dat

6 Dec 10, 2022
Web-Scrapper using Python and Flask

Web-Scrapper "[초급]Python으로 웹 스크래퍼 만들기" 코스 -NomadCoders 기초적인 Python 문법강의부터 시작하여 웹사이트의 html파일에서 원하는 내용을 Scrapping해서 출력, csv 파일로 저장, flask를 이용한 간단한 웹페이지

윤성도 1 Nov 10, 2021
An arxiv spider

An Arxiv Spider 做为一个cser,杰出男孩深知内核对连接到计算机上的硬件设备进行管理的高效方式是中断而不是轮询。每当小伙伴发来一篇刚挂在arxiv上的”热乎“好文章时,杰出男孩都会感叹道:”师兄这是每天都挂在arxiv上呀,跑的好快~“。于是杰出男孩找了找 github,借鉴了一下其

Jie Liu 11 Sep 09, 2022
The first public repository that provides free BUBT website scraping API script on Github.

BUBT WEBSITE SCRAPPING SCRIPT I think this is the first public repository that provides free BUBT website scraping API script on github. When I was do

Md Imam Hossain 3 Feb 10, 2022
Scraping web pages to get data

Scraping Data Get public data and save in database This is project use Python How to run a project 1 - Clone the repository 2 - Install beautifulsoup4

Soccer Project 2 Nov 01, 2021
Crawl BookCorpus

These are scripts to reproduce BookCorpus by yourself.

Sosuke Kobayashi 590 Jan 03, 2023
This is a webscraper for a specific website

This is a webscraper for a specific website. It is tuned to extract the headlines of that website. With some little adjustments the webscraper is able to extract any part of the website.

Rahul Siyanwal 1 Dec 13, 2021
Python script who crawl first shodan page and check DBLTEK vulnerability

🐛 MASS DBLTEK EXPLOIT CHECKER USING SHODAN 🕸 Python script who crawl first shodan page and check DBLTEK vulnerability

Divin 4 Jan 09, 2022
Web Crawlers for Data Labelling of Malicious Domain Detection & IP Reputation Evaluation

Web Crawlers for Data Labelling of Malicious Domain Detection & IP Reputation Evaluation This repository provides two web crawlers to label domain nam

1 Nov 05, 2021
Scrapy uses Request and Response objects for crawling web sites.

Requests and Responses¶ Scrapy uses Request and Response objects for crawling web sites. Typically, Request objects are generated in the spiders and p

Md Rashidul Islam 1 Nov 03, 2021
Goblyn is a Python tool focused to enumeration and capture of website files metadata.

Goblyn Metadata Enumeration What's Goblyn? Goblyn is a tool focused to enumeration and capture of website files metadata. How it works? Goblyn will se

Gustavo 46 Nov 22, 2022
CreamySoup - a helper script for automated SourceMod plugin updates management.

CreamySoup/"Creamy SourceMod Updater" (or just soup for short), a helper script for automated SourceMod plugin updates management.

3 Jan 03, 2022
Scrapes proxies and saves them to a text file

Proxy Scraper Scrapes proxies from https://proxyscrape.com and saves them to a file. Also has a customizable theme system Made by nell and Lamp

nell 2 Dec 22, 2021
Proxy scraper. Format: IP | PORT | COUNTRY | TYPE

proxy scraper 🔎 Installation: git clone https://github.com/ebankoff/proxy_scraper Required pip libraries (pip install library name): lxml beautifulso

Eban'ko 19 Dec 07, 2022
LSpider 一个为被动扫描器定制的前端爬虫

LSpider LSpider - 一个为被动扫描器定制的前端爬虫 什么是LSpider? 一款为被动扫描器而生的前端爬虫~ 由Chrome Headless、LSpider主控、Mysql数据库、RabbitMQ、被动扫描器5部分组合而成。

Knownsec, Inc. 321 Dec 12, 2022
京东秒杀商品抢购Python脚本

Jd_Seckill 非常感谢原作者 https://github.com/zhou-xiaojun/jd_mask 提供的代码 也非常感谢 https://github.com/wlwwu/jd_maotai 进行的优化 主要功能 登陆京东商城(www.jd.com) cookies登录 (需要自

Andy Zou 1.5k Jan 03, 2023
Python web scrapper

Website scrapper Web scrapping project in Python. Created for learning purposes. Start Install python Update configuration with websites Launch script

Nogueira Vitor 1 Dec 19, 2021