✂️🕷️ Spider-Cut is a Network Mapper Framework (NMAP Framework)

Overview

Build with Python Latest Release Last Comitment Python Version


Spider-Cut is a Network Mapper Framework (NMAP Framework)

Installation    |    Usage    |    Creators    |    Donate

Installation

# Kali Linux | WSL

# clone the repo
$ git clone https://github.com/XFORWORKS/SpiderCut

# change the working directory to SpiderCut
$ cd SpiderCut
  
# install the requirements  
$ python -m pip install -r requirements.txt

# run the bigger installation process
$ python setup.py

Usage

# Kali Linux | WSL

$ python spidercut.py --help
usage: spidercut.py [-h] [-v] [-s] [-m] [-rH] [-sn] [-rh] [-eX]
                              [-ex] [-a]

Spider-Cut is a Network Mapper Framework (NMAP Framework)

optional arguments:
  -h, --help       show this help message and exit
  -v, --version    shows the current Version of Spider-Cut
  -s, --single     scan a single Target
  -m, --multiple   scan multiple Targets
  -rH, --rhost     scan a range of Hosts
  -sn, --subnet    scan an entire Subnet
  -rh, --ranhost   scan random Hosts
  -eX, --exclude   Excluding Targets from a scan
  -ex, --excllist  Excluding Targets using a list
  -a, --agscan     perform an Aggressive scan

Created By XFORWORKS (Retr0 & Ctript0)

To run the script:

# Kali Linux | WSL

$ python spidercut.py

Creators

The Script is designed by Cript0 (like the banners, the colors, etc...)

Everything other is from Retr0 (like the commands, the framework idea, etc...)

Donate

PayPal

Releases

Latest Release

You might also like...
Distributed Crawler Management Framework Based on Scrapy, Scrapyd, Django and Vue.js

Gerapy Distributed Crawler Management Framework Based on Scrapy, Scrapyd, Scrapyd-Client, Scrapyd-API, Django and Vue.js. Documentation Documentation

Async Python 3.6+ web scraping micro-framework based on asyncio
Async Python 3.6+ web scraping micro-framework based on asyncio

Ruia 🕸️ Async Python 3.6+ web scraping micro-framework based on asyncio. ⚡ Write less, run faster. Overview Ruia is an async web scraping micro-frame

A high-level distributed crawling framework.

Cola: high-level distributed crawling framework Overview Cola is a high-level distributed crawling framework, used to crawl pages and extract structur

Transistor, a Python web scraping framework for intelligent use cases.
Transistor, a Python web scraping framework for intelligent use cases.

Web data collection and storage for intelligent use cases. transistor About The web is full of data. Transistor is a web scraping framework for collec

Web crawling framework  based on asyncio.
Web crawling framework based on asyncio.

Web crawling framework for everyone. Written with asyncio, uvloop and aiohttp. Requirements Python3.5+ Installation pip install gain pip install uvloo

PyQuery-based scraping micro-framework.

demiurge PyQuery-based scraping micro-framework. Supports Python 2.x and 3.x. Documentation: http://demiurge.readthedocs.org Installing demiurge $ pip

Crawler do site Fundamentus.com com o uso do framework scrapy, tanto da aba detalhada como a de resumo.

Crawler do site Fundamentus.com com o uso do framework scrapy, tanto da aba detalhada como a de resumo. (Todas as infomações)

A simple django-rest-framework api using web scraping

Apicell You can use this api to search in google, bing, pypi and subscene and get results Method : POST Parameter : query Example import request url =

Python framework to scrape Pastebin pastes and analyze them
Python framework to scrape Pastebin pastes and analyze them

pastepwn - Paste-Scraping Python Framework Pastebin is a very helpful tool to store or rather share ascii encoded data online. In the world of OSINT,

Releases(1.0)
  • 1.0(Dec 25, 2021)

    Build with Python Last Comitment Python Version


    Spider-Cut is a Network Mapper Framework (NMAP Framework)

    Installation

    # Kali Linux | WSL
    
    # clone the repo
    $ git clone https://github.com/XFORWORKS/SpiderCut
    
    # change the working directory to SpiderCut
    $ cd SpiderCut
      
    # install the requirements  
    $ python -m pip install -r requirements.txt
    
    # run the bigger installation process
    $ python setup.py
    

    Usage

    # Kali Linux | WSL
    
    $ python spidercut.py --help
    usage: spidercut.py [-h] [-v] [-s] [-m] [-rH] [-sn] [-rh] [-eX]
                                  [-ex] [-a]
    
    Spider-Cut is a Network Mapper Framework (NMAP Framework)
    
    optional arguments:
      -h, --help       show this help message and exit
      -v, --version    shows the current Version of Spider-Cut
      -s, --single     scan a single Target
      -m, --multiple   scan multiple Targets
      -rH, --rhost     scan a range of Hosts
      -sn, --subnet    scan an entire Subnet
      -rh, --ranhost   scan random Hosts
      -eX, --exclude   Excluding Targets from a scan
      -ex, --excllist  Excluding Targets using a list
      -a, --agscan     perform an Aggressive scan
    
    Created By XFORWORKS (Retr0 & Cript0)
    

    To run the script:

    # Kali Linux | WSL
    
    $ python spidercut.py
    

    Creators

    The Script is designed by Cript0 (like the banners, the colors, etc...)

    Everything other is from Retr0 (like the commands, the framework idea, etc...)

    Donate

    PayPal

    Source code(tar.gz)
    Source code(zip)
    SpiderCut.rar(171.16 KB)
    SpiderCut.zip(226.92 KB)
Owner
XforWorks
Nothing to see here...
XforWorks
Twitter Scraper

Twitter's API is annoying to work with, and has lots of limitations — luckily their frontend (JavaScript) has it's own API, which I reverse–engineered. No API rate limits. No restrictions. Extremely

Tayyab Kharl 45 Dec 30, 2022
UdemyBot - A Simple Udemy Free Courses Scrapper

UdemyBot - A Simple Udemy Free Courses Scrapper

Gautam Kumar 112 Nov 12, 2022
用python爬取江苏几大高校的就业网站,并提供3种方式通知给用户,分别是通过微信发送、命令行直接输出、windows气泡通知。

crawler_for_university 用python爬取江苏几大高校的就业网站,并提供3种方式通知给用户,分别是通过微信发送、命令行直接输出、windows气泡通知。 环境依赖 wxpy,requests,bs4等库 功能描述 该项目基于python,通过爬虫爬各高校的就业信息网,爬取招聘信

8 Aug 16, 2021
A high-level distributed crawling framework.

Cola: high-level distributed crawling framework Overview Cola is a high-level distributed crawling framework, used to crawl pages and extract structur

Xuye (Chris) Qin 1.5k Dec 24, 2022
Luis M. Capdevielle 1 Jan 14, 2022
爬虫案例合集。包括但不限于《淘宝、京东、天猫、豆瓣、抖音、快手、微博、微信、阿里、头条、pdd、优酷、爱奇艺、携程、12306、58、搜狐、百度指数、维普万方、Zlibraty、Oalib、小说、招标网、采购网、小红书》

lxSpider 爬虫案例合集。包括但不限于《淘宝、京东、天猫、豆瓣、抖音、快手、微博、微信、阿里、头条、pdd、优酷、爱奇艺、携程、12306、58、搜狐、百度指数、维普万方、Zlibraty、Oalib、小说网站、招标采购网》 简介: 时光荏苒,记不清写了多少案例了。

lx 793 Jan 05, 2023
A Powerful Spider(Web Crawler) System in Python.

pyspider A Powerful Spider(Web Crawler) System in Python. Write script in Python Powerful WebUI with script editor, task monitor, project manager and

Roy Binux 15.7k Jan 04, 2023
robobrowser - A simple, Pythonic library for browsing the web without a standalone web browser.

RoboBrowser: Your friendly neighborhood web scraper Homepage: http://robobrowser.readthedocs.org/ RoboBrowser is a simple, Pythonic library for browsi

Joshua Carp 3.7k Dec 27, 2022
Parse feeds in Python

feedparser - Parse Atom and RSS feeds in Python. Copyright 2010-2020 Kurt McKee Kurt McKee 1.5k Dec 30, 2022

Find thumbnails and original images from URL or HTML file.

Haul Find thumbnails and original images from URL or HTML file. Demo Hauler on Heroku Installation on Ubuntu $ sudo apt-get install build-essential py

Vinta Chen 150 Oct 15, 2022
a way to scrape a database of all of the isef projects

ISEF Database This is a simple web scraper which gets all of the projects and abstract information from here. My goal for this is for someone to get i

William Kaiser 1 Mar 18, 2022
A multithreaded tool for searching and downloading images from popular search engines. It is straightforward to set up and run!

🕳️ CygnusX1 Code by Trong-Dat Ngo. Overviews 🕳️ CygnusX1 is a multithreaded tool 🛠️ , used to search and download images from popular search engine

DatNgo 32 Dec 31, 2022
Scraping web pages to get data

Scraping Data Get public data and save in database This is project use Python How to run a project 1 - Clone the repository 2 - Install beautifulsoup4

Soccer Project 2 Nov 01, 2021
Python Web Scrapper Project

Web Scrapper Projeto desenvolvido em python, sobre tudo com Selenium, BeautifulSoup e Pandas é um web scrapper que puxa uma tabela com as principais e

Jordan Ítalo Amaral 2 Jan 04, 2022
Library to scrape and clean web pages to create massive datasets.

lazynlp A straightforward library that allows you to crawl, clean up, and deduplicate webpages to create massive monolingual datasets. Using this libr

Chip Huyen 2.1k Jan 06, 2023
A Web Scraper built with beautiful soup, that fetches udemy course information. Get udemy course information and convert it to json, csv or xml file

Udemy Scraper A Web Scraper built with beautiful soup, that fetches udemy course information. Installation Virtual Environment Firstly, it is recommen

Aditya Gupta 15 May 17, 2022
Python script that reads Aliexpress offers urls from a Excel filename (.csv) and post then in a Telegram channel using a bot

Aliexpress to telegram post Python script that reads Aliexpress offers urls from a Excel filename (.csv) and post then in a Telegram channel using a b

Fernando 6 Dec 06, 2022
Subscrape - A Python scraper for substrate chains

subscrape A Python scraper for substrate chains that uses Subscan. Usage copy co

ChaosDAO 14 Dec 15, 2022
A Scrapper with python

Scrapper-en-python Scrapper des données signifie récuperer des données pour les traiter ou les analyser. En python, il y'a 2 grands moyens de scrapper

Lun4rIum 1 Dec 05, 2021
Minimal set of tools to conduct stealthy scraping.

Stealthy Scraping Tools Do not use puppeteer and playwright for scraping. Explanation. We only use the CDP to obtain the page source and to get the ab

Nikolai Tschacher 88 Jan 04, 2023