Takes a string and puts it through different languages in Google Translate a requested amount of times, returning nonsense.

Overview

PythonTextObfuscator

Takes a string and puts it through different languages in Google Translate a requested amount of times, returning nonsense. Example

Requirements:

python3

For the Selenium Obfuscator:

    -Selenium
    
    -Firefox
    
    -Geckodriver

In the Selenium Obfuscator:

-The major benefit is that you can translate excel documents, the downside is that after 10 or so document translations, Google blocks your ip for a while.

-Translation is generally slower and more limited using selenium as a browser tab is being used to scrape the data. Also beware of RAM usage.

-May no longer be supported in the future due to its drawbacks.

In the Urllib Obfuscator:

-Translation is generally faster and uses very little resources as only html is downloaded through a request. Multiprocessing also allows simultanious requests and can be used to the full extent without worrying about RAM usage.

—Split by length is faster and uses less requests (better for longer texts)

—Split by newline is slower and uses more requests but adds much more translation variety.

-Reminder: Since google has a url request limit, you'll need to switch VPN locations when the request limit is hit.

    ——Don't worry too much though, as it takes quite a bit of requests to get to that point, and the block only lasts for around an hour.
You might also like...
Translate - a PyTorch Language Library

NOTE PyTorch Translate is now deprecated, please use fairseq instead. Translate - a PyTorch Language Library Translate is a library for machine transl

Auto translate textbox from Japanese to English or Indonesia
Auto translate textbox from Japanese to English or Indonesia

priconne-auto-translate Auto translate textbox from Japanese to English or Indonesia How to use Install python first, Anaconda is recommended Install

translate using your voice
translate using your voice

speech-to-text-translator Usage translate using your voice description this project makes translating a word easy, all you have to do is speak and...

translate using your voice

speech-to-text-translator Usage translate using your voice description this project makes translating a word easy, all you have to do is speak and...

This program do translate english words to portuguese

Python-Dictionary This program is used to translate english words to portuguese. Web-Scraping This program use BeautifulSoap to make web scraping, so

Translate U is capable of translating the text present in an image from one language to the other.
Translate U is capable of translating the text present in an image from one language to the other.

Translate U is capable of translating the text present in an image from one language to the other. The app uses OCR and Google translate to identify and translate across 80+ languages.

Graphical user interface for Argos Translate
Graphical user interface for Argos Translate

Argos Translate GUI Website | GitHub | PyPI Graphical user interface for Argos Translate. Install pip3 install argostranslategui

Use the state-of-the-art m2m100 to translate large data on CPU/GPU/TPU. Super Easy!
Use the state-of-the-art m2m100 to translate large data on CPU/GPU/TPU. Super Easy!

Easy-Translate is a script for translating large text files in your machine using the M2M100 models from Facebook/Meta AI. We also privide a script fo

Search for documents in a domain through Google. The objective is to extract metadata

MetaFinder - Metadata search through Google _____ __ ___________ .__ .___ / \

Comments
  • Attempt to decode JSON with unexpected mimetype: text/plain

    Attempt to decode JSON with unexpected mimetype: text/plain

    I'm not sure what's causing this, as the last time I tried this release, this issue was not present. If it's accessing content server-side, then it might be that the server has had a config change resulting in it returning a different mimetype?

    I get the error message below consistently in the console, with %2E being added to the end of the URL each time. It does seem like some translation does happen; in this case, I inputted "Test", and the URL ended with "Hlola".

    https://translate.alefvanoon.xyz/api/v1/zu/mi/Hlola%2E 0, message='Attempt to decode JSON with unexpected mimetype: text/plain; charset=utf-8', url=URL('https://translate.alefvanoon.xyz/api/v1/zu/mi/Hlola')

    From what I've gathered looking online, the issue lies in either line 13, line 469, or both.

    return (await response.json())['translation'].replace('/','⁄')

    text = (await response.json())['translation'].replace('/','⁄')

    Some of the solutions online referred to adding "content_type=None" or "content_type='text/plain'" into the brackets after "json", but this only seemed to cause further issues for me.

    opened by UltraHylia 2
  • Program Freezes Up and Looping Error

    Program Freezes Up and Looping Error

    When you have Chinese (Simplified) and/or Chinese (Traditional) enabled in the language selector, the program can freeze and an error loops in the console. It happens no matter what other languages are enabled.

    https://user-images.githubusercontent.com/60769253/197659506-38871035-e311-4710-9eb9-ac2d7387841f.mp4

    opened by DerpTaco99921 0
Releases(v0.4)
  • v0.4(Feb 2, 2022)

    Rebuilt from the ground up with a new GUI and translation method.

    Changes:

    -Improved GUI.

    -Translations are retrieved from a front-end to Google Translate called Lingva, which removes the issue with being blocked for doing too many requests.

    -Translations are done in an asynchronous function using aiohttp instead of a process pool, which is optimal for large bulk translations.

    -Removed selenium obfuscation.

    Additions: -Importing and saving text files. -Language Selector to activate or deactivate any individual language. -Language setting for the result. -Three different split methods: ____-Initial ________-Text is split by length before being passed into the obfuscate function. ________-Faster as less requests are made. ________-Different languages for each piece. ________-Tabs not preserved. ____-Continuous ________-Text is split by length inside the obfuscate function. ________-Faster as less requests are made. ________-Same languages for each piece. ________-Tabs not preserved. ____-Newline ________-Text is split by newlines and tabs. ________-Slower as more requests are made. ________-Every single line is translated with different languages. ________-Tabs preserved. -Translation Generator which creates a .csv file containing multiple translations of the same text: ____-Repeat mode obfuscates the original text each time, adding the result in each new column. ____-Continue mode obfuscates the results from each subsequent obfuscation, adding the result in each new column.

    Source code(tar.gz)
    Source code(zip)
    Python.Text.Obfuscator.v0.4.zip(15.75 KB)
  • v0.3.1c-r2(Dec 23, 2021)

  • v0.3.1c(Dec 23, 2021)

    Newlines no longer get messed up in Urllib Obfuscator. Added a choice to split by length or by newlines. —Split by length is faster and uses less requests (better for longer texts) —Split by newline is slower and uses more requests but adds much more translation variety. Reminder: Since google has a URL request limit, you'll need to switch VPN locations when the request limit is hit.

    Source code(tar.gz)
    Source code(zip)
    Python.Text.Obfuscator.v0.3.1c.zip(51.63 KB)
  • v0.3.1b(Dec 23, 2021)

  • v0.3.1a(Dec 23, 2021)

  • v0.3(Dec 23, 2021)

    I made massive improvements to the speed of the obfuscation thanks to learning about urllib.

    For example, I did translated the same ~2300 character long string of text 10 times in the old and new version; the old one took 38.8 seconds while the new one took only 6.8 seconds.

    In addition, the capacity to add a larger amount of characters is far increased as it doesn't require Firefox tabs to be open and eating up ram.

    As a test I translated the entire Among Us Wikipedia page 50 times (with a character count of over 60 thousand!), and it only took only 114 seconds to finish translating. Using the old obfuscator I wouldn't be able to translate more than half that amount, and it would take ages to complete (Like 10 mins or more).

    Unfortunately for this version the Excel Obfuscator is removed until I can figure out how to get it to work in urllib, if I can't then I'll probably add it back it with Selenium.

    At least if you couldn't get selenium to work on your computer for the previous versions you don't have to worry about getting it for this.

    Source code(tar.gz)
    Source code(zip)
    Python.Text.Obfuscator.v0.3.zip(5.73 KB)
  • v0.2.2(Dec 23, 2021)

  • v0.2.1b(Dec 23, 2021)

  • v0.2.1a(Dec 23, 2021)

    Fixed TimeoutExceptions for the string translations (textbox input) obfuscation. You can now do as many translations as you want without worrying about encountering an error. Same for amount of characters (as long as your PC can handle of course). As for excel translations they remain unchanged — since I can't do anything about Google's Document translation limit — so just switch locations on VPN like usual after 10 translations for the Excel Obfuscator.

    Source code(tar.gz)
    Source code(zip)
    Python.Text.Obfuscator.v0.2.1.zip(5.88 KB)
  • v0.2(Dec 23, 2021)

  • v0.1b(Dec 23, 2021)

  • v0.1a(Dec 23, 2021)

Tools and data for measuring the popularity & growth of various programming languages.

growth-data Tools and data for measuring the popularity & growth of various programming languages. Install the dependencies $ pip install -r requireme

3 Jan 06, 2022
Code for CVPR 2021 paper: Revamping Cross-Modal Recipe Retrieval with Hierarchical Transformers and Self-supervised Learning

Revamping Cross-Modal Recipe Retrieval with Hierarchical Transformers and Self-supervised Learning This is the PyTorch companion code for the paper: A

Amazon 69 Jan 03, 2023
Use AutoModelForSeq2SeqLM in Huggingface Transformers to train COMET

Training COMET using seq2seq setting Use AutoModelForSeq2SeqLM in Huggingface Transformers to train COMET. The codes are modified from run_summarizati

tqfang 9 Dec 17, 2022
SciBERT is a BERT model trained on scientific text.

SciBERT is a BERT model trained on scientific text.

AI2 1.2k Dec 24, 2022
BROS: A Pre-trained Language Model Focusing on Text and Layout for Better Key Information Extraction from Documents

BROS (BERT Relying On Spatiality) is a pre-trained language model focusing on text and layout for better key information extraction from documents. Given the OCR results of the document image, which

Clova AI Research 94 Dec 30, 2022
Some embedding layer implementation using ivy library

ivy-manual-embeddings Some embedding layer implementation using ivy library. Just for fun. It is based on NYCTaxiFare dataset from kaggle (cut down to

Ishtiaq Hussain 2 Feb 10, 2022
The proliferation of disinformation across social media has led the application of deep learning techniques to detect fake news.

Fake News Detection Overview The proliferation of disinformation across social media has led the application of deep learning techniques to detect fak

Kushal Shingote 1 Feb 08, 2022
Pipelines de datos, 2021.

Este repo ilustra un proceso sencillo de automatización de transformación y modelado de datos, a través de un pipeline utilizando Luigi. Stack princip

Rodolfo Ferro 8 May 19, 2022
A2T: Towards Improving Adversarial Training of NLP Models (EMNLP 2021 Findings)

A2T: Towards Improving Adversarial Training of NLP Models This is the source code for the EMNLP 2021 (Findings) paper "Towards Improving Adversarial T

QData 17 Oct 15, 2022
Twitter-Sentiment-Analysis - Twitter sentiment analysis for india's top online retailers(2019 to 2022)

Twitter-Sentiment-Analysis Twitter sentiment analysis for india's top online retailers(2019 to 2022) Project Overview : Sentiment Analysis helps us to

Balaji R 1 Jan 01, 2022
A BERT-based reverse-dictionary of Korean proverbs

Wisdomify A BERT-based reverse-dictionary of Korean proverbs. 김유빈 : 모델링 / 데이터 수집 / 프로젝트 설계 / back-end 김종윤 : 데이터 수집 / 프로젝트 설계 / front-end Quick Start C

Eu-Bin KIM 94 Dec 08, 2022
Finally, some decent sample sentences

tts-dataset-prompts This repository aims to be a decent set of sentences for people looking to clone their own voices (e.g. using Tacotron 2). Each se

hecko 19 Dec 13, 2022
The (extremely) naive sentiment classification function based on NBSVM trained on wisesight_sentiment

thai_sentiment The naive sentiment classification function based on NBSVM trained on wisesight_sentiment วิธีติดตั้ง pip install thai_sentiment==0.1.3

Charin 7 Dec 08, 2022
Google and Stanford University released a new pre-trained model called ELECTRA

Google and Stanford University released a new pre-trained model called ELECTRA, which has a much compact model size and relatively competitive performance compared to BERT and its variants. For furth

Yiming Cui 1.2k Dec 30, 2022
NLTK Source

Natural Language Toolkit (NLTK) NLTK -- the Natural Language Toolkit -- is a suite of open source Python modules, data sets, and tutorials supporting

Natural Language Toolkit 11.4k Jan 04, 2023
Research code for ECCV 2020 paper "UNITER: UNiversal Image-TExt Representation Learning"

UNITER: UNiversal Image-TExt Representation Learning This is the official repository of UNITER (ECCV 2020). This repository currently supports finetun

Yen-Chun Chen 680 Dec 24, 2022
:hot_pepper: R²SQL: "Dynamic Hybrid Relation Network for Cross-Domain Context-Dependent Semantic Parsing." (AAAI 2021)

R²SQL The PyTorch implementation of paper Dynamic Hybrid Relation Network for Cross-Domain Context-Dependent Semantic Parsing. (AAAI 2021) Requirement

huybery 60 Dec 31, 2022
Search with BERT vectors in Solr and Elasticsearch

Search with BERT vectors in Solr and Elasticsearch

Dmitry Kan 123 Dec 29, 2022
Source code of the "Graph-Bert: Only Attention is Needed for Learning Graph Representations" paper

Graph-Bert Source code of "Graph-Bert: Only Attention is Needed for Learning Graph Representations". Please check the script.py as the entry point. We

14 Mar 25, 2022
CCKS-Title-based-large-scale-commodity-entity-retrieval-top1

- 基于标题的大规模商品实体检索top1 一、任务介绍 CCKS 2020:基于标题的大规模商品实体检索,任务为对于给定的一个商品标题,参赛系统需要匹配到该标题在给定商品库中的对应商品实体。 输入:输入文件包括若干行商品标题。 输出:输出文本每一行包括此标题对应的商品实体,即给定知识库中商品 ID,

43 Nov 11, 2022