Takes a string and puts it through different languages in Google Translate a requested amount of times, returning nonsense.

Overview

PythonTextObfuscator

Takes a string and puts it through different languages in Google Translate a requested amount of times, returning nonsense. Example

Requirements:

python3

For the Selenium Obfuscator:

    -Selenium
    
    -Firefox
    
    -Geckodriver

In the Selenium Obfuscator:

-The major benefit is that you can translate excel documents, the downside is that after 10 or so document translations, Google blocks your ip for a while.

-Translation is generally slower and more limited using selenium as a browser tab is being used to scrape the data. Also beware of RAM usage.

-May no longer be supported in the future due to its drawbacks.

In the Urllib Obfuscator:

-Translation is generally faster and uses very little resources as only html is downloaded through a request. Multiprocessing also allows simultanious requests and can be used to the full extent without worrying about RAM usage.

—Split by length is faster and uses less requests (better for longer texts)

—Split by newline is slower and uses more requests but adds much more translation variety.

-Reminder: Since google has a url request limit, you'll need to switch VPN locations when the request limit is hit.

    ——Don't worry too much though, as it takes quite a bit of requests to get to that point, and the block only lasts for around an hour.
You might also like...
Translate - a PyTorch Language Library

NOTE PyTorch Translate is now deprecated, please use fairseq instead. Translate - a PyTorch Language Library Translate is a library for machine transl

Auto translate textbox from Japanese to English or Indonesia
Auto translate textbox from Japanese to English or Indonesia

priconne-auto-translate Auto translate textbox from Japanese to English or Indonesia How to use Install python first, Anaconda is recommended Install

translate using your voice
translate using your voice

speech-to-text-translator Usage translate using your voice description this project makes translating a word easy, all you have to do is speak and...

translate using your voice

speech-to-text-translator Usage translate using your voice description this project makes translating a word easy, all you have to do is speak and...

This program do translate english words to portuguese

Python-Dictionary This program is used to translate english words to portuguese. Web-Scraping This program use BeautifulSoap to make web scraping, so

Translate U is capable of translating the text present in an image from one language to the other.
Translate U is capable of translating the text present in an image from one language to the other.

Translate U is capable of translating the text present in an image from one language to the other. The app uses OCR and Google translate to identify and translate across 80+ languages.

Graphical user interface for Argos Translate
Graphical user interface for Argos Translate

Argos Translate GUI Website | GitHub | PyPI Graphical user interface for Argos Translate. Install pip3 install argostranslategui

Use the state-of-the-art m2m100 to translate large data on CPU/GPU/TPU. Super Easy!
Use the state-of-the-art m2m100 to translate large data on CPU/GPU/TPU. Super Easy!

Easy-Translate is a script for translating large text files in your machine using the M2M100 models from Facebook/Meta AI. We also privide a script fo

Search for documents in a domain through Google. The objective is to extract metadata

MetaFinder - Metadata search through Google _____ __ ___________ .__ .___ / \

Comments
  • Attempt to decode JSON with unexpected mimetype: text/plain

    Attempt to decode JSON with unexpected mimetype: text/plain

    I'm not sure what's causing this, as the last time I tried this release, this issue was not present. If it's accessing content server-side, then it might be that the server has had a config change resulting in it returning a different mimetype?

    I get the error message below consistently in the console, with %2E being added to the end of the URL each time. It does seem like some translation does happen; in this case, I inputted "Test", and the URL ended with "Hlola".

    https://translate.alefvanoon.xyz/api/v1/zu/mi/Hlola%2E 0, message='Attempt to decode JSON with unexpected mimetype: text/plain; charset=utf-8', url=URL('https://translate.alefvanoon.xyz/api/v1/zu/mi/Hlola')

    From what I've gathered looking online, the issue lies in either line 13, line 469, or both.

    return (await response.json())['translation'].replace('/','⁄')

    text = (await response.json())['translation'].replace('/','⁄')

    Some of the solutions online referred to adding "content_type=None" or "content_type='text/plain'" into the brackets after "json", but this only seemed to cause further issues for me.

    opened by UltraHylia 2
  • Program Freezes Up and Looping Error

    Program Freezes Up and Looping Error

    When you have Chinese (Simplified) and/or Chinese (Traditional) enabled in the language selector, the program can freeze and an error loops in the console. It happens no matter what other languages are enabled.

    https://user-images.githubusercontent.com/60769253/197659506-38871035-e311-4710-9eb9-ac2d7387841f.mp4

    opened by DerpTaco99921 0
Releases(v0.4)
  • v0.4(Feb 2, 2022)

    Rebuilt from the ground up with a new GUI and translation method.

    Changes:

    -Improved GUI.

    -Translations are retrieved from a front-end to Google Translate called Lingva, which removes the issue with being blocked for doing too many requests.

    -Translations are done in an asynchronous function using aiohttp instead of a process pool, which is optimal for large bulk translations.

    -Removed selenium obfuscation.

    Additions: -Importing and saving text files. -Language Selector to activate or deactivate any individual language. -Language setting for the result. -Three different split methods: ____-Initial ________-Text is split by length before being passed into the obfuscate function. ________-Faster as less requests are made. ________-Different languages for each piece. ________-Tabs not preserved. ____-Continuous ________-Text is split by length inside the obfuscate function. ________-Faster as less requests are made. ________-Same languages for each piece. ________-Tabs not preserved. ____-Newline ________-Text is split by newlines and tabs. ________-Slower as more requests are made. ________-Every single line is translated with different languages. ________-Tabs preserved. -Translation Generator which creates a .csv file containing multiple translations of the same text: ____-Repeat mode obfuscates the original text each time, adding the result in each new column. ____-Continue mode obfuscates the results from each subsequent obfuscation, adding the result in each new column.

    Source code(tar.gz)
    Source code(zip)
    Python.Text.Obfuscator.v0.4.zip(15.75 KB)
  • v0.3.1c-r2(Dec 23, 2021)

  • v0.3.1c(Dec 23, 2021)

    Newlines no longer get messed up in Urllib Obfuscator. Added a choice to split by length or by newlines. —Split by length is faster and uses less requests (better for longer texts) —Split by newline is slower and uses more requests but adds much more translation variety. Reminder: Since google has a URL request limit, you'll need to switch VPN locations when the request limit is hit.

    Source code(tar.gz)
    Source code(zip)
    Python.Text.Obfuscator.v0.3.1c.zip(51.63 KB)
  • v0.3.1b(Dec 23, 2021)

  • v0.3.1a(Dec 23, 2021)

  • v0.3(Dec 23, 2021)

    I made massive improvements to the speed of the obfuscation thanks to learning about urllib.

    For example, I did translated the same ~2300 character long string of text 10 times in the old and new version; the old one took 38.8 seconds while the new one took only 6.8 seconds.

    In addition, the capacity to add a larger amount of characters is far increased as it doesn't require Firefox tabs to be open and eating up ram.

    As a test I translated the entire Among Us Wikipedia page 50 times (with a character count of over 60 thousand!), and it only took only 114 seconds to finish translating. Using the old obfuscator I wouldn't be able to translate more than half that amount, and it would take ages to complete (Like 10 mins or more).

    Unfortunately for this version the Excel Obfuscator is removed until I can figure out how to get it to work in urllib, if I can't then I'll probably add it back it with Selenium.

    At least if you couldn't get selenium to work on your computer for the previous versions you don't have to worry about getting it for this.

    Source code(tar.gz)
    Source code(zip)
    Python.Text.Obfuscator.v0.3.zip(5.73 KB)
  • v0.2.2(Dec 23, 2021)

  • v0.2.1b(Dec 23, 2021)

  • v0.2.1a(Dec 23, 2021)

    Fixed TimeoutExceptions for the string translations (textbox input) obfuscation. You can now do as many translations as you want without worrying about encountering an error. Same for amount of characters (as long as your PC can handle of course). As for excel translations they remain unchanged — since I can't do anything about Google's Document translation limit — so just switch locations on VPN like usual after 10 translations for the Excel Obfuscator.

    Source code(tar.gz)
    Source code(zip)
    Python.Text.Obfuscator.v0.2.1.zip(5.88 KB)
  • v0.2(Dec 23, 2021)

  • v0.1b(Dec 23, 2021)

  • v0.1a(Dec 23, 2021)

multi-label,classifier,text classification,多标签文本分类,文本分类,BERT,ALBERT,multi-label-classification,seq2seq,attention,beam search

multi-label,classifier,text classification,多标签文本分类,文本分类,BERT,ALBERT,multi-label-classification,seq2seq,attention,beam search

hellonlp 30 Dec 12, 2022
Simple program that translates the name of files into English

Simple program that translates the name of files into English. Useful for when editing/inspecting programs that were developed in a foreign language.

0 Dec 22, 2021
Transformer - A TensorFlow Implementation of the Transformer: Attention Is All You Need

[UPDATED] A TensorFlow Implementation of Attention Is All You Need When I opened this repository in 2017, there was no official code yet. I tried to i

Kyubyong Park 3.8k Dec 26, 2022
RuCLIP tiny (Russian Contrastive Language–Image Pretraining) is a neural network trained to work with different pairs (images, texts).

RuCLIPtiny Zero-shot image classification model for Russian language RuCLIP tiny (Russian Contrastive Language–Image Pretraining) is a neural network

Shahmatov Arseniy 26 Sep 20, 2022
The RWKV Language Model

RWKV-LM We propose the RWKV language model, with alternating time-mix and channel-mix layers: The R, K, V are generated by linear transforms of input,

PENG Bo 877 Jan 05, 2023
Contains the code and data for our #ICSE2022 paper titled as "CodeFill: Multi-token Code Completion by Jointly Learning from Structure and Naming Sequences"

CodeFill This repository contains the code for our paper titled as "CodeFill: Multi-token Code Completion by Jointly Learning from Structure and Namin

Software Analytics Lab 11 Oct 31, 2022
NeMo: a toolkit for conversational AI

NVIDIA NeMo Introduction NeMo is a toolkit for creating Conversational AI applications. NeMo product page. Introductory video. The toolkit comes with

NVIDIA Corporation 5.3k Jan 04, 2023
Code for the project carried out fulfilling the course requirements for Fall 2021 NLP at NYU

Introduction Fairseq(-py) is a sequence modeling toolkit that allows researchers and developers to train custom models for translation, summarization,

Sai Himal Allu 1 Apr 25, 2022
MicBot - MicBot uses Google Translate to speak everyone's chat messages

MicBot MicBot uses Google Translate to speak everyone's chat messages. It can al

2 Mar 09, 2022
SAVI2I: Continuous and Diverse Image-to-Image Translation via Signed Attribute Vectors

SAVI2I: Continuous and Diverse Image-to-Image Translation via Signed Attribute Vectors [Paper] [Project Website] Pytorch implementation for SAVI2I. We

Qi Mao 44 Dec 30, 2022
IndoBERTweet is the first large-scale pretrained model for Indonesian Twitter. Published at EMNLP 2021 (main conference)

IndoBERTweet 🐦 🇮🇩 1. Paper Fajri Koto, Jey Han Lau, and Timothy Baldwin. IndoBERTweet: A Pretrained Language Model for Indonesian Twitter with Effe

IndoLEM 40 Nov 30, 2022
MEDIALpy: MEDIcal Abbreviations Lookup in Python

A small python package that allows the user to look up common medical abbreviations.

Aberystwyth Systems Biology 7 Nov 09, 2022
Guide: Finetune GPT2-XL (1.5 Billion Parameters) and GPT-NEO (2.7 B) on a single 16 GB VRAM V100 Google Cloud instance with Huggingface Transformers using DeepSpeed

Guide: Finetune GPT2-XL (1.5 Billion Parameters) and GPT-NEO (2.7 Billion Parameters) on a single 16 GB VRAM V100 Google Cloud instance with Huggingfa

289 Jan 06, 2023
Unofficial implementation of Google's FNet: Mixing Tokens with Fourier Transforms

FNet: Mixing Tokens with Fourier Transforms Pytorch implementation of Fnet : Mixing Tokens with Fourier Transforms. Citation: @misc{leethorp2021fnet,

Rishikesh (ऋषिकेश) 217 Dec 05, 2022
Search-Engine - 📖 AI based search engine

Search Engine AI based search engine that was trained on 25000 samples, feel free to train on up to 1.2M sample from kaggle dataset, link below StackS

Vladislav Kruglikov 2 Nov 29, 2022
OCR을 이용하여 인원수를 인식 후 줌을 Kill 해줍니다

How To Use killtheZoom-2.0 Windows 0. https://joyhong.tistory.com/79 이 글을 보면서 tesseract를 C:\Program Files\Tesseract-OCR 경로로 설치해주세요(한국어 언어 추가 필요) 상단의 초

김정인 9 Sep 13, 2021
A curated list of FOSS tools to improve the Hacker News experience

Awesome-Hackernews Hacker News is a social news website focusing on computer technologies, hacking and startups. It promotes any content likely to "gr

Bryton Lacquement 141 Dec 27, 2022
Topic Inference with Zeroshot models

zeroshot_topics Table of Contents Installation Usage License Installation zeroshot_topics is distributed on PyPI as a universal wheel and is available

Rita Anjana 55 Nov 28, 2022
Galois is an auto code completer for code editors (or any text editor) based on OpenAI GPT-2.

Galois is an auto code completer for code editors (or any text editor) based on OpenAI GPT-2. It is trained (finetuned) on a curated list of approximately 45K Python (~470MB) files gathered from the

Galois Autocompleter 91 Sep 23, 2022
This repository has a implementations of data augmentation for NLP for Japanese.

daaja This repository has a implementations of data augmentation for NLP for Japanese: EDA: Easy Data Augmentation Techniques for Boosting Performance

Koga Kobayashi 60 Nov 11, 2022