I will encrypt your phishing pages or letters for longer FUD

В этой теме можно использовать автоматический гарант!

Anunnaki

ripper
КИДАЛА
Регистрация
28.05.2020
Сообщения
39
Реакции
4
Пожалуйста, обратите внимание, что пользователь заблокирован
Most Phishing Campaigns encounter the problem of having their Phishing Pages / Websites /Letters backlisted (e.g RED PAGE)

960x0.jpg


I offer my service to make your phishing pages or letters to LAST ALOT LONGER before it gets flagged as "deceptive site"

PM me or message on Telegram below

Official Telegram Profile: @ScalarNetwork
Official Telegram Channel: @ScalarNetworkOfficial
 
Most Phishing Campaigns encounter the problem of having their Phishing Pages / Websites /Letters backlisted (e.g RED PAGE)

Посмотреть вложение 39319

I offer my service to make your phishing pages or letters to LAST ALOT LONGER before it gets flagged as "deceptive site"

PM me or message on Telegram below

Official Telegram Profile: @ScalarNetwork
Official Telegram Channel: @ScalarNetworkOfficial
How do you think you can guarantee this?
Modifying the sources is one temporany solution, but it will work until Safebrowsing takes a fingeprint of it and categorizes it as malicious.

Does your approach manage to bypass this behaviour?
Does it provide any kind of dynamic content generation?
 
Пожалуйста, обратите внимание, что пользователь заблокирован
How do you think you can guarantee this?
Modifying the sources is one temporany solution, but it will work until Safebrowsing takes a fingeprint of it and categorizes it as malicious.

Does your approach manage to bypass this behaviour?
Does it provide any kind of dynamic content generation?

sorry for responding late. i've been away for a while. I would've loved to explain my methods openly, but i can't since i'm commercializing it. I would be glad to show a sample privately.
 
How do you think you can guarantee this?
Modifying the sources is one temporany solution, but it will work until Safebrowsing takes a fingeprint of it and categorizes it as malicious.

Does your approach manage to bypass this behaviour?
Does it provide any kind of dynamic content generation?
I think using a redirect link with ip2location to block proxies IP ( dedicated + residential ) should take care of keeping the page undetected as long as the first link "redirect" still clean

Is there any good obfuscator one page that can prevent the page from getting detected the slowest possible?
 


Напишите ответ...
Верх