how to manage huge quantity of url login password?
by ygrek - Monday April 8, 2024 at 11:04 AM
#1
Hi everyone, 
since their is more and more way to get url:login:password file i was wandering what was the best way to sort them, remove duplicate and then analyse them 
I mean it's ok to sort -u for file lower than 20gb but once their is more, let's say 500gb+ it's way harder and take so much time, i would have wanted to be able to gest a count by domain/mail/password simply for some stats. anyone have that kind of sorter? it's okay to also create a sort of flow to split by domain every digested files so each folder named after the domain contain a file or a bunch of files size limited inside without dup
thank you
Reply
#2
if you want to sort and remove duplicates from +500gb you need to use linux sort
Reply
#3
the sort -u is not realy the best way, for now i am using some sql converter for the simples data and multi threaded grep for the files but that's not the best option
Reply
#4
if you sort all the lines in one file its easy to remove duplicate from you just spilt it into 20 gb files and remove duplicates from each file
i have removed duplicate from 350gb file like this

also you need to use merge sort
sort each file and then use merge sort command to get one big sorted file
Reply


Possibly Related Threads…
Thread Author Replies Views Last Post
  DarkGPT Tutorial Easy idontknowmyname 188 8,174 22 minutes ago
Last Post: tomyss67
  STOP PAYING FOR CAPTCHA SERVICES!!! UNLIMITED CAPTCHA SOLVER TUTORIAL HASBULLA 87 14,338 5 hours ago
Last Post: zedfghjytgfvbhgfvc
  WormGPT? D3N1S 259 41,551 8 hours ago
Last Post: vx3n
  0day-Mari Bot Godfather1 77 7,390 Yesterday, 09:30 AM
Last Post: Diezxx
  [2026] Bypass AV / EDR Spearr 63 1,341 May 07, 2026, 07:19 PM
Last Post: AKASHIC

Forum Jump:


 Users browsing this forum: 1 Guest(s)