this is based on calsyslab project
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

114 lines
3.6 KiB

1 year ago
1 year ago
1 year ago
1 year ago
1 year ago
1 year ago
1 year ago
1 year ago
1 year ago
1 year ago
  1. __author__ = 'DarkWeb'
  2. '''
  3. Starting point of the Darkweb Markets Mining
  4. '''
  5. from datetime import *
  6. from MarketPlaces.DarkFox.crawler_selenium import crawler as crawlerDarkFox
  7. from MarketPlaces.MikesGrandStore.crawler_selenium import crawler as crawlerMikesGrandStore
  8. from MarketPlaces.DarkTor.crawler_selenium import crawler as crawlerDarkTor
  9. from MarketPlaces.AnonymousMarketplace.crawler_selenium import crawler as crawlerAnonymousMarketplace
  10. from MarketPlaces.CityMarket.crawler_selenium import crawler as crawlerCityMarket
  11. from MarketPlaces.M00nkeyMarket.crawler_selenium import crawler as crawlerM00nkeyMarket
  12. from MarketPlaces.ViceCity.crawler_selenium import crawler as crawlerViceCity
  13. from MarketPlaces.CypherMarketplace.crawler_selenium import crawler as crawlerCypher
  14. from MarketPlaces.PabloEscobarMarket.crawler_selenium import crawler as crawlerPabloEscobar
  15. from MarketPlaces.Ares.crawler_selenium import crawler as crawlerAres
  16. import configparser
  17. import os
  18. import subprocess
  19. config = configparser.ConfigParser()
  20. config.read('../../setup.ini')
  21. CURRENT_DATE = str("%02d" % date.today().month) + str("%02d" % date.today().day) + str("%04d" % date.today().year)
  22. # reads list of marketplaces
  23. def getMarkets():
  24. mkts = []
  25. with open('MarketsList.txt') as f:
  26. mkts = f.readlines()
  27. return mkts
  28. # Creates needed directories for marketplace if doesn't exist
  29. def createDirectory(mkt):
  30. # Package should already be there, holding crawler and parser
  31. # pagesDir = '../' + mkt + '/HTML_Pages'
  32. pagesDir = os.path.join(config.get('Project', 'shared_folder'), "MarketPlaces/" + mkt + "/HTML_Pages")
  33. if not os.path.isdir(pagesDir):
  34. os.makedirs(pagesDir)
  35. currentDateDir = pagesDir + '/' + CURRENT_DATE
  36. if not os.path.isdir(currentDateDir):
  37. os.mkdir(currentDateDir)
  38. listingDir = currentDateDir + '/Listing'
  39. if not os.path.isdir(listingDir):
  40. os.mkdir(listingDir)
  41. listReadDir = listingDir + '/Read'
  42. if not os.path.isdir(listReadDir):
  43. os.mkdir(listReadDir)
  44. descriptionDir = currentDateDir + '/Description'
  45. if not os.path.isdir(descriptionDir):
  46. os.mkdir(descriptionDir)
  47. descReadDir = descriptionDir + '/Read'
  48. if not os.path.isdir(descReadDir):
  49. os.mkdir(descReadDir)
  50. # Opens Tor Browser
  51. def opentor():
  52. global pid
  53. print("Connecting Tor...")
  54. pro = subprocess.Popen(config.get('TOR', 'firefox_binary_path'))
  55. pid = pro.pid
  56. # time.sleep(7.5)
  57. input('Press ENTER when Tor is connected to continue')
  58. return
  59. if __name__ == '__main__':
  60. opentor()
  61. mktsList = getMarkets()
  62. for mkt in mktsList:
  63. mkt = mkt.replace('\n','')
  64. print("\nCreating listing and description directories ... for " + mkt)
  65. createDirectory(mkt)
  66. print("Directories created.")
  67. if mkt == "DarkFox":
  68. # for base in json["DarkFox"]["base"]:
  69. # if crawlerDarkFox(base["url"], base["categories"]):
  70. # break
  71. crawlerDarkFox()
  72. elif mkt == "MikesGrandStore":
  73. crawlerMikesGrandStore()
  74. elif mkt == "DarkTor":
  75. crawlerDarkTor()
  76. elif mkt == "AnonymousMarketplace":
  77. crawlerAnonymousMarketplace()
  78. elif mkt == "CityMarket":
  79. crawlerCityMarket()
  80. elif mkt == "M00nkeyMarket":
  81. crawlerM00nkeyMarket()
  82. elif mkt == "ViceCity":
  83. crawlerViceCity()
  84. elif mkt == "CypherMarketplace":
  85. crawlerCypher()
  86. elif mkt == "PabloEscobarMarket":
  87. crawlerPabloEscobar()
  88. elif mkt == "Ares":
  89. crawlerAres()
  90. print("\nScraping process completed!")