Advertisement
WhosYourDaddySec

GhostSec Deep Scan ToolSet (Baseline)

Dec 14th, 2023
161
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
Python 13.88 KB | None | 0 0
  1. import requests
  2. from bs4 import BeautifulSoup
  3. from urllib.parse import urlparse, urljoin
  4. import builtwith
  5. import whois
  6. from sslyze import server_connectivity, plugins
  7. from PyInquirer import prompt
  8. from termcolor import colored
  9.  
  10.  
  11. class GhostSecDeepScanToolSet:
  12.     def __init__(self):
  13.         self.visited_urls = set()
  14.  
  15.     def show_menu(self):
  16.         questions = [
  17.             {
  18.                 'type': 'list',
  19.                 'name': 'option',
  20.                 'message': colored('GhostSec Deep Scan ToolSet - Select an option:', 'cyan'),
  21.                 'choices': [
  22.                     colored('Perform User-Specific Search', 'green'),
  23.                     colored('Scan Multiple Sites', 'green'),
  24.                     colored('Advanced Search Options', 'green'),
  25.                     colored('Spider Function', 'green'),
  26.                     colored('Deep Scan and Technology Analysis', 'green'),
  27.                     colored('Security Protocols', 'green'),
  28.                     colored('Error Handling Protocols', 'green'),
  29.                     colored('Whois Lookup', 'green'),
  30.                     colored('Save Data to File', 'green'),
  31.                     colored('Content Analysis', 'green'),
  32.                     colored('SSL Analysis', 'green'),
  33.                     colored('Check robots.txt', 'green'),
  34.                     colored('Check sitemap.xml', 'green'),
  35.                     colored('DNS Lookup', 'green'),
  36.                     colored('Perform CVE Scan', 'green'),
  37.                     colored('Exit', 'red')
  38.                 ],
  39.             }
  40.         ]
  41.  
  42.         try:
  43.             answers = prompt(questions)
  44.             return answers['option']
  45.         except KeyboardInterrupt:
  46.             print("\nExiting GhostSec Deep Scan ToolSet.")
  47.             exit()
  48.  
  49.     def execute_option(self, option):
  50.         try:
  51.             if option == colored('Perform User-Specific Search', 'green'):
  52.                 username = input(colored("Enter username: ", 'cyan'))
  53.                 user_query = input(colored("Enter search query: ", 'cyan'))
  54.                 self.harvest_data(username, user_query)
  55.  
  56.             elif option == colored('Scan Multiple Sites', 'green'):
  57.                 sites = input(colored("Enter sites separated by commas: ", 'cyan')).split(',')
  58.                 user_query = input(colored("Enter search query: ", 'cyan'))
  59.                 self.scan_multiple_sites(sites, user_query)
  60.  
  61.             elif option == colored('Advanced Search Options', 'green'):
  62.                 self.advanced_search_options()
  63.  
  64.             elif option == colored('Spider Function', 'green'):
  65.                 start_url = input(colored("Enter starting URL: ", 'cyan'))
  66.                 max_depth = int(input(colored("Enter maximum depth: ", 'cyan')))
  67.                 self.crawl(start_url, max_depth)
  68.  
  69.             elif option == colored('Deep Scan and Technology Analysis', 'green'):
  70.                 url = input(colored("Enter URL for deep scan: ", 'cyan'))
  71.                 self.deep_scan(url)
  72.  
  73.             elif option == colored('Security Protocols', 'green'):
  74.                 self.security_protocols()
  75.  
  76.             elif option == colored('Error Handling Protocols', 'green'):
  77.                 self.error_handling_protocols()
  78.  
  79.             elif option == colored('Whois Lookup', 'green'):
  80.                 url = input(colored("Enter the website URL for Whois lookup: ", 'cyan'))
  81.                 self.whois_lookup(url)
  82.  
  83.             elif option == colored('Save Data to File', 'green'):
  84.                 filename = input(colored("Enter the filename to save data: ", 'cyan'))
  85.                 content = input(colored("Enter the content to save: ", 'cyan'))
  86.                 self.save_to_file(filename, content)
  87.  
  88.             elif option == colored('Content Analysis', 'green'):
  89.                 # Placeholder for content analysis logic
  90.                 pass
  91.  
  92.             elif option == colored('SSL Analysis', 'green'):
  93.                 url = input(colored("Enter the website URL for SSL analysis: ", 'cyan'))
  94.                 self.ssl_analysis(url)
  95.  
  96.             elif option == colored('Check robots.txt', 'green'):
  97.                 url = input(colored("Enter the website URL to check robots.txt: ", 'cyan'))
  98.                 self.check_robots_txt(url)
  99.  
  100.             elif option == colored('Check sitemap.xml', 'green'):
  101.                 url = input(colored("Enter the website URL to check sitemap.xml: ", 'cyan'))
  102.                 self.check_sitemap_xml(url)
  103.  
  104.             elif option == colored('DNS Lookup', 'green'):
  105.                 domain = input(colored("Enter the domain for DNS lookup: ", 'cyan'))
  106.                 self.dns_lookup(domain)
  107.  
  108.             elif option == colored('Perform CVE Scan', 'green'):
  109.                 url = input(colored("Enter the website URL for CVE scan: ", 'cyan'))
  110.                 self.perform_cve_scan(url)
  111.  
  112.             elif option == colored('Exit', 'red'):
  113.                 print(colored("Exiting GhostSec Deep Scan ToolSet.", 'red'))
  114.                 exit()
  115.  
  116.         except KeyboardInterrupt:
  117.             print("\nExiting GhostSec Deep Scan ToolSet.")
  118.             exit()
  119.         except Exception as e:
  120.             print(colored(f"An error occurred: {e}", 'red'))
  121.  
  122.     def run(self):
  123.         while True:
  124.             option = self.show_menu()
  125.             self.execute_option(option)
  126.  
  127.     def harvest_data(self, username, user_query):
  128.         search_query = f"site:{username}.com {user_query}"
  129.         search_results = self.perform_web_search(search_query)
  130.         harvested_data = [self.extract_data(result) for result in search_results]
  131.         self.display_data(harvested_data)
  132.  
  133.     def scan_multiple_sites(self, sites, user_query):
  134.         for site in sites self.harvest_data(site, user_query)
  135.  
  136.     def advanced_search_options(self):
  137.         print(colored("Advanced Search Options:", 'cyan'))
  138.         print(colored("1. Search by File Type", 'cyan'))
  139.         print(colored("2. Search by Date Range", 'cyan'))
  140.         print(colored("3. Search by Custom Criteria", 'cyan'))
  141.         option = int(input(colored("Enter your choice (1, 2, or 3): ", 'cyan')))
  142.  
  143.         if option == 1:
  144.             file_type = input(colored("Enter the file type (e.g., pdf): ", 'cyan'))
  145.             self.file_type_search(file_type)
  146.         elif option == 2:
  147.             start_date = input(colored("Enter the start date (YYYY-MM-DD): ", 'cyan'))
  148.             end_date = input(colored("Enter the end date (YYYY-MM-DD): ", 'cyan'))
  149.             self.date_range_search(start_date, end_date)
  150.         elif option == 3:
  151.             custom_criteria = input(colored("Enter custom search criteria: ", 'cyan'))
  152.             self.custom_criteria_search(custom_criteria)
  153.         else:
  154.             print(colored("Invalid option. Returning to the main menu.", 'red'))
  155.  
  156.     def file_type_search(self, file_type):
  157.         # Placeholder for file type-specific search logic
  158.         pass
  159.  
  160.     def date_range_search(self, start_date, end_date):
  161.         # Placeholder for date range-specific search logic
  162.         pass
  163.  
  164.     def custom_criteria_search(self, custom_criteria):
  165.         # Placeholder for custom criteria-specific search logic
  166.         pass
  167.  
  168.     def crawl(self, start_url, max_depth):
  169.         self.visited_urls.clear()  # Clear visited URLs for each new crawl
  170.         self.crawl_recursive(start_url, 0, max_depth)
  171.  
  172.     def crawl_recursive(self, url, depth, max_depth):
  173.         if depth > max_depth or url in self.visited_urls or not self.is_valid_url(url):
  174.             return
  175.  
  176.         self.visited_urls.add(url)
  177.         print(colored(f"Exploring: {url}", 'cyan'))
  178.  
  179.         # Add logic to extract and process data from the page
  180.         # For example, you can call a function to harvest data from the page
  181.  
  182.         # Extract links and continue crawling
  183.         links = self.extract_links(url)
  184.         for link in links:
  185.             self.crawl_recursive(link, depth + 1, max_depth)
  186.  
  187.     def deep_scan(self, url):
  188.         try:
  189.             html_source, technologies = self.fetch_deep_scan_data(url)
  190.             if html_source and technologies:
  191.                 print(colored("HTML Source Code:", 'cyan'))
  192.                 print(html_source)
  193.                 print(colored("Technologies Used:", 'cyan'))
  194.                 print(technologies)
  195.             else:
  196.                 print(colored("Deep scan failed. Check the URL and try again.", 'red'))
  197.         except Exception as e:
  198.             print(colored(f"An error occurred during deep scan: {e}", 'red'))
  199.  
  200.     def fetch_deep_scan_data(self, url):
  201.         try:
  202.             response = requests.get(url)
  203.             response.raise_for_status()  # Raise an exception for bad response status
  204.  
  205.             soup = BeautifulSoup(response.text, 'html.parser')
  206.  
  207.             # Parse technologies using builtwith library
  208.             technologies = builtwith.builtwith(url)
  209.  
  210.             return soup.prettify(), technologies
  211.         except requests.exceptions.RequestException as req_err:
  212.             print(colored(f"Error: {req_err}", 'red'))
  213.             return None, None
  214.  
  215.     def security_protocols(self):
  216.         print(colored("Security Protocols:", 'cyan'))
  217.         print(colored("1. Check for SSL/TLS", 'cyan'))
  218.         print(colored("2. Check for HTTP Strict Transport Security (HSTS)", 'cyan'))
  219.         option = int(input(colored("Enter your choice (1 or 2): ", 'cyan')))
  220.  
  221.         if option == 1:
  222.             url = input(colored("Enter the website URL to check for SSL/TLS: ", 'cyan'))
  223.             self.check_ssl_tls(url)
  224.         elif option == 2:
  225.             url = input(colored("Enter the website URL to check for HSTS: ", 'cyan'))
  226.             self.check_hsts(url)
  227.         else:
  228.             print(colored("Invalid option. Returning to the main menu.", 'red'))
  229.  
  230.     def check_ssl_tls(self, url):
  231.         try:
  232.             result = requests.get(url)
  233.             if result.url.startswith('https'):
  234.                 print(colored("The website is using SSL/TLS.", 'green'))
  235.             else:
  236.                 print(colored("The website is not using SSL/TLS.", 'yellow'))
  237.         except requests.exceptions.RequestException as req_err:
  238.             print(colored(f"Error checking SSL/TLS: {req_err}", 'red'))
  239.  
  240.     def check_hsts(self, url):
  241.         try:
  242.             headers = requests.head(url).headers
  243.             if 'Strict-Transport-Security' in headers:
  244.                 print(colored("The website has HTTP Strict Transport Security (HSTS) enabled.", 'green'))
  245.             else:
  246.                 print(colored("The website does not have HSTS enabled.", 'yellow'))
  247.         except requests.exceptions.RequestException as req_err:
  248.             print(colored(f"Error checking HSTS: {req_err}", 'red'))
  249.  
  250.     def error_handling_protocols(self):
  251.         print(colored("Error Handling Protocols:", 'cyan'))
  252.         print(colored("1. Check for Broken Links", 'cyan'))
  253.         option = int(input(colored("Enter your choice (1): ", 'cyan')))
  254.  
  255.         if option == 1:
  256.             url = input(colored("Enter the website URL to check for broken links: ", 'cyan'))
  257.             self.check_broken_links(url)
  258.         else:
  259.             print(colored("Invalid option. Returning to the main menu.", 'red'))
  260.  
  261.     def check_broken_links(self, url):
  262.         try:
  263.             response = requests.get(url)
  264.             response.raise_for_status()
  265.             print(colored("No broken links found.", 'green'))
  266.         except requests.exceptions.HTTPError as http_err:
  267.             print(colored(f"HTTP Error: {http_err}", 'yellow'))
  268.         except requests.exceptions.RequestException as req_err:
  269.             print(colored(f"Request Error: {req_err}", 'yellow'))
  270.  
  271.     def whois_lookup(self, url):
  272.         try:
  273.             domain = urlparse(url).hostname
  274.             whois_info = whois.whois(domain)
  275.             print(colored("Whois Information:", 'cyan'))
  276.             print(whois_info)
  277.         except Exception as whois_err:
  278.             print(colored(f"Whois Lookup Error: {whois_err}", 'red'))
  279.  
  280.     def save_to_file(self, filename, content):
  281.         try:
  282.             with open(filename, 'w', encoding='utf-8') as file:
  283.                 file.write(content)
  284.             print(colored(f"Information saved to {filename}", 'green'))
  285.         except Exception as file_err:
  286.             print(colored(f"Error saving to file: {file_err}", 'red'))
  287.  
  288.     def ssl_analysis(self, url):
  289.         # Placeholder for SSL analysis logic
  290.         pass
  291.  
  292.     def check_robots_txt(self, url):
  293.         # Placeholder for robots.txt checking logic
  294.         pass
  295.  
  296.     def check_sitemap_xml(self, url):
  297.         # Placeholder for sitemap.xml checking logic
  298.         pass
  299.  
  300.     def dns_lookup(self, domain):
  301.         # Placeholder for DNS lookup logic
  302.         pass
  303.  
  304.     def perform_cve_scan(self, url):
  305.         # Placeholder for CVE scan logic
  306.         pass
  307.  
  308.     def perform_web_search(self, query, search_engine="google"):
  309.         search_url = f"https://{search_engine}.com/search?q={query}"
  310.         response = requests.get(search_url)
  311.         soup = BeautifulSoup(response.text, 'html.parser')
  312.         return soup.find_all('div', class_='tF2Cxc')
  313.  
  314.     def extract_data(self, result):
  315.         title = result.find('h3').text
  316.         link = result.find('a')['href']
  317.         content = result.find('div', class_='B6fmyf').text  # Adjust class based on actual structure
  318.         return {'Title': title, 'Link': link, 'Content': content}
  319.  
  320.     def display_data(self, data_list):
  321.         for data in data_list:
  322.             print(f"Title: {data['Title']}\nLink: {data['Link']}\nContent: {data['Content']}\n")
  323.  
  324.     def extract_links(self, url):
  325.         response = requests.get(url)
  326.         soup = BeautifulSoup(response.text, 'html.parser')
  327.         links = soup.find_all('a', href=True)
  328.         return [urljoin(url, link['href']) for link in links]
  329.  
  330.     def run(self):
  331.         while True:
  332.             option = self.show_menu()
  333.             self.execute_option(option)
  334.  
  335.  
  336. if __name__ == "__main__":
  337.     ghostsec_toolset = GhostSecDeepScanToolSet()
  338.     ghostsec_toolset.run()
  339.  
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement