mirror of
https://github.com/discourse/discourse.git
synced 2024-11-23 07:38:01 +08:00
f0fea5991f
Selenium uses Keep-Alive since version 3.141, so the net-http-persistent gem shouldn't be needed anymore.
264 lines
6.7 KiB
Ruby
Executable File
264 lines
6.7 KiB
Ruby
Executable File
#!/usr/bin/env ruby
|
|
# frozen_string_literal: true
|
|
|
|
require "bundler/inline"
|
|
|
|
gemfile(true) do
|
|
source "https://rubygems.org"
|
|
|
|
gem "nokogiri"
|
|
gem "webdrivers"
|
|
end
|
|
|
|
require "fileutils"
|
|
require "nokogiri"
|
|
require "optparse"
|
|
require "webdrivers"
|
|
require "set"
|
|
require "yaml"
|
|
|
|
DEFAULT_OUTPUT_PATH = "/shared/import/data"
|
|
|
|
def driver
|
|
@driver ||= begin
|
|
chrome_args = ["headless", "disable-gpu"]
|
|
chrome_args << "no-sandbox" if inside_container?
|
|
options = Selenium::WebDriver::Chrome::Options.new(args: chrome_args)
|
|
Selenium::WebDriver.for(:chrome, options: options)
|
|
end
|
|
end
|
|
|
|
def inside_container?
|
|
File.foreach("/proc/1/cgroup") do |line|
|
|
return true if line.include?("docker")
|
|
end
|
|
|
|
false
|
|
end
|
|
|
|
MAX_GET_RETRIES = 5
|
|
MAX_FIND_RETRIES = 3
|
|
|
|
def get(url)
|
|
begin
|
|
retries ||= 0
|
|
driver.get(url)
|
|
rescue Net::ReadTimeout
|
|
sleep retries
|
|
retry if (retries += 1) < MAX_GET_RETRIES
|
|
end
|
|
end
|
|
|
|
def extract(css, parent_element = driver)
|
|
begin
|
|
retries ||= 0
|
|
parent_element.find_elements(css: css).map { |element| yield(element) }
|
|
rescue Net::ReadTimeout, Selenium::WebDriver::Error::StaleElementReferenceError
|
|
sleep retries
|
|
retry if (retries += 1) < MAX_FIND_RETRIES
|
|
end
|
|
end
|
|
|
|
def find(css, parent_element = driver)
|
|
begin
|
|
retries ||= 0
|
|
parent_element.find_element(css: css)
|
|
rescue Net::ReadTimeout, Selenium::WebDriver::Error::ElementNotVisibleError
|
|
sleep retries
|
|
retry if (retries += 1) < MAX_FIND_RETRIES
|
|
end
|
|
end
|
|
|
|
def crawl_categories
|
|
1.step(nil, 100).each do |start|
|
|
url = "https://groups.google.com/forum/?_escaped_fragment_=categories/#{@groupname}[#{start}-#{start + 99}]"
|
|
get(url)
|
|
|
|
topic_urls = extract(".subject a[href*='#{@groupname}']") { |a| a["href"].sub("/d/topic/", "/forum/?_escaped_fragment_=topic/") }
|
|
break if topic_urls.size == 0
|
|
|
|
topic_urls.each { |topic_url| crawl_topic(topic_url) }
|
|
end
|
|
end
|
|
|
|
def crawl_topic(url)
|
|
if @scraped_topic_urls.include?(url)
|
|
puts "Skipping #{url}"
|
|
return
|
|
end
|
|
|
|
puts "Scraping #{url}"
|
|
get(url)
|
|
|
|
extract(".subject a[href*='#{@groupname}']") do |a|
|
|
[
|
|
a["href"].sub("/d/msg/", "/forum/message/raw?msg="),
|
|
a["title"].empty?
|
|
]
|
|
end.each { |msg_url, might_be_deleted| crawl_message(msg_url, might_be_deleted) }
|
|
|
|
@scraped_topic_urls << url
|
|
rescue
|
|
puts "Failed to scrape topic at #{url}"
|
|
raise
|
|
end
|
|
|
|
def crawl_message(url, might_be_deleted)
|
|
get(url)
|
|
|
|
filename = File.join(@path, "#{url[/#{@groupname}\/(.+)/, 1].sub("/", "-")}.eml")
|
|
content = find("pre")["innerText"]
|
|
|
|
if !@first_message_checked
|
|
@first_message_checked = true
|
|
|
|
if content.match?(/From:.*\.\.\.@.*/i) && !@force_import
|
|
exit_with_error(<<~MSG)
|
|
It looks like you do not have permissions to see email addresses. Aborting.
|
|
Use the --force option to import anyway.
|
|
MSG
|
|
end
|
|
end
|
|
|
|
File.write(filename, content)
|
|
rescue Selenium::WebDriver::Error::NoSuchElementError
|
|
raise unless might_be_deleted
|
|
puts "Message might be deleted. Skipping #{url}"
|
|
rescue
|
|
puts "Failed to scrape message at #{url}"
|
|
raise
|
|
end
|
|
|
|
def login
|
|
puts "Logging in..."
|
|
get("https://www.google.com/accounts/Login")
|
|
|
|
sleep(1)
|
|
email_element = wait_for_element("input[type='email']")
|
|
exit_with_error("Failed to detect 'email' input on login page") if !email_element
|
|
|
|
driver.action.move_to(email_element)
|
|
email_element.send_keys(@email)
|
|
email_element.send_keys("\n")
|
|
|
|
sleep(1)
|
|
password_element = wait_for_element("input[type='password']")
|
|
exit_with_error("Failed to detect 'password' input on login page") if !password_element
|
|
|
|
driver.action.move_to(password_element)
|
|
password_element.send_keys(@password)
|
|
password_element.send_keys("\n")
|
|
|
|
sleep(1)
|
|
|
|
if driver.current_url.include?("challenge")
|
|
puts "", "2-Step Verification is required."
|
|
puts "Unlock on your phone and press Enter"
|
|
puts "or enter the code from your authenticator app"
|
|
puts "or enter the code you received via SMS (without the G- prefix)"
|
|
|
|
print "Enter code: "
|
|
|
|
code = gets.chomp
|
|
|
|
if code.empty?
|
|
# Verification via phone?
|
|
begin
|
|
wait_for_url { |url| !url.include?("challenge") }
|
|
rescue Selenium::WebDriver::Error::TimeOutError
|
|
exit_with_error("Failed to login. Did you tap 'Yes' on your phone to allow the login?")
|
|
end
|
|
else
|
|
code_element = wait_for_element("input[type='tel']")
|
|
exit_with_error("Failed to detect 'code' input on login page") if !code_element
|
|
|
|
code_element.send_keys(code)
|
|
code_element.send_keys("\n")
|
|
|
|
begin
|
|
wait_for_url { |url| !url.include?("challenge") }
|
|
rescue Selenium::WebDriver::Error::TimeOutError
|
|
exit_with_error("Failed to login. Wrong code?")
|
|
end
|
|
end
|
|
end
|
|
|
|
sleep(1)
|
|
user_element = wait_for_element("a[aria-label*='#{@email}']")
|
|
exit_with_error("Failed to login") if !user_element
|
|
end
|
|
|
|
def wait_for_url
|
|
wait = Selenium::WebDriver::Wait.new(timeout: 5)
|
|
wait.until { yield(driver.current_url) }
|
|
end
|
|
|
|
def wait_for_element(css)
|
|
wait = Selenium::WebDriver::Wait.new(timeout: 5)
|
|
wait.until { driver.find_element(css: css).displayed? }
|
|
find(css)
|
|
rescue Selenium::WebDriver::Error::TimeOutError
|
|
nil
|
|
end
|
|
|
|
def exit_with_error(*messages)
|
|
STDERR.puts messages
|
|
exit 1
|
|
end
|
|
|
|
def crawl
|
|
start_time = Time.now
|
|
status_filename = File.join(@path, "status.yml")
|
|
@scraped_topic_urls = File.exists?(status_filename) ? YAML.load_file(status_filename) : Set.new
|
|
|
|
login
|
|
|
|
begin
|
|
crawl_categories
|
|
ensure
|
|
File.write(status_filename, @scraped_topic_urls.to_yaml)
|
|
end
|
|
|
|
elapsed = Time.now - start_time
|
|
puts "", "", "Done (%02dh %02dmin %02dsec)" % [elapsed / 3600, elapsed / 60 % 60, elapsed % 60]
|
|
end
|
|
|
|
def parse_arguments
|
|
puts ""
|
|
|
|
@force_import = false
|
|
|
|
parser = OptionParser.new do |opts|
|
|
opts.banner = "Usage: google_groups.rb [options]"
|
|
|
|
opts.on("-e", "--email EMAIL", "email address of group admin or manager") { |v| @email = v }
|
|
opts.on("-p", "--password PASSWORD", "password of group admin or manager") { |v| @password = v }
|
|
opts.on("-g", "--groupname GROUPNAME") { |v| @groupname = v }
|
|
opts.on("--path PATH", "output path for emails") { |v| @path = v }
|
|
opts.on("-f", "--force", "force import when user isn't allowed to see email addresses") { @force_import = true }
|
|
opts.on("-h", "--help") do
|
|
puts opts
|
|
exit
|
|
end
|
|
end
|
|
|
|
begin
|
|
parser.parse!
|
|
rescue OptionParser::ParseError => e
|
|
exit_with_error(e.message, "", parser)
|
|
end
|
|
|
|
mandatory = [:email, :password, :groupname]
|
|
missing = mandatory.select { |name| instance_variable_get("@#{name}").nil? }
|
|
|
|
if missing.any?
|
|
exit_with_error("Missing arguments: #{missing.join(', ')}", "", parser)
|
|
end
|
|
|
|
@path = File.join(DEFAULT_OUTPUT_PATH, @groupname) if @path.nil?
|
|
FileUtils.mkpath(@path)
|
|
end
|
|
|
|
parse_arguments
|
|
crawl
|