# frozen_string_literal: true begin require_relative "base" require "sqlite3" require "json" rescue LoadError STDERR.puts "", "ERROR: Failed to load required gems.", "", "You need to enable the `generic_import` group in your Gemfile.", "Execute the following command to do so:", "", "\tbundle config set --local with generic_import && bundle install", "" exit 1 end class BulkImport::Generic < BulkImport::Base AVATAR_DIRECTORY = ENV["AVATAR_DIRECTORY"] UPLOAD_DIRECTORY = ENV["UPLOAD_DIRECTORY"] def initialize(db_path, uploads_db_path = nil) super() @source_db = create_connection(db_path) @uploads_db = create_connection(uploads_db_path) if uploads_db_path end def start run # will call execute, and then "complete" the migration # Now that the migration is complete, do some more work: Discourse::Application.load_tasks puts "running 'import:ensure_consistency' rake task." Rake::Task["import:ensure_consistency"].invoke end def execute enable_required_plugins import_site_settings import_uploads # needs to happen before users, because keeping group names is more important than usernames import_groups import_users import_user_emails import_user_profiles import_user_options import_user_fields import_user_field_values import_single_sign_on_records import_muted_users import_user_histories import_user_notes import_user_note_counts import_user_followers import_user_avatars update_uploaded_avatar_id import_group_members import_tag_groups import_tags import_tag_users import_categories import_category_custom_fields import_category_tag_groups import_category_permissions import_topics import_posts import_post_custom_fields import_polls import_poll_options import_poll_votes import_topic_tags import_topic_allowed_users import_likes import_votes import_answers import_gamification_scores import_post_events import_badge_groupings import_badges import_user_badges import_upload_references import_optimized_images import_topic_users update_topic_users import_user_stats import_permalink_normalizations import_permalinks end def execute_after import_category_about_topics @source_db.close @uploads_db.close if @uploads_db end def enable_required_plugins puts "", "Enabling required plugins..." required_plugin_names = @source_db.get_first_value(<<~SQL)&.then(&JSON.method(:parse)) SELECT value FROM config WHERE name = 'enable_required_plugins' SQL return if required_plugin_names.blank? plugins_by_name = Discourse.plugins_by_name required_plugin_names.each do |plugin_name| if (plugin = plugins_by_name[plugin_name]) if !plugin.enabled? && plugin.configurable? SiteSetting.set(plugin.enabled_site_setting, true) end puts " #{plugin_name} plugin enabled" else puts " ERROR: The #{plugin_name} plugin is required, but not installed." exit 1 end end end def import_site_settings puts "", "Importing site settings..." rows = query(<<~SQL) SELECT name, value, action FROM site_settings ORDER BY ROWID SQL all_settings = SiteSetting.all_settings rows.each do |row| name = row["name"].to_sym setting = all_settings.find { |s| s[:setting] == name } next unless setting case row["action"] when "update" SiteSetting.set_and_log(name, row["value"]) when "append" raise "Cannot append to #{name} setting" if setting[:type] != "list" items = (SiteSetting.get(name) || "").split("|") items << row["value"] unless items.include?(row["value"]) SiteSetting.set_and_log(name, items.join("|")) end end rows.close end def import_categories puts "", "Importing categories..." categories = query(<<~SQL) WITH RECURSIVE tree AS ( SELECT c.id, c.parent_category_id, c.name, c.description, c.color, c.text_color, c.read_restricted, c.slug, c.existing_id, c.position, c.logo_upload_id, 0 AS level FROM categories c WHERE c.parent_category_id IS NULL UNION ALL SELECT c.id, c.parent_category_id, c.name, c.description, c.color, c.text_color, c.read_restricted, c.slug, c.existing_id, c.position, c.logo_upload_id, tree.level + 1 AS level FROM categories c, tree WHERE c.parent_category_id = tree.id ) SELECT id, parent_category_id, name, description, color, text_color, read_restricted, slug, existing_id, logo_upload_id, COALESCE(position, ROW_NUMBER() OVER (PARTITION BY parent_category_id ORDER BY parent_category_id NULLS FIRST, name)) AS position FROM tree ORDER BY level, position, id SQL create_categories(categories) do |row| next if category_id_from_imported_id(row["id"]).present? { imported_id: row["id"], existing_id: row["existing_id"], name: row["name"], description: row["description"], parent_category_id: row["parent_category_id"] ? category_id_from_imported_id(row["parent_category_id"]) : nil, slug: row["slug"], read_restricted: row["read_restricted"], uploaded_logo_id: row["logo_upload_id"] ? upload_id_from_original_id(row["logo_upload_id"]) : nil, } end categories.close end def import_category_about_topics puts "", %|Creating "About..." topics for categories...| start_time = Time.now Category.ensure_consistency! Site.clear_cache categories = query(<<~SQL) SELECT id, about_topic_title FROM categories WHERE about_topic_title IS NOT NULL ORDER BY id SQL categories.each do |row| if (about_topic_title = row["about_topic_title"]).present? if (category_id = category_id_from_imported_id(row["id"])) topic = Category.find(category_id).topic topic.title = about_topic_title topic.save!(validate: false) end end end categories.close puts " Creating took #{(Time.now - start_time).to_i} seconds." end def import_category_custom_fields puts "", "Importing category custom fields..." category_custom_fields = query(<<~SQL) SELECT * FROM category_custom_fields ORDER BY category_id, name SQL field_names = query("SELECT DISTINCT name FROM category_custom_fields") { _1.map { |row| row["name"] } } existing_category_custom_fields = CategoryCustomField.where(name: field_names).pluck(:category_id, :name).to_set create_category_custom_fields(category_custom_fields) do |row| category_id = category_id_from_imported_id(row["category_id"]) next if category_id.nil? next if existing_category_custom_fields.include?([category_id, row["name"]]) { category_id: category_id, name: row["name"], value: row["value"] } end category_custom_fields.close end def import_category_tag_groups puts "", "Importing category tag groups..." category_tag_groups = query(<<~SQL) SELECT c.id AS category_id, t.value AS tag_group_id FROM categories c, JSON_EACH(c.tag_group_ids) t ORDER BY category_id, tag_group_id SQL existing_category_tag_groups = CategoryTagGroup.pluck(:category_id, :tag_group_id).to_set create_category_tag_groups(category_tag_groups) do |row| category_id = category_id_from_imported_id(row["category_id"]) tag_group_id = @tag_group_mapping[row["tag_group_id"]] next unless category_id && tag_group_id next unless existing_category_tag_groups.add?([category_id, tag_group_id]) { category_id: category_id, tag_group_id: tag_group_id } end category_tag_groups.close end def import_category_permissions puts "", "Importing category permissions..." permissions = query(<<~SQL) SELECT c.id AS category_id, p.value -> 'group_id' AS group_id, p.value -> 'permission_type' AS permission_type FROM categories c, JSON_EACH(c.permissions) p SQL existing_category_group_ids = CategoryGroup.pluck(:category_id, :group_id).to_set create_category_groups(permissions) do |row| category_id = category_id_from_imported_id(row["category_id"]) group_id = group_id_from_imported_id(row["group_id"]) next if existing_category_group_ids.include?([category_id, group_id]) { category_id: category_id, group_id: group_id, permission_type: row["permission_type"] } end permissions.close end def import_groups puts "", "Importing groups..." groups = query(<<~SQL) SELECT * FROM groups ORDER BY id SQL create_groups(groups) do |row| next if group_id_from_imported_id(row["id"]).present? { imported_id: row["id"], name: row["name"], full_name: row["full_name"], visibility_level: row["visibility_level"], members_visibility_level: row["members_visibility_level"], mentionable_level: row["mentionable_level"], messageable_level: row["messageable_level"], } end groups.close end def import_group_members puts "", "Importing group members..." group_members = query(<<~SQL) SELECT * FROM group_members ORDER BY ROWID SQL existing_group_user_ids = GroupUser.pluck(:group_id, :user_id).to_set create_group_users(group_members) do |row| group_id = group_id_from_imported_id(row["group_id"]) user_id = user_id_from_imported_id(row["user_id"]) next if existing_group_user_ids.include?([group_id, user_id]) { group_id: group_id, user_id: user_id } end group_members.close end def import_users puts "", "Importing users..." users = query(<<~SQL) SELECT * FROM users ORDER BY id SQL create_users(users) do |row| next if user_id_from_imported_id(row["id"]).present? sso_record = JSON.parse(row["sso_record"]) if row["sso_record"].present? if row["suspension"].present? suspension = JSON.parse(row["suspension"]) suspended_at = suspension["suspended_at"] suspended_till = suspension["suspended_till"] end if row["anonymized"] == 1 while true anon_suffix = (SecureRandom.random_number * 100_000_000).to_i break if !@anonymized_user_suffixes.include?(anon_suffix) end row["username"] = "anon_#{anon_suffix}" row["email"] = "#{row["username"]}#{UserAnonymizer::EMAIL_SUFFIX}" row["name"] = nil row["registration_ip_address"] = nil @anonymized_user_suffixes << anon_suffix end { imported_id: row["id"], username: row["username"], original_username: row["original_username"], name: row["name"], email: row["email"], external_id: sso_record&.fetch("external_id", nil), created_at: to_datetime(row["created_at"]), last_seen_at: to_datetime(row["last_seen_at"]), admin: row["admin"], moderator: row["moderator"], suspended_at: suspended_at, suspended_till: suspended_till, registration_ip_address: row["registration_ip_address"], } end users.close end def import_user_emails puts "", "Importing user emails..." existing_user_ids = UserEmail.pluck(:user_id).to_set users = query(<<~SQL) SELECT id, email, created_at FROM users ORDER BY id SQL create_user_emails(users) do |row| user_id = user_id_from_imported_id(row["id"]) next if user_id && existing_user_ids.include?(user_id) { user_id: user_id, email: row["email"], created_at: to_datetime(row["created_at"]) } end users.close end def import_user_profiles puts "", "Importing user profiles..." users = query(<<~SQL) SELECT id, bio FROM users WHERE bio IS NOT NULL ORDER BY id SQL existing_user_ids = UserProfile.pluck(:user_id).to_set create_user_profiles(users) do |row| user_id = user_id_from_imported_id(row["id"]) next if user_id && existing_user_ids.include?(user_id) { user_id: user_id, bio_raw: row["bio"] } end users.close end def import_user_options puts "", "Importing user options..." users = query(<<~SQL) SELECT id, timezone, email_level, email_messages_level, email_digests FROM users WHERE timezone IS NOT NULL ORDER BY id SQL existing_user_ids = UserOption.pluck(:user_id).to_set create_user_options(users) do |row| user_id = user_id_from_imported_id(row["id"]) next if user_id && existing_user_ids.include?(user_id) { user_id: user_id, timezone: row["timezone"], email_level: row["email_level"], email_messages_level: row["email_messages_level"], email_digests: row["email_digests"], } end users.close end def import_user_fields puts "", "Importing user fields..." user_fields = query(<<~SQL) SELECT * FROM user_fields ORDER BY ROWID SQL existing_user_field_names = UserField.pluck(:name).to_set user_fields.each do |row| next if existing_user_field_names.include?(row["name"]) options = row.delete("options") field = UserField.create!(row) if options.present? JSON.parse(options).each { |option| field.user_field_options.create!(value: option) } end end user_fields.close end def import_user_field_values puts "", "Importing user field values..." discourse_field_mapping = UserField.pluck(:name, :id).to_h user_fields = query("SELECT id, name FROM user_fields") field_id_mapping = user_fields .map do |row| discourse_field_id = discourse_field_mapping[row["name"]] field_name = "#{User::USER_FIELD_PREFIX}#{discourse_field_id}" [row["id"], field_name] end .to_h user_fields.close # TODO make restriction to non-anonymized users configurable values = query(<<~SQL) SELECT v.* FROM user_field_values v JOIN users u ON v.user_id = u.id WHERE u.anonymized = FALSE SQL existing_user_fields = UserCustomField.where("name LIKE '#{User::USER_FIELD_PREFIX}%'").pluck(:user_id, :name).to_set create_user_custom_fields(values) do |row| user_id = user_id_from_imported_id(row["user_id"]) field_name = field_id_mapping[row["field_id"]] next if user_id && field_name && existing_user_fields.include?([user_id, field_name]) { user_id: user_id, name: field_name, value: row["value"] } end values.close end def import_single_sign_on_records puts "", "Importing SSO records..." users = query(<<~SQL) SELECT id, sso_record FROM users WHERE sso_record IS NOT NULL ORDER BY id SQL existing_user_ids = SingleSignOnRecord.pluck(:user_id).to_set create_single_sign_on_records(users) do |row| user_id = user_id_from_imported_id(row["id"]) next if user_id && existing_user_ids.include?(user_id) sso_record = JSON.parse(row["sso_record"], symbolize_names: true) sso_record[:user_id] = user_id sso_record end users.close end def import_topics puts "", "Importing topics..." topics = query(<<~SQL) SELECT * FROM topics ORDER BY id SQL create_topics(topics) do |row| unless row["category_id"] && (category_id = category_id_from_imported_id(row["category_id"])) next end next if topic_id_from_imported_id(row["id"]).present? { archetype: row["private_message"] ? Archetype.private_message : Archetype.default, imported_id: row["id"], title: row["title"], user_id: user_id_from_imported_id(row["user_id"]), created_at: to_datetime(row["created_at"]), category_id: category_id, closed: to_boolean(row["closed"]), views: row["views"], subtype: row["subtype"], } end topics.close end def import_topic_allowed_users # FIXME: This is not working correctly because it imports only the first user from the list! # Groups are ignored completely. And there is no check for existing records. puts "", "Importing topic_allowed_users..." topics = query(<<~SQL) SELECT * FROM topics WHERE private_message IS NOT NULL ORDER BY id SQL added = 0 create_topic_allowed_users(topics) do |row| next unless (topic_id = topic_id_from_imported_id(row["id"])) imported_user_id = JSON.parse(row["private_message"])["user_ids"].first user_id = user_id_from_imported_id(imported_user_id) added += 1 { # FIXME: missing imported_id topic_id: topic_id, user_id: user_id, } end topics.close puts " Added #{added} topic_allowed_users records." end def import_posts puts "", "Importing posts..." posts = query(<<~SQL) SELECT * FROM posts ORDER BY topic_id, id SQL group_names = Group.pluck(:id, :name).to_h # TODO: Investigate feasibility of loading all users on large sites user_names = User.pluck(:id, :username).to_h create_posts(posts) do |row| next if row["raw"].blank? next unless (topic_id = topic_id_from_imported_id(row["topic_id"])) next if post_id_from_imported_id(row["id"]).present? # TODO Ensure that we calculate the `like_count` if the column is empty, but the DB contains likes. # Otherwise #import_user_stats will not be able to calculate the correct `likes_received` value. { imported_id: row["id"], topic_id: topic_id, user_id: user_id_from_imported_id(row["user_id"]), created_at: to_datetime(row["created_at"]), raw: post_raw(row, group_names, user_names), like_count: row["like_count"], reply_to_post_number: row["reply_to_post_id"] ? post_number_from_imported_id(row["reply_to_post_id"]) : nil, } end posts.close end def post_raw(row, group_names, user_names) raw = row["raw"] placeholders = row["placeholders"]&.then { |json| JSON.parse(json) } if (polls = placeholders&.fetch("polls", nil)) poll_mapping = polls.map { |poll| [poll["poll_id"], poll["placeholder"]] }.to_h poll_details = query(<<~SQL, { post_id: row["id"] }) SELECT p.*, ROW_NUMBER() OVER (PARTITION BY p.post_id, p.name ORDER BY p.id) AS seq, JSON_GROUP_ARRAY(DISTINCT TRIM(po.text)) AS options FROM polls p JOIN poll_options po ON p.id = po.poll_id WHERE p.post_id = :post_id ORDER BY p.id, po.position, po.id SQL poll_details.each do |poll| if (placeholder = poll_mapping[poll["id"]]) raw.gsub!(placeholder, poll_bbcode(poll)) end end poll_details.close end if (mentions = placeholders&.fetch("mentions", nil)) mentions.each do |mention| name = if mention["type"] == "user" user_names[user_id_from_imported_id(mention["id"])] elsif mention["type"] == "group" group_names[group_id_from_imported_id(mention["id"])] end puts "#{mention["type"]} not found -- #{mention["id"]}" unless name raw.gsub!(mention["placeholder"], "@#{name}") end end if (event = placeholders&.fetch("event", nil)) event_details = @source_db.get_first_row(<<~SQL, { event_id: event["event_id"] }) SELECT * FROM events WHERE id = :event_id SQL raw.gsub!(event["placeholder"], event_bbcode(event_details)) if event_details end if row["upload_ids"].present? && @uploads_db upload_ids = JSON.parse(row["upload_ids"]) upload_ids_placeholders = (["?"] * upload_ids.size).join(",") query( "SELECT id, markdown FROM uploads WHERE id IN (#{upload_ids_placeholders})", upload_ids, db: @uploads_db, ).tap do |result_set| result_set.each { |upload| raw.gsub!("[upload|#{upload["id"]}]", upload["markdown"] || "") } result_set.close end end raw end def process_raw(original_raw) original_raw end def poll_name(row) name = +(row["name"] || "poll") name << "-#{row["seq"]}" if row["seq"] > 1 name end def poll_bbcode(row) return unless defined?(::Poll) name = poll_name(row) type = ::Poll.types.key(row["type"]) regular_type = type == ::Poll.types[:regular] number_type = type == ::Poll.types[:number] result_visibility = ::Poll.results.key(row["results"]) min = row["min"] max = row["max"] step = row["step"] visibility = ::Poll.visibilities.key(row["visibility"]) chart_type = ::Poll.chart_types.key(row["chart_type"]) groups = row["groups"] auto_close = to_datetime(row["close_at"]) title = row["title"] options = JSON.parse(row["options"]) text = +"[poll" text << " name=#{name}" if name != "poll" text << " type=#{type}" text << " results=#{result_visibility}" text << " min=#{min}" if min && !regular_type text << " max=#{max}" if max && !regular_type text << " step=#{step}" if step && !number_type text << " public=true" if visibility == Poll.visibilities[:everyone] text << " chartType=#{chart_type}" if chart_type.present? && !regular_type text << " groups=#{groups.join(",")}" if groups.present? text << " close=#{auto_close.utc.iso8601}" if auto_close text << "]\n" text << "# #{title}\n" if title.present? text << options.map { |o| "* #{o}" }.join("\n") if options.present? && !number_type text << "\n[/poll]\n" text end def event_bbcode(event) return unless defined?(::DiscoursePostEvent) starts_at = to_datetime(event["starts_at"]) ends_at = to_datetime(event["ends_at"]) status = ::DiscoursePostEvent::Event.statuses[event["status"]].to_s name = if (name = event["name"].presence) name.ljust(::DiscoursePostEvent::Event::MIN_NAME_LENGTH, ".").truncate( ::DiscoursePostEvent::Event::MAX_NAME_LENGTH, ) end url = event["url"] custom_fields = event["custom_fields"] ? JSON.parse(event["custom_fields"]) : nil text = +"[event" text << %{ start="#{starts_at.utc.strftime("%Y-%m-%d %H:%M")}"} if starts_at text << %{ end="#{ends_at.utc.strftime("%Y-%m-%d %H:%M")}"} if ends_at text << %{ timezone="UTC"} text << %{ status="#{status}"} if status text << %{ name="#{name}"} if name text << %{ url="#{url}"} if url custom_fields.each { |key, value| text << %{ #{key}="#{value}"} } if custom_fields text << "]\n" text << "[/event]\n" text end def import_post_custom_fields puts "", "Importing post custom fields..." post_custom_fields = query(<<~SQL) SELECT * FROM post_custom_fields ORDER BY post_id, name SQL field_names = query("SELECT DISTINCT name FROM post_custom_fields") { _1.map { |row| row["name"] } } existing_post_custom_fields = PostCustomField.where(name: field_names).pluck(:post_id, :name).to_set create_post_custom_fields(post_custom_fields) do |row| post_id = post_id_from_imported_id(row["post_id"]) next if post_id.nil? next if existing_post_custom_fields.include?([post_id, row["name"]]) { post_id: post_id, name: row["name"], value: row["value"] } end post_custom_fields.close end def import_polls unless defined?(::Poll) puts "", "Skipping polls, because the poll plugin is not installed." return end puts "", "Importing polls..." polls = query(<<~SQL) SELECT *, ROW_NUMBER() OVER (PARTITION BY post_id, name ORDER BY id) AS seq FROM polls ORDER BY id SQL create_polls(polls) do |row| next if poll_id_from_original_id(row["id"]).present? post_id = post_id_from_imported_id(row["post_id"]) next unless post_id { original_id: row["id"], post_id: post_id, name: poll_name(row), closed_at: to_datetime(row["closed_at"]), type: row["type"], status: row["status"], results: row["results"], visibility: row["visibility"], min: row["min"], max: row["max"], step: row["step"], anonymous_voters: row["anonymous_voters"], created_at: to_datetime(row["created_at"]), chart_type: row["chart_type"], groups: row["groups"], title: row["title"], } end polls.close puts "", "Importing polls into post custom fields..." polls = query(<<~SQL) SELECT post_id, MIN(created_at) AS created_at FROM polls GROUP BY post_id ORDER BY post_id SQL field_name = DiscoursePoll::HAS_POLLS value = "true" existing_fields = PostCustomField.where(name: field_name).pluck(:post_id).to_set create_post_custom_fields(polls) do |row| next unless (post_id = post_id_from_imported_id(row["post_id"])) next if existing_fields.include?(post_id) { post_id: post_id, name: field_name, value: value, created_at: to_datetime(row["created_at"]), } end polls.close end def import_poll_options unless defined?(::Poll) puts "", "Skipping polls, because the poll plugin is not installed." return end puts "", "Importing poll options..." poll_options = query(<<~SQL) SELECT poll_id, TRIM(text) AS text, MIN(created_at) AS created_at, GROUP_CONCAT(id) AS option_ids FROM poll_options GROUP BY 1, 2 ORDER BY poll_id, position, id SQL create_poll_options(poll_options) do |row| poll_id = poll_id_from_original_id(row["poll_id"]) next unless poll_id option_ids = row["option_ids"].split(",") option_ids.each { |option_id| next if poll_option_id_from_original_id(option_id).present? } { original_ids: option_ids, poll_id: poll_id, html: row["text"], created_at: to_datetime(row["created_at"]), } end poll_options.close end def import_poll_votes unless defined?(::Poll) puts "", "Skipping polls, because the poll plugin is not installed." return end puts "", "Importing poll votes..." poll_votes = query(<<~SQL) SELECT po.poll_id, pv.poll_option_id, pv.user_id, pv.created_at FROM poll_votes pv JOIN poll_options po ON pv.poll_option_id = po.id ORDER BY pv.poll_option_id, pv.user_id SQL existing_poll_votes = PollVote.pluck(:poll_option_id, :user_id).to_set create_poll_votes(poll_votes) do |row| poll_id = poll_id_from_original_id(row["poll_id"]) poll_option_id = poll_option_id_from_original_id(row["poll_option_id"]) user_id = user_id_from_imported_id(row["user_id"]) next unless poll_id && poll_option_id && user_id next unless existing_poll_votes.add?([poll_option_id, user_id]) { poll_id: poll_id, poll_option_id: poll_option_id, user_id: user_id, created_at: row["created_at"], } end poll_votes.close end def import_likes puts "", "Importing likes..." likes = query(<<~SQL) SELECT post_id, user_id, created_at FROM likes ORDER BY post_id, user_id SQL post_action_type_id = PostActionType.types[:like] existing_likes = PostAction.where(post_action_type_id: post_action_type_id).pluck(:post_id, :user_id).to_set create_post_actions(likes) do |row| post_id = post_id_from_imported_id(row["post_id"]) user_id = user_id_from_imported_id(row["user_id"]) next unless post_id && user_id next unless existing_likes.add?([post_id, user_id]) { post_id: post_id, user_id: user_id, post_action_type_id: post_action_type_id, created_at: to_datetime(row["created_at"]), } end likes.close puts "", "Updating like counts of topics..." start_time = Time.now DB.exec(<<~SQL) WITH likes AS ( SELECT topic_id, SUM(like_count) AS like_count FROM posts WHERE like_count > 0 GROUP BY topic_id ) UPDATE topics SET like_count = likes.like_count FROM likes WHERE topics.id = likes.topic_id AND topics.like_count <> likes.like_count SQL puts " Update took #{(Time.now - start_time).to_i} seconds." end def import_topic_users puts "", "Importing topic users..." topic_users = query(<<~SQL) SELECT * FROM topic_users ORDER BY user_id, topic_id SQL existing_topics = TopicUser.pluck(:topic_id).to_set create_topic_users(topic_users) do |row| user_id = user_id_from_imported_id(row["user_id"]) topic_id = topic_id_from_imported_id(row["topic_id"]) next unless user_id && topic_id next if existing_topics.include?(topic_id) { user_id: user_id, topic_id: topic_id, last_read_post_number: row["last_read_post_number"], last_visited_at: to_datetime(row["last_visited_at"]), first_visited_at: to_datetime(row["first_visited_at"]), notification_level: row["notification_level"], notifications_changed_at: to_datetime(row["notifications_changed_at"]), notifications_reason_id: row["notifications_reason_id"] || TopicUser.notification_reasons[:user_changed], total_msecs_viewed: row["total_msecs_viewed"] || 0, } end topic_users.close end def update_topic_users puts "", "Updating topic users..." start_time = Time.now params = { post_action_type_id: PostActionType.types[:like], msecs_viewed_per_post: 10_000, notification_level_topic_created: NotificationLevels.topic_levels[:watching], notification_level_posted: NotificationLevels.topic_levels[:tracking], reason_topic_created: TopicUser.notification_reasons[:created_topic], reason_posted: TopicUser.notification_reasons[:created_post], } DB.exec(<<~SQL, params) INSERT INTO topic_users (user_id, topic_id, posted, last_read_post_number, first_visited_at, last_visited_at, notification_level, notifications_changed_at, notifications_reason_id, total_msecs_viewed, last_posted_at) SELECT p.user_id, p.topic_id, TRUE AS posted, MAX(p.post_number) AS last_read_post_number, MIN(p.created_at) AS first_visited_at, MAX(p.created_at) AS last_visited_at, CASE WHEN MIN(p.post_number) = 1 THEN :notification_level_topic_created ELSE :notification_level_posted END AS notification_level, MIN(p.created_at) AS notifications_changed_at, CASE WHEN MIN(p.post_number) = 1 THEN :reason_topic_created ELSE :reason_posted END AS notifications_reason_id, MAX(p.post_number) * :msecs_viewed_per_post AS total_msecs_viewed, MAX(p.created_at) AS last_posted_at FROM posts p JOIN topics t ON p.topic_id = t.id WHERE p.user_id > 0 AND p.deleted_at IS NULL AND NOT p.hidden AND t.deleted_at IS NULL AND t.visible GROUP BY p.user_id, p.topic_id ON CONFLICT (user_id, topic_id) DO UPDATE SET posted = excluded.posted, last_read_post_number = GREATEST(topic_users.last_read_post_number, excluded.last_read_post_number), first_visited_at = LEAST(topic_users.first_visited_at, excluded.first_visited_at), last_visited_at = GREATEST(topic_users.last_visited_at, excluded.last_visited_at), notification_level = GREATEST(topic_users.notification_level, excluded.notification_level), notifications_changed_at = CASE WHEN COALESCE(excluded.notification_level, 0) > COALESCE(topic_users.notification_level, 0) THEN COALESCE(excluded.notifications_changed_at, topic_users.notifications_changed_at) ELSE topic_users.notifications_changed_at END, notifications_reason_id = CASE WHEN COALESCE(excluded.notification_level, 0) > COALESCE(topic_users.notification_level, 0) THEN COALESCE(excluded.notifications_reason_id, topic_users.notifications_reason_id) ELSE topic_users.notifications_reason_id END, total_msecs_viewed = CASE WHEN topic_users.total_msecs_viewed = 0 THEN excluded.total_msecs_viewed ELSE topic_users.total_msecs_viewed END, last_posted_at = GREATEST(topic_users.last_posted_at, excluded.last_posted_at) SQL DB.exec(<<~SQL, params) INSERT INTO topic_users (user_id, topic_id, last_read_post_number, first_visited_at, last_visited_at, total_msecs_viewed, liked) SELECT pa.user_id, p.topic_id, MAX(p.post_number) AS last_read_post_number, MIN(pa.created_at) AS first_visited_at, MAX(pa.created_at) AS last_visited_at, MAX(p.post_number) * :msecs_viewed_per_post AS total_msecs_viewed, TRUE AS liked FROM post_actions pa JOIN posts p ON pa.post_id = p.id JOIN topics t ON p.topic_id = t.id WHERE pa.post_action_type_id = :post_action_type_id AND pa.user_id > 0 AND pa.deleted_at IS NULL AND p.deleted_at IS NULL AND NOT p.hidden AND t.deleted_at IS NULL AND t.visible GROUP BY pa.user_id, p.topic_id ON CONFLICT (user_id, topic_id) DO UPDATE SET last_read_post_number = GREATEST(topic_users.last_read_post_number, excluded.last_read_post_number), first_visited_at = LEAST(topic_users.first_visited_at, excluded.first_visited_at), last_visited_at = GREATEST(topic_users.last_visited_at, excluded.last_visited_at), total_msecs_viewed = CASE WHEN topic_users.total_msecs_viewed = 0 THEN excluded.total_msecs_viewed ELSE topic_users.total_msecs_viewed END, liked = excluded.liked SQL puts " Updated topic users in #{(Time.now - start_time).to_i} seconds." end def import_user_stats puts "", "Importing user stats..." start_time = Time.now # TODO Merge with #update_user_stats from import.rake and check if there are privacy concerns # E.g. maybe we need to exclude PMs from the calculation? DB.exec(<<~SQL) WITH post_stats AS ( SELECT p.user_id, COUNT(p.id) AS post_count, MIN(p.created_at) AS first_post_created_at, SUM(like_count) AS likes_received FROM posts p GROUP BY p.user_id ), topic_stats AS ( SELECT t.user_id, COUNT(t.id) AS topic_count FROM topics t GROUP BY t.user_id ), like_stats AS ( SELECT pa.user_id, COUNT(*) AS likes_given FROM post_actions pa WHERE pa.post_action_type_id = 2 GROUP BY pa.user_id ) INSERT INTO user_stats (user_id, new_since, post_count, topic_count, first_post_created_at, likes_received, likes_given) SELECT u.id, u.created_at AS new_since, COALESCE(ps.post_count, 0) AS post_count, COALESCE(ts.topic_count, 0) AS topic_count, ps.first_post_created_at, COALESCE(ps.likes_received, 0) AS likes_received, COALESCE(ls.likes_given, 0) AS likes_given FROM users u LEFT JOIN post_stats ps ON u.id = ps.user_id LEFT JOIN topic_stats ts ON u.id = ts.user_id LEFT JOIN like_stats ls ON u.id = ls.user_id WHERE NOT EXISTS ( SELECT 1 FROM user_stats us WHERE us.user_id = u.id ) ON CONFLICT DO NOTHING SQL puts " Imported user stats in #{(Time.now - start_time).to_i} seconds." end def import_muted_users puts "", "Importing muted users..." muted_users = query(<<~SQL) SELECT * FROM muted_users SQL existing_user_ids = MutedUser.pluck(:user_id).to_set create_muted_users(muted_users) do |row| user_id = user_id_from_imported_id(row["user_id"]) next if user_id && existing_user_ids.include?(user_id) { user_id: user_id, muted_user_id: user_id_from_imported_id(row["muted_user_id"]) } end muted_users.close end def import_user_histories puts "", "Importing user histories..." user_histories = query(<<~SQL) SELECT id, JSON_EXTRACT(suspension, '$.reason') AS reason FROM users WHERE suspension IS NOT NULL SQL action_id = UserHistory.actions[:suspend_user] existing_user_ids = UserHistory.where(action: action_id).pluck(:target_user_id).to_set create_user_histories(user_histories) do |row| user_id = user_id_from_imported_id(row["id"]) next if user_id && existing_user_ids.include?(user_id) { action: action_id, acting_user_id: Discourse::SYSTEM_USER_ID, target_user_id: user_id, details: row["reason"], } end user_histories.close end def import_user_notes puts "", "Importing user notes..." unless defined?(::DiscourseUserNotes) puts " Skipping import of user notes because the plugin is not installed." return end user_notes = query(<<~SQL) SELECT user_id, JSON_GROUP_ARRAY(JSON_OBJECT('raw', raw, 'created_by', created_by_user_id, 'created_at', created_at)) AS note_json_text FROM user_notes GROUP BY user_id ORDER BY user_id, id SQL existing_user_ids = PluginStoreRow .where(plugin_name: "user_notes") .pluck(:key) .map { |key| key.delete_prefix("notes:").to_i } .to_set create_plugin_store_rows(user_notes) do |row| user_id = user_id_from_imported_id(row["user_id"]) next if !user_id || existing_user_ids.include?(user_id) notes = JSON.parse(row["note_json_text"], symbolize_names: true) notes.each do |note| note[:id] = SecureRandom.hex(16) note[:user_id] = user_id note[:created_by] = ( if note[:created_by] user_id_from_imported_id(note[:created_by]) else Discourse::SYSTEM_USER_ID end ) note[:created_at] = to_datetime(note[:created_at]) end { plugin_name: "user_notes", key: "notes:#{user_id}", type_name: "JSON", value: notes.to_json, } end user_notes.close end def import_user_note_counts puts "", "Importing user note counts..." unless defined?(::DiscourseUserNotes) puts " Skipping import of user notes because the plugin is not installed." return end user_note_counts = query(<<~SQL) SELECT user_id, COUNT(*) AS count FROM user_notes GROUP BY user_id ORDER BY user_id SQL existing_user_ids = UserCustomField.where(name: "user_notes_count").pluck(:user_id).to_set create_user_custom_fields(user_note_counts) do |row| user_id = user_id_from_imported_id(row["user_id"]) next if !user_id || existing_user_ids.include?(user_id) { user_id: user_id, name: "user_notes_count", value: row["count"].to_s } end user_note_counts.close end def import_user_followers puts "", "Importing user followers..." unless defined?(::Follow) puts " Skipping import of user followers because the plugin is not installed." return end user_followers = query(<<~SQL) SELECT * FROM user_followers ORDER BY user_id, follower_id SQL existing_followers = UserFollower.pluck(:user_id, :follower_id).to_set notification_level = Follow::Notification.levels[:watching] create_user_followers(user_followers) do |row| user_id = user_id_from_imported_id(row["user_id"]) follower_id = user_id_from_imported_id(row["follower_id"]) next if !user_id || !follower_id || existing_followers.include?([user_id, follower_id]) { user_id: user_id, follower_id: follower_id, level: notification_level, created_at: to_datetime(row["created_at"]), } end user_followers.close end def import_uploads return if !@uploads_db puts "", "Importing uploads..." uploads = query(<<~SQL, db: @uploads_db) SELECT id, upload FROM uploads WHERE upload IS NOT NULL ORDER BY rowid SQL create_uploads(uploads) do |row| next if upload_id_from_original_id(row["id"]).present? upload = JSON.parse(row["upload"], symbolize_names: true) upload[:original_id] = row["id"] upload end uploads.close end def import_optimized_images return if !@uploads_db puts "", "Importing optimized images..." optimized_images = query(<<~SQL, db: @uploads_db) SELECT oi.id AS upload_id, x.value AS optimized_image FROM optimized_images oi, JSON_EACH(oi.optimized_images) x WHERE optimized_images IS NOT NULL ORDER BY oi.rowid, x.value -> 'id' SQL DB.exec(<<~SQL) DELETE FROM optimized_images oi WHERE EXISTS ( SELECT 1 FROM migration_mappings mm WHERE mm.type = 1 AND mm.discourse_id::BIGINT = oi.upload_id ) SQL existing_optimized_images = OptimizedImage.pluck(:upload_id, :height, :width).to_set create_optimized_images(optimized_images) do |row| upload_id = upload_id_from_original_id(row["upload_id"]) next unless upload_id optimized_image = JSON.parse(row["optimized_image"], symbolize_names: true) unless existing_optimized_images.add?( [upload_id, optimized_image[:height], optimized_image[:width]], ) next end optimized_image[:upload_id] = upload_id optimized_image end optimized_images.close end def import_user_avatars return if !@uploads_db puts "", "Importing user avatars..." avatars = query(<<~SQL) SELECT id, avatar_upload_id FROM users WHERE avatar_upload_id IS NOT NULL ORDER BY id SQL existing_user_ids = UserAvatar.pluck(:user_id).to_set create_user_avatars(avatars) do |row| user_id = user_id_from_imported_id(row["id"]) upload_id = upload_id_from_original_id(row["avatar_upload_id"]) next if !upload_id || !user_id || existing_user_ids.include?(user_id) { user_id: user_id, custom_upload_id: upload_id } end avatars.close end def import_upload_references puts "", "Importing upload references for user avatars..." start_time = Time.now DB.exec(<<~SQL) INSERT INTO upload_references (upload_id, target_type, target_id, created_at, updated_at) SELECT ua.custom_upload_id, 'UserAvatar', ua.id, ua.created_at, ua.updated_at FROM user_avatars ua WHERE ua.custom_upload_id IS NOT NULL AND NOT EXISTS ( SELECT 1 FROM upload_references ur WHERE ur.upload_id = ua.custom_upload_id AND ur.target_type = 'UserAvatar' AND ur.target_id = ua.id ) ON CONFLICT DO NOTHING SQL puts " Import took #{(Time.now - start_time).to_i} seconds." puts "", "Importing upload references for categories..." start_time = Time.now DB.exec(<<~SQL) INSERT INTO upload_references (upload_id, target_type, target_id, created_at, updated_at) SELECT upload_id, 'Category', target_id, created_at, updated_at FROM ( SELECT uploaded_logo_id AS upload_id, id AS target_id, created_at, updated_at FROM categories WHERE uploaded_logo_id IS NOT NULL UNION SELECT uploaded_logo_dark_id AS upload_id, id AS target_id, created_at, updated_at FROM categories WHERE uploaded_logo_dark_id IS NOT NULL UNION SELECT uploaded_background_id AS upload_id, id AS target_id, created_at, updated_at FROM categories WHERE uploaded_background_id IS NOT NULL ) x WHERE NOT EXISTS ( SELECT 1 FROM upload_references ur WHERE ur.upload_id = x.upload_id AND ur.target_type = 'Category' AND ur.target_id = x.target_id ) ON CONFLICT DO NOTHING SQL puts " Import took #{(Time.now - start_time).to_i} seconds." puts "", "Importing upload references for badges..." start_time = Time.now DB.exec(<<~SQL) INSERT INTO upload_references (upload_id, target_type, target_id, created_at, updated_at) SELECT image_upload_id, 'Badge', id, created_at, updated_at FROM badges b WHERE image_upload_id IS NOT NULL AND NOT EXISTS ( SELECT 1 FROM upload_references ur WHERE ur.upload_id = b.image_upload_id AND ur.target_type = 'Badge' AND ur.target_id = b.id ) ON CONFLICT DO NOTHING SQL puts " Import took #{(Time.now - start_time).to_i} seconds." puts "", "Importing upload references for posts..." post_uploads = query(<<~SQL) SELECT p.id AS post_id, u.value AS upload_id FROM posts p, JSON_EACH(p.upload_ids) u WHERE upload_ids IS NOT NULL SQL existing_upload_references = UploadReference.where(target_type: "Post").pluck(:upload_id, :target_id).to_set create_upload_references(post_uploads) do |row| upload_id = upload_id_from_original_id(row["upload_id"]) post_id = post_id_from_imported_id(row["post_id"]) next unless upload_id && post_id next unless existing_upload_references.add?([upload_id, post_id]) { upload_id: upload_id, target_type: "Post", target_id: post_id } end post_uploads.close end def update_uploaded_avatar_id puts "", "Updating user's uploaded_avatar_id column..." start_time = Time.now DB.exec(<<~SQL) UPDATE users u SET uploaded_avatar_id = ua.custom_upload_id FROM user_avatars ua WHERE u.uploaded_avatar_id IS NULL AND u.id = ua.user_id AND ua.custom_upload_id IS NOT NULL SQL puts " Update took #{(Time.now - start_time).to_i} seconds." end def import_tag_groups puts "", "Importing tag groups..." SiteSetting.tags_listed_by_group = true @tag_group_mapping = {} tag_groups = query(<<~SQL) SELECT * FROM tag_groups ORDER BY id SQL tag_groups.each do |row| tag_group = TagGroup.find_or_create_by!(name: row["name"]) @tag_group_mapping[row["id"]] = tag_group.id if (permissions = row["permissions"]) tag_group.permissions = JSON .parse(permissions) .map do |p| group_id = p["existing_group_id"] || group_id_from_imported_id(p["group_id"]) group_id ? [group_id, p["permission_type"]] : nil end .compact tag_group.save! end end tag_groups.close end def import_tags puts "", "Importing tags..." SiteSetting.max_tag_length = 100 if SiteSetting.max_tag_length < 100 @tag_mapping = {} tags = query(<<~SQL) SELECT * FROM tags ORDER BY id SQL tags.each do |row| cleaned_tag_name = DiscourseTagging.clean_tag(row["name"]) tag = Tag.find_or_create_by!(name: cleaned_tag_name) @tag_mapping[row["id"]] = tag.id if row["tag_group_id"] TagGroupMembership.find_or_create_by!( tag_id: tag.id, tag_group_id: @tag_group_mapping[row["tag_group_id"]], ) end end tags.close end def import_topic_tags puts "", "Importing topic tags..." if !@tag_mapping puts " Skipping import of topic tags because tags have not been imported." return end topic_tags = query(<<~SQL) SELECT * FROM topic_tags ORDER BY topic_id, tag_id SQL existing_topic_tags = TopicTag.pluck(:topic_id, :tag_id).to_set create_topic_tags(topic_tags) do |row| topic_id = topic_id_from_imported_id(row["topic_id"]) tag_id = @tag_mapping[row["tag_id"]] next unless topic_id && tag_id next unless existing_topic_tags.add?([topic_id, tag_id]) { topic_id: topic_id, tag_id: tag_id } end topic_tags.close end def import_votes puts "", "Importing votes for posts..." unless defined?(::PostVoting) puts " Skipping import of votes for posts because the plugin is not installed." return end votes = query(<<~SQL) SELECT * FROM votes WHERE votable_type = 'Post' SQL votable_type = "Post" existing_votes = PostVotingVote.where(votable_type: votable_type).pluck(:user_id, :votable_id).to_set create_post_voting_votes(votes) do |row| user_id = user_id_from_imported_id(row["user_id"]) post_id = post_id_from_imported_id(row["votable_id"]) next unless user_id && post_id next unless existing_votes.add?([user_id, post_id]) { user_id: user_id, direction: row["direction"], votable_type: votable_type, votable_id: post_id, created_at: to_datetime(row["created_at"]), } end votes.close puts "", "Updating vote counts of posts..." start_time = Time.now DB.exec(<<~SQL) WITH votes AS ( SELECT votable_id AS post_id, SUM(CASE direction WHEN 'up' THEN 1 ELSE -1 END) AS vote_count FROM post_voting_votes GROUP BY votable_id ) UPDATE posts SET qa_vote_count = votes.vote_count FROM votes WHERE votes.post_id = posts.id AND votes.vote_count <> posts.qa_vote_count SQL puts " Update took #{(Time.now - start_time).to_i} seconds." end def import_answers puts "", "Importing solutions into post custom fields..." solutions = query(<<~SQL) SELECT * FROM solutions ORDER BY topic_id SQL field_name = "is_accepted_answer" value = "true" existing_fields = PostCustomField.where(name: field_name).pluck(:post_id).to_set create_post_custom_fields(solutions) do |row| next unless (post_id = post_id_from_imported_id(row["post_id"])) next unless existing_fields.add?(post_id) { post_id: post_id, name: field_name, value: value, created_at: to_datetime(row["created_at"]), } end puts "", "Importing solutions into topic custom fields..." solutions.reset field_name = "accepted_answer_post_id" existing_fields = TopicCustomField.where(name: field_name).pluck(:topic_id).to_set create_topic_custom_fields(solutions) do |row| post_id = post_id_from_imported_id(row["post_id"]) topic_id = topic_id_from_imported_id(row["topic_id"]) next unless post_id && topic_id next unless existing_fields.add?(topic_id) { topic_id: topic_id, name: field_name, value: post_id.to_s, created_at: to_datetime(row["created_at"]), } end puts "", "Importing solutions into user actions..." existing_fields = nil solutions.reset action_type = UserAction::SOLVED existing_actions = UserAction.where(action_type: action_type).pluck(:target_post_id).to_set create_user_actions(solutions) do |row| post_id = post_id_from_imported_id(row["post_id"]) next unless post_id && existing_actions.add?(post_id) topic_id = topic_id_from_imported_id(row["topic_id"]) user_id = user_id_from_imported_id(row["user_id"]) next unless topic_id && user_id acting_user_id = row["acting_user_id"] ? user_id_from_imported_id(row["acting_user_id"]) : nil { action_type: action_type, user_id: user_id, target_topic_id: topic_id, target_post_id: post_id, acting_user_id: acting_user_id || Discourse::SYSTEM_USER_ID, } end solutions.close end def import_gamification_scores puts "", "Importing gamification scores..." unless defined?(::DiscourseGamification) puts " Skipping import of gamification scores because the plugin is not installed." return end # TODO Make this configurable from_date = Date.today DiscourseGamification::GamificationLeaderboard.all.each do |leaderboard| leaderboard.update!(from_date: from_date) end scores = query(<<~SQL) SELECT * FROM gamification_score_events ORDER BY id SQL # TODO Better way of detecting existing scores? existing_scores = DiscourseGamification::GamificationScoreEvent.pluck(:user_id, :date).to_set create_gamification_score_events(scores) do |row| user_id = user_id_from_imported_id(row["user_id"]) next unless user_id date = to_date(row["date"]) || from_date next if existing_scores.include?([user_id, date]) { user_id: user_id, date: date, points: row["points"], description: row["description"], created_at: to_datetime(row["created_at"]), } end scores.close end def import_post_events puts "", "Importing events..." unless defined?(::DiscoursePostEvent) puts " Skipping import of events because the plugin is not installed." return end post_events = query(<<~SQL) SELECT * FROM events ORDER BY id SQL default_custom_fields = "{}" timezone = "UTC" public_group_invitees = "{#{::DiscoursePostEvent::Event::PUBLIC_GROUP}}" standalone_invitees = "{}" existing_events = DiscoursePostEvent::Event.pluck(:id).to_set create_post_events(post_events) do |row| post_id = post_id_from_imported_id(row["post_id"]) next if !post_id || existing_events.include?(post_id) { id: post_id, status: row["status"], original_starts_at: to_datetime(row["starts_at"]), original_ends_at: to_datetime(row["ends_at"]), name: row["name"], url: row["url"] ? row["url"][0..999] : nil, custom_fields: row["custom_fields"] || default_custom_fields, timezone: timezone, raw_invitees: ( if row["status"] == ::DiscoursePostEvent::Event.statuses[:public] public_group_invitees else standalone_invitees end ), } end puts "", "Importing event dates..." post_events.reset existing_events = DiscoursePostEvent::EventDate.pluck(:event_id).to_set create_post_event_dates(post_events) do |row| post_id = post_id_from_imported_id(row["post_id"]) next if !post_id || existing_events.include?(post_id) { event_id: post_id, starts_at: to_datetime(row["starts_at"]), ends_at: to_datetime(row["ends_at"]), } end puts "", "Importing topic event custom fields..." post_events.reset field_name = DiscoursePostEvent::TOPIC_POST_EVENT_STARTS_AT existing_fields = TopicCustomField.where(name: field_name).pluck(:topic_id).to_set create_topic_custom_fields(post_events) do |row| date = to_datetime(row["starts_at"]) next unless date topic_id = topic_id_from_imported_post_id(row["post_id"]) next if !topic_id || existing_fields.include?(topic_id) { topic_id: topic_id, name: field_name, value: date.utc.strftime("%Y-%m-%d %H:%M:%S") } end post_events.reset field_name = DiscoursePostEvent::TOPIC_POST_EVENT_ENDS_AT existing_fields = TopicCustomField.where(name: field_name).pluck(:topic_id).to_set create_topic_custom_fields(post_events) do |row| date = to_datetime(row["ends_at"]) next unless date topic_id = topic_id_from_imported_post_id(row["post_id"]) next if !topic_id || existing_fields.include?(topic_id) { topic_id: topic_id, name: field_name, value: date.utc.strftime("%Y-%m-%d %H:%M:%S") } end post_events.close end def import_tag_users puts "", "Importing tag users..." tag_users = query(<<~SQL) SELECT * FROM tag_users ORDER BY tag_id, user_id SQL existing_tag_users = TagUser.distinct.pluck(:user_id).to_set create_tag_users(tag_users) do |row| tag_id = @tag_mapping[row["tag_id"]] user_id = user_id_from_imported_id(row["user_id"]) next unless tag_id && user_id next if existing_tag_users.include?(user_id) { tag_id: tag_id, user_id: user_id, notification_level: row["notification_level"] } end tag_users.close end def import_badge_groupings puts "", "Importing badge groupings..." rows = query(<<~SQL) SELECT DISTINCT badge_group FROM badges ORDER BY badge_group SQL @badge_group_mapping = {} max_position = BadgeGrouping.maximum(:position) || 0 rows.each do |row| grouping = BadgeGrouping.find_or_create_by!(name: row["badge_group"]) do |bg| bg.position = max_position += 1 end @badge_group_mapping[row["badge_group"]] = grouping.id end rows.close end def import_badges puts "", "Importing badges..." badges = query(<<~SQL) SELECT * FROM badges ORDER BY id SQL existing_badge_names = Badge.pluck(:name).to_set create_badges(badges) do |row| next if badge_id_from_original_id(row["id"]).present? badge_name = row["name"] unless existing_badge_names.add?(badge_name) badge_name = badge_name + "_1" badge_name.next! until existing_badge_names.add?(badge_name) end { original_id: row["id"], name: badge_name, description: row["description"], badge_type_id: row["badge_type_id"], badge_grouping_id: @badge_group_mapping[row["badge_group"]], long_description: row["long_description"], image_upload_id: row["image_upload_id"] ? upload_id_from_original_id(row["image_upload_id"]) : nil, query: row["query"], } end badges.close end def import_user_badges puts "", "Importing user badges..." user_badges = query(<<~SQL) SELECT user_id, badge_id, granted_at, ROW_NUMBER() OVER (PARTITION BY user_id, badge_id ORDER BY granted_at) - 1 AS seq FROM user_badges ORDER BY user_id, badge_id, granted_at SQL existing_user_badges = UserBadge.distinct.pluck(:user_id, :badge_id, :seq).to_set create_user_badges(user_badges) do |row| user_id = user_id_from_imported_id(row["user_id"]) badge_id = badge_id_from_original_id(row["badge_id"]) seq = row["seq"] next unless user_id && badge_id next if existing_user_badges.include?([user_id, badge_id, seq]) { user_id: user_id, badge_id: badge_id, granted_at: to_datetime(row["granted_at"]), seq: seq } end user_badges.close puts "", "Updating badge grant counts..." start_time = Time.now DB.exec(<<~SQL) WITH grants AS ( SELECT badge_id, COUNT(*) AS grant_count FROM user_badges GROUP BY badge_id ) UPDATE badges SET grant_count = grants.grant_count FROM grants WHERE badges.id = grants.badge_id AND badges.grant_count <> grants.grant_count SQL puts " Update took #{(Time.now - start_time).to_i} seconds." end def import_permalink_normalizations puts "", "Importing permalink normalizations..." start_time = Time.now rows = query(<<~SQL) SELECT normalization FROM permalink_normalizations ORDER BY normalization SQL normalizations = SiteSetting.permalink_normalizations normalizations = normalizations.blank? ? [] : normalizations.split("|") rows.each do |row| normalization = row["normalization"] normalizations << normalization unless normalizations.include?(normalization) end SiteSetting.permalink_normalizations = normalizations.join("|") rows.close puts " Import took #{(Time.now - start_time).to_i} seconds." end def import_permalinks puts "", "Importing permalinks for topics..." rows = query(<<~SQL) SELECT id, old_relative_url FROM topics WHERE old_relative_url IS NOT NULL ORDER BY id SQL existing_permalinks = Permalink.where("topic_id IS NOT NULL").pluck(:topic_id).to_set create_permalinks(rows) do |row| topic_id = topic_id_from_imported_id(row["id"]) next if !topic_id || existing_permalinks.include?(topic_id) { url: row["old_relative_url"], topic_id: topic_id } end rows.close puts "", "Importing permalinks for posts..." rows = query(<<~SQL) SELECT id, old_relative_url FROM posts WHERE old_relative_url IS NOT NULL ORDER BY id SQL existing_permalinks = Permalink.where("post_id IS NOT NULL").pluck(:post_id).to_set create_permalinks(rows) do |row| post_id = post_id_from_imported_id(row["id"]) next if !post_id || existing_permalinks.include?(post_id) { url: row["old_relative_url"], post_id: post_id } end rows.close puts "", "Importing permalinks for categories..." rows = query(<<~SQL) SELECT id, old_relative_url FROM categories WHERE old_relative_url IS NOT NULL ORDER BY id SQL existing_permalinks = Permalink.where("category_id IS NOT NULL").pluck(:category_id).to_set create_permalinks(rows) do |row| category_id = category_id_from_imported_id(row["id"]) next if !category_id || existing_permalinks.include?(category_id) { url: row["old_relative_url"], category_id: category_id } end rows.close if @tag_mapping puts "", "Importing permalinks for tags..." rows = query(<<~SQL) SELECT id, old_relative_url FROM tags WHERE old_relative_url IS NOT NULL ORDER BY id SQL existing_permalinks = Permalink.where("tag_id IS NOT NULL").pluck(:tag_id).to_set create_permalinks(rows) do |row| tag_id = @tag_mapping[row["id"]] next if !tag_id || existing_permalinks.include?(tag_id) { url: row["old_relative_url"], tag_id: tag_id } end rows.close else puts " Skipping import of topic tags because tags have not been imported." end puts "", "Importing permalinks for external/relative URLs..." rows = query(<<~SQL) SELECT url, external_url FROM permalinks WHERE external_url IS NOT NULL ORDER BY url SQL existing_permalinks = Permalink.where("external_url IS NOT NULL").pluck(:external_url).to_set create_permalinks(rows) do |row| next if existing_permalinks.include?(row["external_url"]) { url: row["url"], external_url: row["external_url"] } end rows.close end def create_connection(path) sqlite = SQLite3::Database.new(path, results_as_hash: true) sqlite.busy_timeout = 60_000 # 60 seconds sqlite.journal_mode = "wal" sqlite.synchronous = "normal" sqlite end def query(sql, *bind_vars, db: @source_db) result_set = db.prepare(sql).execute(*bind_vars) if block_given? result = yield result_set result_set.close result else result_set end end def to_date(text) text.present? ? Date.parse(text) : nil end def to_datetime(text) text.present? ? DateTime.parse(text) : nil end def to_boolean(value) value == 1 end end BulkImport::Generic.new(ARGV[0], ARGV[1]).start