# frozen_string_literal: true require_dependency "s3_helper" def brotli_s3_path(path) ext = File.extname(path) "#{path[0..-ext.length]}br#{ext}" end def gzip_s3_path(path) ext = File.extname(path) "#{path[0..-ext.length]}gz#{ext}" end def should_skip?(path) return false if ENV['FORCE_S3_UPLOADS'] @existing_assets ||= Set.new(helper.list("assets/").map(&:key)) @existing_assets.include?(path) end def upload(path, remote_path, content_type, content_encoding = nil) options = { cache_control: 'max-age=31556952, public, immutable', content_type: content_type, acl: 'public-read', tagging: '' } if content_encoding options[:content_encoding] = content_encoding end if should_skip?(remote_path) puts "Skipping: #{remote_path}" else puts "Uploading: #{remote_path}" File.open(path) do |file| helper.upload(file, remote_path, options) end end end def use_db_s3_config ENV["USE_DB_S3_CONFIG"] end def helper @helper ||= begin bucket, options = if use_db_s3_config [ SiteSetting.s3_upload_bucket.downcase, S3Helper.s3_options(SiteSetting) ] else [ GlobalSetting.s3_bucket.downcase, S3Helper.s3_options(GlobalSetting) ] end S3Helper.new(bucket, '', options) end end def assets cached = Rails.application.assets&.cached manifest = Sprockets::Manifest.new(cached, Rails.root + 'public/assets', Rails.application.config.assets.manifest) results = [] manifest.assets.each do |_, path| fullpath = (Rails.root + "public/assets/#{path}").to_s content_type = MiniMime.lookup_by_filename(fullpath).content_type asset_path = "assets/#{path}" results << [fullpath, asset_path, content_type] if File.exist?(fullpath + '.br') results << [fullpath + '.br', brotli_s3_path(asset_path), content_type, 'br'] end if File.exist?(fullpath + '.gz') results << [fullpath + '.gz', gzip_s3_path(asset_path), content_type, 'gzip'] end if File.exist?(fullpath + '.map') results << [fullpath + '.map', asset_path + '.map', 'application/json'] end end results end def asset_paths Set.new(assets.map { |_, asset_path| asset_path }) end def ensure_s3_configured! unless GlobalSetting.use_s3? || use_db_s3_config STDERR.puts "ERROR: Ensure S3 is configured in config/discourse.conf or environment vars" exit 1 end end task 's3:correct_acl' => :environment do ensure_s3_configured! puts "ensuring public-read is set on every upload and optimized image" i = 0 base_url = Discourse.store.absolute_base_url objects = Upload.pluck(:id, :url).map { |array| array << :upload } objects.concat(OptimizedImage.pluck(:id, :url).map { |array| array << :optimized_image }) puts "#{objects.length} objects found" objects.each do |id, url, type| i += 1 if !url.start_with?(base_url) puts "Skipping #{type} #{id} since it is not stored on s3, url is #{url}" else begin key = url[(base_url.length + 1)..-1] object = Discourse.store.s3_helper.object(key) object.acl.put(acl: "public-read") rescue => e puts "Skipping #{type} #{id} url is #{url} #{e}" end end if i % 100 == 0 puts "#{i} done" end end end task 's3:upload_assets' => :environment do ensure_s3_configured! puts "installing CORS rule" helper.ensure_cors! assets.each do |asset| upload(*asset) end end task 's3:expire_missing_assets' => :environment do ensure_s3_configured! count = 0 keep = 0 in_manifest = asset_paths puts "Ensuring AWS assets are tagged correctly for removal" helper.list('assets/').each do |f| if !in_manifest.include?(f.key) helper.tag_file(f.key, old: true) count += 1 else # ensure we do not delete this by mistake helper.tag_file(f.key, {}) keep += 1 end end puts "#{count} assets were flagged for removal in 10 days (#{keep} assets will be retained)" puts "Ensuring AWS rule exists for purging old assets" helper.update_lifecycle("delete_old_assets", 10, tag: { key: 'old', value: 'true' }) end