mirror of
https://github.com/scinote-eln/scinote-web.git
synced 2024-11-14 05:05:55 +08:00
168 lines
5.3 KiB
Ruby
168 lines
5.3 KiB
Ruby
# frozen_string_literal: true
|
|
|
|
# handles the import of repository records
|
|
# requires 4 parameters:
|
|
# @sheet: the csv file with imported rows
|
|
# @mappings: mappings for columns
|
|
# @user: current_user
|
|
# @repository: the repository in which we import the items
|
|
module RepositoryImportParser
|
|
class Importer
|
|
IMPORT_BATCH_SIZE = 500
|
|
|
|
def initialize(sheet, mappings, user, repository)
|
|
@columns = []
|
|
@name_index = -1
|
|
@total_new_rows = 0
|
|
@new_rows_added = 0
|
|
@header_skipped = false
|
|
@repository = repository
|
|
@sheet = sheet
|
|
@rows = SpreadsheetParser.spreadsheet_enumerator(@sheet)
|
|
@mappings = mappings
|
|
@user = user
|
|
@repository_columns = @repository.repository_columns
|
|
end
|
|
|
|
def run
|
|
fetch_columns
|
|
return check_for_duplicate_columns if check_for_duplicate_columns
|
|
|
|
import_rows!
|
|
end
|
|
|
|
private
|
|
|
|
def fetch_columns
|
|
@mappings.each_with_index do |(_, value), index|
|
|
if value == '-1'
|
|
# Fill blank space, so our indices stay the same
|
|
@columns << nil
|
|
@name_index = index
|
|
else
|
|
@columns << @repository_columns.where(data_type: Extends::REPOSITORY_IMPORTABLE_TYPES)
|
|
.preload(Extends::REPOSITORY_IMPORT_COLUMN_PRELOADS)
|
|
.find_by(id: value)
|
|
end
|
|
end
|
|
end
|
|
|
|
def check_for_duplicate_columns
|
|
col_compact = @columns.compact
|
|
if col_compact.map(&:id).uniq.length != col_compact.length
|
|
{ status: :error, nr_of_added: @new_rows_added, total_nr: @total_new_rows }
|
|
end
|
|
end
|
|
|
|
def import_rows!
|
|
errors = false
|
|
|
|
@repository.transaction do
|
|
batch_counter = 0
|
|
full_row_import_batch = []
|
|
|
|
@rows.each do |row|
|
|
# Skip empty rows
|
|
next if row.blank?
|
|
|
|
unless @header_skipped
|
|
@header_skipped = true
|
|
next
|
|
end
|
|
@total_new_rows += 1
|
|
|
|
new_full_row = {}
|
|
SpreadsheetParser.parse_row(row, @sheet).each_with_index do |value, index|
|
|
if index == @name_index
|
|
new_row =
|
|
RepositoryRow.new(name: try_decimal_to_string(value),
|
|
repository: @repository,
|
|
created_by: @user,
|
|
last_modified_by: @user)
|
|
unless new_row.valid?
|
|
errors = true
|
|
break
|
|
end
|
|
|
|
new_full_row[:repository_row] = new_row
|
|
next
|
|
end
|
|
next unless @columns[index]
|
|
|
|
new_full_row[index] = value
|
|
end
|
|
|
|
if new_full_row[:repository_row].present?
|
|
full_row_import_batch << new_full_row
|
|
batch_counter += 1
|
|
end
|
|
|
|
next if batch_counter < IMPORT_BATCH_SIZE
|
|
|
|
import_batch_to_database(full_row_import_batch)
|
|
full_row_import_batch = []
|
|
batch_counter = 0
|
|
end
|
|
|
|
# Import of the remaining rows
|
|
import_batch_to_database(full_row_import_batch) if full_row_import_batch.any?
|
|
end
|
|
|
|
if errors
|
|
return { status: :error,
|
|
nr_of_added: @new_rows_added,
|
|
total_nr: @total_new_rows }
|
|
end
|
|
{ status: :ok, nr_of_added: @new_rows_added, total_nr: @total_new_rows }
|
|
end
|
|
|
|
def import_batch_to_database(full_row_import_batch)
|
|
repository_rows = full_row_import_batch.collect { |row| row[:repository_row] }
|
|
@new_rows_added += RepositoryRow.import(repository_rows, recursive: false, validate: false).ids.length
|
|
repository_rows.each { |row| row.run_callbacks(:create) }
|
|
|
|
import_mappings = Hash[@columns.map { |column| column&.data_type&.to_sym }
|
|
.compact
|
|
.uniq
|
|
.map { |data_type| [data_type, []] }]
|
|
|
|
full_row_import_batch.each do |row|
|
|
next unless row[:repository_row].id
|
|
|
|
row.reject { |k| k == :repository_row }.each do |index, value|
|
|
column = @columns[index]
|
|
value = try_decimal_to_string(value) unless column.repository_number_value?
|
|
next if value.nil?
|
|
|
|
cell_value_attributes = { created_by: @user,
|
|
last_modified_by: @user,
|
|
repository_cell_attributes: { repository_row: row[:repository_row],
|
|
repository_column: column,
|
|
importing: true } }
|
|
cell_value = column.data_type.constantize.import_from_text(
|
|
value,
|
|
cell_value_attributes,
|
|
@user.as_json(root: true, only: :settings).deep_symbolize_keys
|
|
)
|
|
next if cell_value.nil?
|
|
|
|
cell_value.repository_cell.value = cell_value
|
|
|
|
import_mappings[column.data_type.to_sym] << cell_value
|
|
end
|
|
end
|
|
|
|
import_mappings.each do |data_type, cell_values|
|
|
data_type.to_s.constantize.import(cell_values, recursive: true, validate: false)
|
|
end
|
|
end
|
|
|
|
def try_decimal_to_string(value)
|
|
if value.is_a?(BigDecimal)
|
|
value.frac.zero? ? value.to_i.to_s : value.to_s
|
|
else
|
|
value
|
|
end
|
|
end
|
|
end
|
|
end
|