scinote-web/app/models/experiment.rb

632 lines
20 KiB
Ruby
Raw Normal View History

2017-06-23 21:19:08 +08:00
class Experiment < ApplicationRecord
include ArchivableModel
include SearchableModel
include SearchableByNameModel
belongs_to :project, inverse_of: :experiments, touch: true, optional: true
2017-06-28 21:21:32 +08:00
belongs_to :created_by,
foreign_key: :created_by_id,
class_name: 'User',
optional: true
belongs_to :last_modified_by,
foreign_key: :last_modified_by_id,
class_name: 'User',
optional: true
belongs_to :archived_by,
foreign_key: :archived_by_id, class_name: 'User', optional: true
belongs_to :restored_by,
foreign_key: :restored_by_id,
class_name: 'User',
optional: true
has_many :my_modules, inverse_of: :experiment, dependent: :destroy
has_many :active_my_modules,
-> { where(archived: false).order(:workflow_order) },
class_name: 'MyModule'
has_many :my_module_groups, inverse_of: :experiment, dependent: :destroy
2016-08-03 18:49:24 +08:00
has_many :report_elements, inverse_of: :experiment, dependent: :destroy
2019-02-25 22:21:13 +08:00
# Associations for old activity type
2017-04-20 16:56:25 +08:00
has_many :activities, inverse_of: :experiment
2016-08-09 22:37:44 +08:00
has_attached_file :workflowimg
validates_attachment :workflowimg,
2016-08-11 15:19:12 +08:00
content_type: { content_type: ['image/png'] },
if: :workflowimg_check
auto_strip_attributes :name, :description, nullify: false
validates :name,
length: { minimum: Constants::NAME_MIN_LENGTH,
maximum: Constants::NAME_MAX_LENGTH },
uniqueness: { scope: :project, case_sensitive: false }
validates :description, length: { maximum: Constants::TEXT_MAX_LENGTH }
validates :project, presence: true
validates :created_by, presence: true
2016-07-25 20:47:13 +08:00
validates :last_modified_by, presence: true
validates :uuid, uniqueness: { scope: :project },
unless: proc { |e| e.uuid.blank? }
with_options if: :archived do |experiment|
experiment.validates :archived_by, presence: true
experiment.validates :archived_on, presence: true
end
scope :is_archived, ->(is_archived) { where("archived = ?", is_archived) }
2017-01-09 18:33:28 +08:00
def self.search(
user,
include_archived,
query = nil,
page = 1,
2017-05-05 22:41:23 +08:00
current_team = nil,
options = {}
2017-01-09 18:33:28 +08:00
)
project_ids =
Project
2016-10-07 00:36:55 +08:00
.search(user, include_archived, nil, Constants::SEARCH_NO_LIMIT)
2017-04-11 20:55:44 +08:00
.pluck(:id)
2017-01-24 23:34:21 +08:00
if current_team
projects_ids =
Project
.search(user,
include_archived,
nil,
1,
2017-01-24 23:34:21 +08:00
current_team)
.select('id')
new_query =
Experiment
.where('experiments.project_id IN (?)', projects_ids)
2017-05-05 22:41:23 +08:00
.where_attributes_like([:name, :description], query, options)
return include_archived ? new_query : new_query.is_archived(false)
2017-01-09 18:33:28 +08:00
elsif include_archived
new_query =
Experiment
.where(project: project_ids)
2017-05-05 22:41:23 +08:00
.where_attributes_like([:name, :description], query, options)
else
new_query =
Experiment
2017-01-09 18:33:28 +08:00
.is_archived(false)
.where(project: project_ids)
2017-05-05 22:41:23 +08:00
.where_attributes_like([:name, :description], query, options)
end
# Show all results if needed
if page == Constants::SEARCH_NO_LIMIT
new_query
else
new_query
2016-10-07 00:36:55 +08:00
.limit(Constants::SEARCH_LIMIT)
.offset((page - 1) * Constants::SEARCH_LIMIT)
end
end
def self.viewable_by_user(user, teams)
where(project: Project.viewable_by_user(user, teams))
end
def navigable?
!project.archived?
end
def active_modules
my_modules.where(archived: false)
end
def archived_modules
my_modules.where(archived: true)
end
def assigned_samples
Sample.joins(:my_modules).where(my_modules: { id: my_modules })
end
def unassigned_samples(assigned_samples)
2017-01-24 23:34:21 +08:00
Sample.where(team_id: team).where.not(id: assigned_samples)
end
def update_canvas(
to_archive,
to_add,
to_rename,
to_move,
to_move_groups,
to_clone,
connections,
positions,
current_user
)
cloned_modules = []
begin
with_lock do
# First, add new modules
new_ids, cloned_pairs, originals = add_modules(
to_add, to_clone, current_user
)
cloned_modules = cloned_pairs.collect { |mn, _| mn }
# Rename modules
2019-03-15 17:30:51 +08:00
rename_modules(to_rename, current_user)
# Add activities that modules were created
originals.each do |my_module|
log_activity(:create_module, current_user, my_module)
end
# Add activities that modules were cloned
cloned_pairs.each do |mn, mo|
Activities::CreateActivityService
.call(activity_type: :clone_module,
owner: current_user,
team: project.team,
project: mn.experiment.project,
subject: mn,
message_items: { my_module_original: mo.id,
my_module_new: mn.id })
end
# Then, archive modules that need to be archived
archive_modules(to_archive, current_user) if to_archive.any?
# Update connections, positions & module group variables
# with actual IDs retrieved from the new modules creation
2016-08-23 16:00:34 +08:00
updated_to_move = {}
to_move.each do |id, value|
updated_to_move[new_ids.fetch(id, id)] = value
end
updated_to_move_groups = {}
to_move_groups.each do |ids, value|
mapped = []
ids.each do |id|
mapped << new_ids.fetch(id, id)
end
updated_to_move_groups[mapped] = value
end
updated_connections = []
connections.each do |a, b|
updated_connections << [new_ids.fetch(a, a), new_ids.fetch(b, b)]
end
updated_positions = {}
positions.each do |id, pos|
updated_positions[new_ids.fetch(id, id)] = pos
end
# Update connections
update_module_connections(updated_connections)
# Update module positions (no validation needed here)
update_module_positions(updated_positions)
# Normalize module positions
normalize_module_positions
# Finally, update module groups
update_module_groups(current_user)
2016-08-19 21:36:12 +08:00
# Finally move any modules to another experiment
2019-03-12 00:17:59 +08:00
move_modules(updated_to_move, current_user)
# Everyhing is set, now we can move any module groups
move_module_groups(updated_to_move_groups, current_user)
end
rescue ActiveRecord::ActiveRecordError,
ArgumentError,
ActiveRecord::RecordNotSaved => ex
logger.error ex.message
return false
end
true
end
def generate_workflow_img
if workflowimg.present?
self.workflowimg = nil
save
end
Experiments::GenerateWorkflowImageService.delay.call(experiment_id: id)
end
2017-01-24 23:34:21 +08:00
# Get projects where user is either owner or user in the same team
2016-08-16 14:47:04 +08:00
# as this experiment
def projects_with_role_above_user(current_user)
2017-01-24 23:34:21 +08:00
team = project.team
projects = team.projects.where(archived: false)
2016-08-16 14:47:04 +08:00
current_user.user_projects
.where(project: projects)
.where('role < 2')
.map(&:project)
end
# Projects to which this experiment can be moved (inside the same
2017-01-24 23:34:21 +08:00
# team and not archived), all users assigned on experiment.project has
2016-08-16 14:47:04 +08:00
# to be assigned on such project
def moveable_projects(current_user)
projects = projects_with_role_above_user(current_user)
projects = projects.each_with_object([]) do |p, arr|
arr << p if (project.users - p.users).empty?
arr
end
projects - [project]
end
private
# Archive all modules. Receives an array of module integer IDs.
def archive_modules(module_ids)
my_modules.where(id: module_ids).each(&:archive!)
my_modules.reload
end
# Archive all modules. Receives an array of module integer IDs
# and current user.
def archive_modules(module_ids, current_user)
my_modules.where(id: module_ids).each do |my_module|
my_module.archive!(current_user)
2019-03-29 22:41:34 +08:00
log_activity(:archive_module, current_user, my_module)
end
my_modules.reload
end
# Add modules, and returns a map of "virtual" IDs with
# actual IDs of saved modules.
# to_add is an array of hashes, each containing 'name',
# 'x', 'y' and 'id'.
# to_clone is a hash, storing new cloned modules as keys,
# and original modules as values.
def add_modules(to_add, to_clone, current_user)
originals = []
cloned_pairs = {}
ids_map = {}
to_add.each do |m|
original = MyModule.find_by_id(to_clone.fetch(m[:id], nil)) if to_clone.present?
if original.present?
my_module = original.deep_clone(current_user)
cloned_pairs[my_module] = original
else
my_module = MyModule.new(experiment: self)
originals << my_module
end
my_module.name = m[:name]
my_module.x = m[:x]
my_module.y = m[:y]
my_module.created_by = current_user
my_module.last_modified_by = current_user
my_module.save!
ids_map[m[:id]] = my_module.id.to_s
end
my_modules.reload
return ids_map, cloned_pairs, originals
end
# Rename modules; this method accepts a map where keys
# represent IDs of modules, and values new names for
# such modules. If a module with given ID doesn't exist,
# it's obviously not updated.
2019-03-12 00:17:59 +08:00
def rename_modules(to_rename, current_user)
to_rename.each do |id, new_name|
my_module = MyModule.find_by_id(id)
if my_module.present?
my_module.name = new_name
my_module.save!
log_activity(:rename_task, current_user, my_module)
end
end
end
# Move modules; this method accepts a map where keys
# represent IDs of modules, and values represent experiment
# IDs of new names to which the given modules should be moved.
# If a module with given ID doesn't exist (or experiment ID)
# it's obviously not updated. Any connection on module is destroyed.
2019-03-12 00:17:59 +08:00
def move_modules(to_move, current_user)
to_move.each do |id, experiment_id|
my_module = my_modules.find_by_id(id)
experiment = project.experiments.find_by_id(experiment_id)
2019-03-12 00:17:59 +08:00
experiment_org = my_module.experiment
2016-08-23 17:08:10 +08:00
next unless my_module.present? && experiment.present?
2016-08-23 17:08:10 +08:00
my_module.experiment = experiment
2016-08-23 17:08:10 +08:00
# Calculate new module position
new_pos = my_module.get_new_position
my_module.x = new_pos[:x]
my_module.y = new_pos[:y]
2016-08-23 17:08:10 +08:00
unless my_module.outputs.destroy_all && my_module.inputs.destroy_all
raise ActiveRecord::ActiveRecordError
end
2016-08-23 17:08:10 +08:00
2019-03-12 00:17:59 +08:00
next unless my_module.save
Activities::CreateActivityService.call(activity_type: :move_task,
owner: current_user,
subject: my_module,
project: project,
team: project.team,
message_items: {
my_module: my_module.id,
experiment_original:
experiment_org.id,
experiment_new: experiment.id
})
2016-09-13 22:32:33 +08:00
end
# Generate workflow image for the experiment in which we moved the task
generate_workflow_img_for_moved_modules(to_move)
end
# Move module groups; this method accepts a map where keys
# represent IDs of modules which are in module group,
# and values represent experiment
# IDs of new names to which the given module group should be moved.
# If a module with given ID doesn't exist (or experiment ID)
# it's obviously not updated. Position for entire module group is updated
# to bottom left corner.
def move_module_groups(to_move, current_user)
to_move.each do |ids, experiment_id|
modules = my_modules.find(ids)
groups = Set.new(modules.map(&:my_module_group))
experiment = project.experiments.find_by_id(experiment_id)
groups.each do |group|
next unless group && experiment.present?
# Find the lowest point for current modules(max_y) and the leftmost
# module(min_x)
if experiment.active_modules.empty?
max_y = 0
min_x = 0
else
max_y = experiment.active_modules.maximum(:y) + MyModule::HEIGHT
min_x = experiment.active_modules.minimum(:x)
end
# Set new positions
curr_min_x = modules.min_by(&:x).x
curr_min_y = modules.min_by(&:y).y
modules.each { |m| m.x += -curr_min_x + min_x }
modules.each { |m| m.y += -curr_min_y + max_y }
modules.each do |m|
experiment_org = m.experiment
m.experiment = experiment
m.save!
# Add activity
Activities::CreateActivityService.call(
activity_type: :move_task,
owner: current_user,
subject: m,
project: experiment.project,
team: experiment.project.team,
message_items: {
my_module: m.id,
experiment_original: experiment_org.id,
experiment_new: experiment.id
}
)
end
group.experiment = experiment
group.save!
end
end
2016-09-13 22:32:33 +08:00
# Generate workflow image for the experiment in which we moved the workflow
generate_workflow_img_for_moved_modules(to_move)
end
# Generates workflow img when the workflow or module is moved
# to other experiment
def generate_workflow_img_for_moved_modules(to_move)
Experiment.where(id: to_move.values.uniq).each do |exp|
exp.generate_workflow_img
2016-09-13 22:32:33 +08:00
end
end
# Update connections for all modules in this project.
# Input is an array of arrays, where first element represents
# source node, and second element represents target node.
# Example input: [ [1, 2], [2, 3], [4, 5], [2, 5] ]
def update_module_connections(connections)
require 'rgl/base'
require 'rgl/adjacency'
require 'rgl/topsort'
dg = RGL::DirectedAdjacencyGraph.new
connections.each do |a, b|
# Check if both vertices exist
if (my_modules.find_all { |m| [a.to_i, b.to_i].include? m.id }).count == 2
dg.add_edge(a, b)
end
end
# Check if cycles exist!
topsort = dg.topsort_iterator.to_a
if topsort.length.zero? && dg.edges.size > 1
raise ArgumentError, 'Cycles exist.'
end
# First, delete existing connections
# but keep a copy of previous state
previous_sources = {}
previous_sources.default = []
my_modules.includes(inputs: { from: [:inputs, outputs: :to] }).each do |m|
previous_sources[m.id] = []
m.inputs.each do |c|
previous_sources[m.id] << c.from
end
end
# There are no callbacks in Connection, so delete_all should be safe
2017-12-20 18:44:11 +08:00
Connection.where(output_id: my_modules).delete_all
# Add new connections
filtered_edges = dg.edges.collect { |e| [e.source, e.target] }
filtered_edges.each do |a, b|
Connection.create!(input_id: b, output_id: a)
end
# Unassign samples from former downstream modules
# for all destroyed connections
unassign_samples_from_old_downstream_modules(previous_sources)
visited = []
# Assign samples to all new downstream modules
filtered_edges.each do |a, b|
source = my_modules.includes({ inputs: :from }, :samples).find(a.to_i)
target = my_modules.find(b.to_i)
# Do this only for new edges
next unless previous_sources[target.id].exclude?(source)
# Go as high upstream as new edges take us
# and then assign samples to all downsteam samples
assign_samples_to_new_downstream_modules(previous_sources,
visited,
source)
end
# Save topological order of modules (for modules without workflow,
# leave them unordered)
my_modules.includes(:my_module_group).each do |m|
m.workflow_order =
if topsort.include? m.id.to_s
topsort.find_index(m.id.to_s)
else
-1
end
m.save!
end
# Make sure to reload my modules, which now have updated connections
# and samples
my_modules.reload
true
end
# When connections are deleted, unassign samples that
# are not inherited anymore
def unassign_samples_from_old_downstream_modules(sources)
my_modules.each do |my_module|
sources[my_module.id].each do |src|
# Only do this for newly deleted connections
next unless src.outputs.map(&:to).exclude? my_module
my_module.downstream_modules.each do |dm|
# Get unique samples for all upstream modules
um = dm.upstream_modules
um.shift # remove current module
ums = um.map(&:samples).flatten.uniq
2017-10-11 23:24:31 +08:00
src.samples.find_each do |sample|
dm.samples.destroy(sample) if ums.exclude? sample
end
end
end
end
end
# Assign samples to new connections recursively
def assign_samples_to_new_downstream_modules(sources, visited, my_module)
# If samples are already assigned for this module, stop going upstream
return if visited.include?(my_module)
visited << my_module
# Edge case, when module is source or it doesn't have any new input
# connections
if my_module.inputs.blank? ||
(my_module.inputs.map(&:from) - sources[my_module.id]).empty?
my_module.downstream_modules.each do |dm|
new_samples = my_module.samples.where.not(id: dm.samples)
dm.samples << new_samples
end
else
my_module.inputs.each do |input|
# Go upstream for new in connections
if sources[my_module.id].exclude?(input.from)
assign_samples_to_new_downstream_modules(sources, visited, input.from)
end
end
end
end
# Updates positions of modules.
# Input is a map where keys are module IDs, and values are
# hashes like { x: <x>, y: <y> }.
def update_module_positions(positions)
modules = my_modules.where(id: positions.keys)
modules.each do |m|
m.update_columns(x: positions[m.id.to_s][:x], y: positions[m.id.to_s][:y])
end
my_modules.reload
end
# Normalize module positions in this project.
def normalize_module_positions
# This method normalizes module positions so x-s and y-s
# are all positive
2017-10-11 23:24:31 +08:00
x_diff = my_modules.pluck(:x).min
y_diff = my_modules.pluck(:y).min
my_modules.each do |m|
m.update_columns(x: m.x - x_diff, y: m.y - y_diff)
end
end
# Recalculate module groups in this project. Input is
# a hash of module ids and their corresponding module names.
def update_module_groups(current_user)
require 'rgl/base'
require 'rgl/adjacency'
require 'rgl/connected_components'
dg = RGL::DirectedAdjacencyGraph[]
group_ids = Set.new
active_modules.includes(:my_module_group, outputs: :to).each do |m|
group_ids << m.my_module_group.id unless m.my_module_group.blank?
dg.add_vertex m.id unless dg.has_vertex? m.id
m.outputs.each do |o|
dg.add_edge m.id, o.to.id
end
end
workflows = []
dg.to_undirected.each_connected_component do |w|
workflows << my_modules.find(w)
end
# Remove any existing module groups from modules
2017-12-20 18:44:11 +08:00
unless MyModuleGroup.where(id: group_ids.to_a).destroy_all
raise ActiveRecord::ActiveRecordError
end
# Second, create new groups
workflows.each do |modules|
# Single modules are not considered part of any workflow
next unless modules.length > 1
MyModuleGroup.create!(experiment: self,
my_modules: modules,
created_by: current_user)
end
my_module_groups.reload
true
end
2016-08-11 15:19:12 +08:00
def workflowimg_check
workflowimg_content_type
rescue
false
end
def log_activity(type_of, current_user, my_module)
Activities::CreateActivityService
.call(activity_type: type_of,
owner: current_user,
team: my_module.experiment.project.team,
project: my_module.experiment.project,
subject: my_module,
message_items: { my_module: my_module.id })
end
end