mirror of
https://github.com/scinote-eln/scinote-web.git
synced 2024-11-10 17:36:33 +08:00
548 lines
16 KiB
Ruby
548 lines
16 KiB
Ruby
|
class Project < ActiveRecord::Base
|
||
|
include ArchivableModel, SearchableModel
|
||
|
|
||
|
enum visibility: { hidden: 0, visible: 1 }
|
||
|
|
||
|
validates :name,
|
||
|
presence: true,
|
||
|
length: { minimum: 4, maximum: 30 },
|
||
|
uniqueness: { scope: :organization, case_sensitive: false }
|
||
|
validates :visibility, presence: true
|
||
|
validates :organization, presence: true
|
||
|
|
||
|
belongs_to :created_by, foreign_key: 'created_by_id', class_name: 'User'
|
||
|
belongs_to :last_modified_by, foreign_key: 'last_modified_by_id', class_name: 'User'
|
||
|
belongs_to :archived_by, foreign_key: 'archived_by_id', class_name: 'User'
|
||
|
belongs_to :restored_by, foreign_key: 'restored_by_id', class_name: 'User'
|
||
|
has_many :user_projects, inverse_of: :project
|
||
|
has_many :users, through: :user_projects
|
||
|
has_many :my_modules, inverse_of: :project
|
||
|
has_many :project_comments, inverse_of: :project
|
||
|
has_many :comments, through: :project_comments
|
||
|
has_many :activities, inverse_of: :project
|
||
|
has_many :my_module_groups, inverse_of: :project
|
||
|
has_many :tags, inverse_of: :project
|
||
|
has_many :reports, inverse_of: :project, dependent: :destroy
|
||
|
has_many :report_elements, inverse_of: :project, dependent: :destroy
|
||
|
belongs_to :organization, inverse_of: :projects
|
||
|
|
||
|
def self.search(user, include_archived, query = nil, page = 1)
|
||
|
org_ids =
|
||
|
Organization
|
||
|
.joins(:user_organizations)
|
||
|
.where("user_organizations.user_id = ?", user.id)
|
||
|
.select("id")
|
||
|
.distinct
|
||
|
|
||
|
if include_archived
|
||
|
new_query = Project
|
||
|
.distinct
|
||
|
.joins(:user_projects)
|
||
|
.where("projects.organization_id IN (?)", org_ids)
|
||
|
.where("projects.visibility = 1 OR user_projects.user_id = ?", user.id)
|
||
|
.where_attributes_like(:name, query)
|
||
|
|
||
|
else
|
||
|
new_query = Project
|
||
|
.distinct
|
||
|
.joins(:user_projects)
|
||
|
.where("projects.organization_id IN (?)", org_ids)
|
||
|
.where("projects.visibility = 1 OR user_projects.user_id = ?", user.id)
|
||
|
.where_attributes_like(:name, query)
|
||
|
.where("projects.archived = ?", false)
|
||
|
end
|
||
|
|
||
|
# Show all results if needed
|
||
|
if page == SHOW_ALL_RESULTS
|
||
|
new_query
|
||
|
else
|
||
|
new_query
|
||
|
.limit(SEARCH_LIMIT)
|
||
|
.offset((page - 1) * SEARCH_LIMIT)
|
||
|
end
|
||
|
end
|
||
|
|
||
|
def last_activities(count = 20)
|
||
|
activities.order(:created_at).last(count)
|
||
|
end
|
||
|
|
||
|
# Get project comments order by created_at time. Results are paginated
|
||
|
# using last comment id and per_page parameters.
|
||
|
def last_comments(last_id = 1, per_page = 20)
|
||
|
last_id = 9999999999999 if last_id <= 1
|
||
|
Comment.joins(:project_comment)
|
||
|
.where(project_comments: {project_id: id})
|
||
|
.where('comments.id < ?', last_id)
|
||
|
.order(created_at: :desc)
|
||
|
.limit(per_page)
|
||
|
end
|
||
|
|
||
|
def active_modules
|
||
|
self.my_modules.where(:archived => false)
|
||
|
end
|
||
|
|
||
|
def archived_modules
|
||
|
self.my_modules.where(:archived => true)
|
||
|
end
|
||
|
|
||
|
def unassigned_users
|
||
|
User.find_by_sql(
|
||
|
"SELECT DISTINCT users.id, users.full_name FROM users " +
|
||
|
"INNER JOIN user_organizations ON users.id = user_organizations.user_id " +
|
||
|
"WHERE user_organizations.organization_id = " + organization_id.to_s +
|
||
|
" AND users.id NOT IN " +
|
||
|
"(SELECT DISTINCT user_id FROM user_projects WHERE user_projects.project_id = " + id.to_s + ")"
|
||
|
)
|
||
|
end
|
||
|
|
||
|
def assigned_modules(user)
|
||
|
role = self.user_role(user)
|
||
|
if role.blank?
|
||
|
return MyModule.none
|
||
|
elsif role == "owner"
|
||
|
return self.my_modules.where(archived: false)
|
||
|
else
|
||
|
return self.my_modules.where(archived: false)
|
||
|
.joins(:user_my_modules)
|
||
|
.where("user_my_modules.user_id IN (?)", user.id)
|
||
|
.distinct
|
||
|
end
|
||
|
end
|
||
|
|
||
|
def user_role(user)
|
||
|
unless self.users.include? user
|
||
|
return nil
|
||
|
end
|
||
|
|
||
|
return (self.user_projects.select { |up| up.user == user }).first.role
|
||
|
end
|
||
|
|
||
|
def modules_without_group
|
||
|
MyModule.where(project_id: id).where(my_module_group: nil)
|
||
|
.where(archived: false)
|
||
|
end
|
||
|
|
||
|
def active_module_groups
|
||
|
self.my_module_groups.joins(:my_modules)
|
||
|
.where('my_modules.archived = ?', false)
|
||
|
.distinct
|
||
|
end
|
||
|
|
||
|
def assigned_samples
|
||
|
Sample.joins(:my_modules).where(my_modules: {id: my_modules} )
|
||
|
end
|
||
|
|
||
|
def unassigned_samples(assigned_samples)
|
||
|
Sample.where(organization_id: organization).where.not(id: assigned_samples)
|
||
|
end
|
||
|
|
||
|
def space_taken
|
||
|
st = 0
|
||
|
my_modules.find_each do |my_module|
|
||
|
st += my_module.space_taken
|
||
|
end
|
||
|
st
|
||
|
end
|
||
|
|
||
|
def update_canvas(
|
||
|
to_archive,
|
||
|
to_add,
|
||
|
to_rename,
|
||
|
to_clone,
|
||
|
connections,
|
||
|
positions,
|
||
|
current_user,
|
||
|
module_groups
|
||
|
)
|
||
|
cloned_modules = []
|
||
|
begin
|
||
|
Project.transaction do
|
||
|
# First, add new modules
|
||
|
new_ids, cloned_pairs, originals = add_modules(
|
||
|
to_add, to_clone, current_user)
|
||
|
cloned_modules = cloned_pairs.collect { |mn, _| mn }
|
||
|
|
||
|
# Rename modules
|
||
|
rename_modules(to_rename)
|
||
|
|
||
|
# Add activities that modules were created
|
||
|
originals.each do |m|
|
||
|
Activity.create(
|
||
|
type_of: :create_module,
|
||
|
user: current_user,
|
||
|
project: self,
|
||
|
my_module: m,
|
||
|
message: I18n.t(
|
||
|
"activities.create_module",
|
||
|
user: current_user.full_name,
|
||
|
module: m.name
|
||
|
)
|
||
|
)
|
||
|
end
|
||
|
|
||
|
# Add activities that modules were cloned
|
||
|
cloned_pairs.each do |mn, mo|
|
||
|
Activity.create(
|
||
|
type_of: :clone_module,
|
||
|
project: mn.project,
|
||
|
my_module: mn,
|
||
|
user: current_user,
|
||
|
message: I18n.t(
|
||
|
"activities.clone_module",
|
||
|
user: current_user.full_name,
|
||
|
module_new: mn.name,
|
||
|
module_original: mo.name
|
||
|
)
|
||
|
)
|
||
|
end
|
||
|
|
||
|
# Then, archive modules that need to be archived
|
||
|
archive_modules(to_archive, current_user)
|
||
|
|
||
|
# Update connections, positions & module group variables
|
||
|
# with actual IDs retrieved from the new modules creation
|
||
|
updated_connections = []
|
||
|
connections.each do |a,b|
|
||
|
updated_connections << [new_ids.fetch(a, a), new_ids.fetch(b, b)]
|
||
|
end
|
||
|
updated_positions = Hash.new
|
||
|
positions.each do |id, pos|
|
||
|
updated_positions[new_ids.fetch(id, id)] = pos
|
||
|
end
|
||
|
updated_module_groups = {}
|
||
|
module_groups.each do |id, name|
|
||
|
updated_module_groups[new_ids.fetch(id, id)] = name
|
||
|
end
|
||
|
|
||
|
# Update connections
|
||
|
update_module_connections(updated_connections)
|
||
|
|
||
|
# Update module positions (no validation needed here)
|
||
|
update_module_positions(updated_positions)
|
||
|
|
||
|
# Normalize module positions
|
||
|
normalize_module_positions
|
||
|
|
||
|
# Finally, update module groups
|
||
|
update_module_groups(updated_module_groups, current_user)
|
||
|
end
|
||
|
rescue ActiveRecord::ActiveRecordError, ArgumentError, ActiveRecord::RecordNotSaved
|
||
|
return false
|
||
|
end
|
||
|
|
||
|
return true
|
||
|
end
|
||
|
|
||
|
# Writes to user log.
|
||
|
def log(message)
|
||
|
final = "[%s] %s" % [name, message]
|
||
|
organization.log(final)
|
||
|
end
|
||
|
|
||
|
private
|
||
|
|
||
|
# Archive all modules. Receives an array of module integer IDs.
|
||
|
def archive_modules(module_ids)
|
||
|
module_ids.each do |m_id|
|
||
|
my_module = my_modules.find_by_id(m_id)
|
||
|
unless my_module.blank?
|
||
|
my_module.archive!
|
||
|
end
|
||
|
end
|
||
|
my_modules.reload
|
||
|
end
|
||
|
|
||
|
# Archive all modules. Receives an array of module integer IDs and current user.
|
||
|
def archive_modules(module_ids, current_user)
|
||
|
module_ids.each do |m_id|
|
||
|
my_module = my_modules.find_by_id(m_id)
|
||
|
unless my_module.blank?
|
||
|
my_module.archive!(current_user)
|
||
|
end
|
||
|
end
|
||
|
my_modules.reload
|
||
|
end
|
||
|
|
||
|
# Add modules, and returns a map of "virtual" IDs with
|
||
|
# actual IDs of saved modules.
|
||
|
# to_add is an array of hashes, each containing 'name',
|
||
|
# 'x', 'y' and 'id'.
|
||
|
# to_clone is a hash, storing new cloned modules as keys,
|
||
|
# and original modules as values.
|
||
|
def add_modules(to_add, to_clone, current_user)
|
||
|
originals = []
|
||
|
cloned_pairs = {}
|
||
|
ids_map = Hash.new
|
||
|
to_add.each do |m|
|
||
|
original = MyModule.find_by_id(to_clone.fetch(m[:id], nil))
|
||
|
if original.present? then
|
||
|
my_module = original.deep_clone(current_user)
|
||
|
cloned_pairs[my_module] = original
|
||
|
else
|
||
|
my_module = MyModule.new(
|
||
|
project: self)
|
||
|
originals << my_module
|
||
|
end
|
||
|
|
||
|
my_module.name = m[:name]
|
||
|
my_module.x = m[:x]
|
||
|
my_module.y = m[:y]
|
||
|
my_module.created_by = current_user
|
||
|
my_module.last_modified_by = current_user
|
||
|
my_module.save!
|
||
|
|
||
|
ids_map[m[:id]] = my_module.id.to_s
|
||
|
end
|
||
|
my_modules.reload
|
||
|
return ids_map, cloned_pairs, originals
|
||
|
end
|
||
|
|
||
|
# Rename modules; this method accepts a map where keys
|
||
|
# represent IDs of modules, and values new names for
|
||
|
# such modules. If a module with given ID doesn't exist,
|
||
|
# it's obviously not updated.
|
||
|
def rename_modules(to_rename)
|
||
|
to_rename.each do |id, new_name|
|
||
|
my_module = MyModule.find_by_id(id)
|
||
|
if my_module.present?
|
||
|
my_module.name = new_name
|
||
|
my_module.save!
|
||
|
end
|
||
|
end
|
||
|
end
|
||
|
|
||
|
# Update connections for all modules in this project.
|
||
|
# Input is an array of arrays, where first element represents
|
||
|
# source node, and second element represents target node.
|
||
|
# Example input: [ [1, 2], [2, 3], [4, 5], [2, 5] ]
|
||
|
def update_module_connections(connections)
|
||
|
require 'rgl/base'
|
||
|
require 'rgl/adjacency'
|
||
|
require 'rgl/topsort'
|
||
|
|
||
|
dg = RGL::DirectedAdjacencyGraph.new
|
||
|
connections.each do |a,b|
|
||
|
# Check if both vertices exist
|
||
|
if (my_modules.find_all {|m| [a.to_i, b.to_i].include? m.id }).count == 2
|
||
|
dg.add_edge(a, b)
|
||
|
end
|
||
|
end
|
||
|
|
||
|
# Check if cycles exist!
|
||
|
topsort = dg.topsort_iterator.to_a
|
||
|
if topsort.length == 0 and dg.edges.size > 1
|
||
|
raise ArgumentError, "Cycles exist."
|
||
|
end
|
||
|
|
||
|
# First, delete existing connections
|
||
|
# but keep a copy of previous state
|
||
|
previous_sources = {}
|
||
|
previous_sources.default = []
|
||
|
my_modules.each do |m|
|
||
|
previous_sources[m.id] = []
|
||
|
m.inputs.each do |c|
|
||
|
previous_sources[m.id] << c.from
|
||
|
end
|
||
|
end
|
||
|
my_modules.each do |m|
|
||
|
unless m.outputs.destroy_all
|
||
|
raise ActiveRecord::ActiveRecordError
|
||
|
end
|
||
|
end
|
||
|
|
||
|
|
||
|
# Add new connections
|
||
|
filtered_edges = dg.edges.collect { |e| [e.source, e.target] }
|
||
|
filtered_edges.each do |a, b|
|
||
|
Connection.create!(:input_id => b, :output_id => a)
|
||
|
end
|
||
|
|
||
|
# Unassign samples from former downstream modules
|
||
|
# for all destroyed connections
|
||
|
unassign_samples_from_old_downstream_modules(previous_sources)
|
||
|
|
||
|
visited = []
|
||
|
# Assign samples to all new downstream modules
|
||
|
filtered_edges.each do |a, b|
|
||
|
source = my_modules.find(a.to_i)
|
||
|
target = my_modules.find(b.to_i)
|
||
|
# Do this only for new edges
|
||
|
if previous_sources[target.id].exclude?(source)
|
||
|
# Go as high upstream as new edges take us
|
||
|
# and then assign samples to all downsteam samples
|
||
|
assign_samples_to_new_downstream_modules(previous_sources, visited, source)
|
||
|
end
|
||
|
end
|
||
|
|
||
|
# Save topological order of modules (for modules without workflow,
|
||
|
# leave them unordered)
|
||
|
my_modules.each do |m|
|
||
|
if topsort.include? m.id.to_s
|
||
|
m.workflow_order = topsort.find_index(m.id.to_s)
|
||
|
else
|
||
|
m.workflow_order = -1
|
||
|
end
|
||
|
m.save!
|
||
|
end
|
||
|
|
||
|
# Make sure to reload my modules, which now have updated connections and samples
|
||
|
my_modules.reload
|
||
|
true
|
||
|
end
|
||
|
|
||
|
# When connections are deleted, unassign samples that
|
||
|
# are not inherited anymore
|
||
|
def unassign_samples_from_old_downstream_modules(sources)
|
||
|
my_modules.each do |my_module|
|
||
|
sources[my_module.id].each do |s|
|
||
|
# Only do this for newly deleted connections
|
||
|
if s.outputs.map{|i| i.to}.exclude? my_module
|
||
|
my_module.get_downstream_modules.each do |dm|
|
||
|
# Get unique samples for all upstream modules
|
||
|
um = dm.get_upstream_modules
|
||
|
um.shift # remove current module
|
||
|
ums = um.map{|m| m.samples}.flatten.uniq
|
||
|
s.samples.each do |sample|
|
||
|
dm.samples.delete(sample) if ums.exclude? sample
|
||
|
end
|
||
|
end
|
||
|
end
|
||
|
end
|
||
|
end
|
||
|
end
|
||
|
|
||
|
# Assign samples to new connections recursively
|
||
|
def assign_samples_to_new_downstream_modules(sources, visited, my_module)
|
||
|
# If samples are already assigned for this module, stop going upstream
|
||
|
if visited.include? (my_module)
|
||
|
return
|
||
|
end
|
||
|
visited << my_module
|
||
|
# Edge case, when module is source or it doesn't have any new input connections
|
||
|
if my_module.inputs.blank? or (
|
||
|
my_module.inputs.map{|c| c.from} -
|
||
|
sources[my_module.id]
|
||
|
).empty?
|
||
|
my_module.get_downstream_modules.each do |dm|
|
||
|
new_samples = my_module.samples.select { |el| dm.samples.exclude?(el) }
|
||
|
dm.samples.push(*new_samples)
|
||
|
end
|
||
|
else
|
||
|
my_module.inputs.each do |input|
|
||
|
# Go upstream for new in connections
|
||
|
if sources[my_module.id].exclude?(input.from)
|
||
|
assign_samples_to_new_downstream_modules(input.from)
|
||
|
end
|
||
|
end
|
||
|
end
|
||
|
end
|
||
|
|
||
|
# Updates positions of modules.
|
||
|
# Input is a map where keys are module IDs, and values are
|
||
|
# hashes like { x: <x>, y: <y> }.
|
||
|
def update_module_positions(positions)
|
||
|
positions.each do |id, pos|
|
||
|
unless MyModule.update(id, x: pos[:x], y: pos[:y])
|
||
|
raise ActiveRecord::ActiveRecordError
|
||
|
end
|
||
|
end
|
||
|
my_modules.reload
|
||
|
end
|
||
|
|
||
|
# Normalize module positions in this project.
|
||
|
def normalize_module_positions
|
||
|
# This method normalizes module positions so x-s and y-s
|
||
|
# are all positive
|
||
|
x_diff = (my_modules.collect { |m| m.x }).min
|
||
|
y_diff = (my_modules.collect { |m| m.y }).min
|
||
|
|
||
|
my_modules.each do |m|
|
||
|
unless
|
||
|
m.update_attribute(:x, m.x - x_diff) and
|
||
|
m.update_attribute(:y, m.y - y_diff)
|
||
|
raise ActiveRecord::ActiveRecordError
|
||
|
end
|
||
|
end
|
||
|
end
|
||
|
|
||
|
# Recalculate module groups in this project. Input is
|
||
|
# a hash of module ids and their corresponding module names.
|
||
|
def update_module_groups(module_groups, current_user)
|
||
|
require 'rgl/base'
|
||
|
require 'rgl/adjacency'
|
||
|
require 'rgl/connected_components'
|
||
|
|
||
|
dg = RGL::DirectedAdjacencyGraph[]
|
||
|
group_ids = Set.new
|
||
|
my_modules.where(archived: :false).each do |m|
|
||
|
unless m.my_module_group.blank?
|
||
|
group_ids << m.my_module_group.id
|
||
|
end
|
||
|
unless dg.has_vertex? m.id
|
||
|
dg.add_vertex m.id
|
||
|
end
|
||
|
m.outputs.each do |o|
|
||
|
dg.add_edge m.id, o.to.id
|
||
|
end
|
||
|
end
|
||
|
workflows = []
|
||
|
dg.to_undirected.each_connected_component { |w| workflows << w }
|
||
|
|
||
|
# Retrieve maximum allowed module group name
|
||
|
max_length = (MyModuleGroup.validators_on(:name).select { |v| v.class == ActiveModel::Validations::LengthValidator }).first.options[:maximum]
|
||
|
# For each workflow, generate new names
|
||
|
new_index = 1
|
||
|
wf_names = []
|
||
|
suffix = I18n.t("my_module_groups.new.suffix")
|
||
|
cut_index = -(suffix.length + 1)
|
||
|
workflows.each do |w|
|
||
|
modules = MyModule.find(w)
|
||
|
|
||
|
# Get an array of module names
|
||
|
names = []
|
||
|
modules.each do |m|
|
||
|
names << module_groups.fetch(m.id.to_s, "")
|
||
|
end
|
||
|
names = names.uniq
|
||
|
name = (names.select { |v| v != "" }).join(", ")
|
||
|
|
||
|
if w.length <= 1
|
||
|
name = nil
|
||
|
elsif name.blank?
|
||
|
name = I18n.t("my_module_groups.new.name", index: new_index)
|
||
|
new_index += 1
|
||
|
while MyModuleGroup.find_by(name: name).present?
|
||
|
name = I18n.t("my_module_groups.new.name", index: new_index)
|
||
|
new_index += 1
|
||
|
end
|
||
|
elsif name.length > max_length
|
||
|
# If length is too long, shorten it
|
||
|
name = name[0..(max_length + cut_index)] + suffix
|
||
|
end
|
||
|
|
||
|
wf_names << name
|
||
|
end
|
||
|
|
||
|
# Remove any existing module groups from modules
|
||
|
unless MyModuleGroup.destroy_all(:id => group_ids.to_a)
|
||
|
raise ActiveRecord::ActiveRecordError
|
||
|
end
|
||
|
|
||
|
# Second, create new groups
|
||
|
workflows.each_with_index do |w, i|
|
||
|
# Single modules are not considered part of any workflow
|
||
|
if w.length > 1
|
||
|
group = MyModuleGroup.new(
|
||
|
name: wf_names[i],
|
||
|
project: self,
|
||
|
my_modules: MyModule.find(w))
|
||
|
group.created_by = current_user
|
||
|
group.save!
|
||
|
end
|
||
|
end
|
||
|
|
||
|
my_module_groups.reload
|
||
|
true
|
||
|
end
|
||
|
end
|