For the includers of HyperTagMethods, this changes the authorization code and object_tag member validation code to look at the project_id column for the object, instead of looking a row with the project and object in the access_tag table. This removes all calls to associate_with_project, other than those for Account. It removes the projects association for the includers of HyperTagMethods, and adds a project association to the models that didn't already have one, since there is only a single project for each object now. Most HyperTagMethods code is inlined into Account, since it is only user of the code now. Temporarily, other models will still include HyperTagMethods for the before_destroy hook, but eventually it will go away completely. The associations in Projects that previous used access_tag as a join table, are changed from many_to_many to one_to_many, except for Account (which still uses the join table). Project#has_resources now needs separate queries for all of the resource classes to see if there any associated objects. This causes a lot of fallout in the specs, but unfortunately that is unavoidable due the extensive use of projects.first in the specs to get the related project for the objects, as well as the extensive use of associate_with_project.
191 lines
7.7 KiB
Ruby
191 lines
7.7 KiB
Ruby
# frozen_string_literal: true
|
|
|
|
require_relative "../../model/spec_helper"
|
|
|
|
RSpec.describe Prog::Minio::MinioClusterNexus do
|
|
subject(:nx) {
|
|
expect(Config).to receive(:minio_service_project_id).and_return(minio_project.id).at_least(:once)
|
|
described_class.new(
|
|
described_class.assemble(
|
|
minio_project.id, "minio", "hetzner-fsn1", "minio-admin", 100, 1, 1, 1, "standard-2"
|
|
)
|
|
)
|
|
}
|
|
|
|
let(:minio_project) { Project.create_with_id(name: "default") }
|
|
|
|
describe ".assemble" do
|
|
before do
|
|
allow(Config).to receive(:minio_service_project_id).and_return(minio_project.id)
|
|
end
|
|
|
|
it "validates input" do
|
|
expect {
|
|
described_class.assemble(SecureRandom.uuid, "minio", "hetzner-fsn1", "minio-admin", 100, 1, 1, 1, "standard-2")
|
|
}.to raise_error RuntimeError, "No existing project"
|
|
|
|
expect {
|
|
described_class.assemble(minio_project.id, "minio", "hetzner-xxx", "minio-admin", 100, 1, 1, 1, "standard-2")
|
|
}.to raise_error Validation::ValidationFailed, "Validation failed for following fields: provider"
|
|
|
|
expect {
|
|
described_class.assemble(minio_project.id, "minio/name", "hetzner-fsn1", "minio-admin", 100, 1, 1, 1, "standard-2")
|
|
}.to raise_error Validation::ValidationFailed, "Validation failed for following fields: name"
|
|
|
|
expect {
|
|
described_class.assemble(minio_project.id, "minio", "hetzner-fsn1", "mu", 100, 1, 1, 1, "standard-2")
|
|
}.to raise_error Validation::ValidationFailed, "Validation failed for following fields: username"
|
|
end
|
|
|
|
it "creates a minio cluster" do
|
|
described_class.assemble(minio_project.id, "minio2", "hetzner-fsn1", "minio-admin", 100, 1, 1, 1, "standard-2")
|
|
|
|
expect(MinioCluster.count).to eq 1
|
|
expect(MinioCluster.first.name).to eq "minio2"
|
|
expect(MinioCluster.first.location).to eq "hetzner-fsn1"
|
|
expect(MinioCluster.first.admin_user).to eq "minio-admin"
|
|
expect(MinioCluster.first.admin_password).to match(/^[A-Za-z0-9_-]{20}$/)
|
|
expect(MinioCluster.first.storage_size_gib).to eq 100
|
|
expect(MinioCluster.first.pools.count).to eq 1
|
|
expect(MinioCluster.first.server_count).to eq 1
|
|
expect(MinioCluster.first.drive_count).to eq 1
|
|
expect(MinioCluster.first.pools.first.vm_size).to eq "standard-2"
|
|
expect(MinioCluster.first.project).to eq minio_project
|
|
expect(MinioCluster.first.strand.label).to eq "wait_pools"
|
|
end
|
|
end
|
|
|
|
describe "#wait_pools" do
|
|
before do
|
|
st = instance_double(Strand, label: "wait")
|
|
instance_double(MinioPool, strand: st).tap { |mp| allow(nx.minio_cluster).to receive(:pools).and_return([mp]) }
|
|
end
|
|
|
|
it "hops to wait if all pools are waiting" do
|
|
expect { nx.wait_pools }.to hop("wait")
|
|
end
|
|
|
|
it "naps if not all pools are waiting" do
|
|
allow(nx.minio_cluster.pools.first.strand).to receive(:label).and_return("start")
|
|
expect { nx.wait_pools }.to nap(5)
|
|
end
|
|
end
|
|
|
|
describe "#wait" do
|
|
it "naps" do
|
|
expect { nx.wait }.to nap(30)
|
|
end
|
|
|
|
it "hops to reconfigure if reconfigure is set" do
|
|
expect(nx).to receive(:when_reconfigure_set?).and_yield
|
|
expect { nx.wait }.to hop("reconfigure")
|
|
end
|
|
|
|
it "hops to refresh_certificates if certificate_last_checked_at is before 1 month" do
|
|
expect(nx.minio_cluster).to receive(:certificate_last_checked_at).and_return(Time.now - 60 * 60 * 24 * 30 - 1)
|
|
expect { nx.wait }.to hop("refresh_certificates")
|
|
end
|
|
end
|
|
|
|
describe "#refresh_certificates" do
|
|
let(:ms) do
|
|
instance_double(MinioServer, cert: "server_cert")
|
|
end
|
|
|
|
before do
|
|
allow(nx.minio_cluster).to receive(:servers).and_return([ms])
|
|
end
|
|
|
|
it "moves root_cert_2 to root_cert_1 and creates new root_cert_2 if root_cert_1 is about to expire, also updates server_cert" do
|
|
rc2 = nx.minio_cluster.root_cert_2
|
|
rck2 = nx.minio_cluster.root_cert_key_2
|
|
certificate_last_checked_at = nx.minio_cluster.certificate_last_checked_at
|
|
expect(OpenSSL::X509::Certificate).to receive(:new).with(nx.minio_cluster.root_cert_1).and_call_original
|
|
expect(Time).to receive(:now).and_return(Time.now + 60 * 60 * 24 * 335 * 5 + 1).at_least(:once)
|
|
expect(Util).to receive(:create_root_certificate).with(common_name: "#{nx.minio_cluster.ubid} Root Certificate Authority", duration: 60 * 60 * 24 * 365 * 10).and_return(["cert", "key"])
|
|
expect(ms).to receive(:incr_reconfigure).once
|
|
|
|
expect { nx.refresh_certificates }.to hop("wait")
|
|
expect(nx.minio_cluster.root_cert_1).to eq rc2
|
|
expect(nx.minio_cluster.root_cert_key_1).to eq rck2
|
|
expect(nx.minio_cluster.root_cert_2).to eq "cert"
|
|
expect(nx.minio_cluster.root_cert_key_2).to eq "key"
|
|
expect(nx.minio_cluster.certificate_last_checked_at).to be > certificate_last_checked_at
|
|
end
|
|
|
|
it "doesn't update root_certs if they are not close to expire" do
|
|
rc1 = nx.minio_cluster.root_cert_1
|
|
rck1 = nx.minio_cluster.root_cert_key_1
|
|
rc2 = nx.minio_cluster.root_cert_2
|
|
rck2 = nx.minio_cluster.root_cert_key_2
|
|
certificate_last_checked_at = nx.minio_cluster.certificate_last_checked_at
|
|
|
|
expect(OpenSSL::X509::Certificate).to receive(:new).with(nx.minio_cluster.root_cert_1).and_call_original
|
|
|
|
expect { nx.refresh_certificates }.to hop("wait")
|
|
expect(nx.minio_cluster.root_cert_1).to eq rc1
|
|
expect(nx.minio_cluster.root_cert_key_1).to eq rck1
|
|
expect(nx.minio_cluster.root_cert_2).to eq rc2
|
|
expect(nx.minio_cluster.root_cert_key_2).to eq rck2
|
|
expect(nx.minio_cluster.certificate_last_checked_at).to be > certificate_last_checked_at
|
|
end
|
|
end
|
|
|
|
describe "#reconfigure" do
|
|
it "increments reconfigure semaphore of all minio servers and hops to wait" do
|
|
expect(nx).to receive(:decr_reconfigure)
|
|
ms = instance_double(MinioServer)
|
|
expect(ms).to receive(:incr_reconfigure)
|
|
expect(nx.minio_cluster).to receive(:servers).and_return([ms]).at_least(:once)
|
|
expect(ms).to receive(:incr_restart)
|
|
expect { nx.reconfigure }.to hop("wait")
|
|
end
|
|
end
|
|
|
|
describe "#destroy" do
|
|
it "increments destroy semaphore of minio pools and hops to wait_pools_destroy" do
|
|
expect(nx).to receive(:decr_destroy)
|
|
mp = instance_double(MinioPool, incr_destroy: nil)
|
|
expect(mp).to receive(:incr_destroy)
|
|
expect(nx.minio_cluster).to receive(:pools).and_return([mp])
|
|
expect { nx.destroy }.to hop("wait_pools_destroyed")
|
|
end
|
|
end
|
|
|
|
describe "#wait_pools_destroyed" do
|
|
it "naps if there are still minio pools" do
|
|
expect(nx.minio_cluster).to receive(:pools).and_return([true])
|
|
expect { nx.wait_pools_destroyed }.to nap(10)
|
|
end
|
|
|
|
it "increments private subnet destroy and destroys minio cluster" do
|
|
expect(nx.minio_cluster).to receive(:pools).and_return([])
|
|
fw = instance_double(Firewall)
|
|
ps = instance_double(PrivateSubnet, firewalls: [fw])
|
|
expect(ps).to receive(:incr_destroy)
|
|
expect(fw).to receive(:destroy)
|
|
expect(nx.minio_cluster).to receive(:private_subnet).and_return(ps).at_least(:once)
|
|
expect(nx.minio_cluster).to receive(:destroy)
|
|
expect { nx.wait_pools_destroyed }.to exit({"msg" => "destroyed"})
|
|
end
|
|
end
|
|
|
|
describe "#before_run" do
|
|
it "hops to destroy if destroy is set" do
|
|
expect(nx).to receive(:when_destroy_set?).and_yield
|
|
expect { nx.before_run }.to hop("destroy")
|
|
end
|
|
|
|
it "does not hop to destroy if destroy is not set" do
|
|
expect(nx).to receive(:when_destroy_set?).and_return(false)
|
|
expect { nx.before_run }.not_to hop("destroy")
|
|
end
|
|
|
|
it "does not hop to destroy if strand label is destroy" do
|
|
expect(nx).to receive(:when_destroy_set?).and_yield
|
|
expect(nx.strand).to receive(:label).and_return("destroy")
|
|
expect { nx.before_run }.not_to hop("destroy")
|
|
end
|
|
end
|
|
end
|