Files
ubicloud/prog/test/vm_group.rb
Jeremy Evans 215f09541a Make access_tag only for project <-> accounts join table
For the includers of HyperTagMethods, this changes the authorization
code and object_tag member validation code to look at the project_id
column for the object, instead of looking a row with the project and
object in the access_tag table.

This removes all calls to associate_with_project, other than
those for Account.  It removes the projects association for
the includers of HyperTagMethods, and adds a project association to
the models that didn't already have one, since there is only a single
project for each object now.

Most HyperTagMethods code is inlined into Account, since it is only
user of the code now.  Temporarily, other models will still include
HyperTagMethods for the before_destroy hook, but eventually it will
go away completely.

The associations in Projects that previous used access_tag as a join
table, are changed from many_to_many to one_to_many, except for
Account (which still uses the join table).

Project#has_resources now needs separate queries for all of the
resource classes to see if there any associated objects.

This causes a lot of fallout in the specs, but unfortunately that is
unavoidable due the extensive use of projects.first in the specs to
get the related project for the objects, as well as the extensive
use of associate_with_project.
2025-01-17 08:32:46 -08:00

157 lines
4.0 KiB
Ruby

# frozen_string_literal: true
require "net/ssh"
DEFAULT_BOOT_IMAGE_NAMES = Option::BootImages.map { _1.name }.freeze
class Prog::Test::VmGroup < Prog::Test::Base
def self.assemble(storage_encrypted: true, test_reboot: true, boot_images: DEFAULT_BOOT_IMAGE_NAMES)
Strand.create_with_id(
prog: "Test::VmGroup",
label: "start",
stack: [{
"storage_encrypted" => storage_encrypted,
"test_reboot" => test_reboot,
"vms" => [],
"boot_images" => boot_images
}]
)
end
label def start
hop_setup_vms
end
label def setup_vms
project = Project.create_with_id(name: "project-1")
subnet1_s = Prog::Vnet::SubnetNexus.assemble(
project.id, name: "the-first-subnet", location: "hetzner-fsn1"
)
subnet2_s = Prog::Vnet::SubnetNexus.assemble(
project.id, name: "the-second-subnet", location: "hetzner-fsn1"
)
storage_encrypted = frame.fetch("storage_encrypted", true)
boot_images = frame.fetch("boot_images")
vm1_s = Prog::Vm::Nexus.assemble_with_sshable(
"ubi", project.id,
private_subnet_id: subnet1_s.id,
storage_volumes: [
{encrypted: storage_encrypted, skip_sync: true},
{encrypted: storage_encrypted, size_gib: 5}
],
boot_image: boot_images.sample,
enable_ip4: true
)
vm2_s = Prog::Vm::Nexus.assemble_with_sshable(
"ubi", project.id,
private_subnet_id: subnet1_s.id,
storage_volumes: [{
encrypted: storage_encrypted, skip_sync: false,
max_read_mbytes_per_sec: 200,
max_write_mbytes_per_sec: 150,
max_ios_per_sec: 25600
}],
boot_image: boot_images.sample,
enable_ip4: true
)
vm3_s = Prog::Vm::Nexus.assemble_with_sshable(
"ubi", project.id,
private_subnet_id: subnet2_s.id,
storage_volumes: [{encrypted: storage_encrypted, skip_sync: false}],
boot_image: boot_images.sample,
enable_ip4: true
)
update_stack({
"vms" => [vm1_s.id, vm2_s.id, vm3_s.id],
"subnets" => [subnet1_s.id, subnet2_s.id],
"project_id" => project.id
})
hop_wait_vms
end
label def wait_vms
nap 10 if frame["vms"].any? { Vm[_1].display_state != "running" }
hop_verify_vms
end
label def verify_vms
if retval&.dig("msg") == "Verified VM!"
hop_verify_firewall_rules
end
push Prog::Test::Vm, {subject_id: frame["vms"].first}
end
label def verify_firewall_rules
if retval&.dig("msg") == "Verified Firewall Rules!"
hop_verify_connected_subnets
end
push Prog::Test::FirewallRules, {subject_id: PrivateSubnet[frame["subnets"].first].firewalls.first.id}
end
label def verify_connected_subnets
if retval&.dig("msg") == "Verified Connected Subnets!"
if frame["test_reboot"]
hop_test_reboot
else
hop_destroy_resources
end
end
ps1, ps2 = frame["subnets"].map { PrivateSubnet[_1] }
push Prog::Test::ConnectedSubnets, {subnet_id_multiple: ((ps1.vms.count > 1) ? ps1.id : ps2.id), subnet_id_single: ((ps1.vms.count > 1) ? ps2.id : ps1.id)}
end
label def test_reboot
vm_host.incr_reboot
hop_wait_reboot
end
label def wait_reboot
if vm_host.strand.label == "wait" && vm_host.strand.semaphores.empty?
# Run VM tests again, but avoid rebooting again
update_stack({"test_reboot" => false})
hop_verify_vms
end
nap 20
end
label def destroy_resources
frame["vms"].each { Vm[_1].incr_destroy }
frame["subnets"].each { PrivateSubnet[_1].tap { |ps| ps.firewalls.each { |fw| fw.destroy } }.incr_destroy }
hop_wait_resources_destroyed
end
label def wait_resources_destroyed
unless frame["vms"].all? { Vm[_1].nil? } && frame["subnets"].all? { PrivateSubnet[_1].nil? }
nap 5
end
hop_finish
end
label def finish
Project[frame["project_id"]].destroy
pop "VmGroup tests finished!"
end
label def failed
nap 15
end
def vm_host
@vm_host ||= Vm[frame["vms"].first].vm_host
end
end