diff --git a/app/models/client.rb b/app/models/client.rb index 8271b76..788026e 100644 --- a/app/models/client.rb +++ b/app/models/client.rb @@ -1,63 +1,64 @@ class Client < ActiveRecord::Base self.table_name = :Client self.primary_key = :ClientId alias_attribute :name, :Name alias_attribute :uname, :Uname alias_attribute :auto_prune, :AutoPrune alias_attribute :file_retention, :FileRetention alias_attribute :job_retention, :JobRetention has_many :jobs, foreign_key: :ClientId + has_one :host, foreign_key: :name, primary_key: :Name DAY_SECS = 60 * 60 * 24 # Helper method. It shows the client's job retention, # (which is expressed in seconds) in days. # # @return [Integer] def job_retention_days job_retention / DAY_SECS end # Helper method. It shows the client's file retention, # (which is expressed in seconds) in days. # # @return [Integer] def file_retention_days file_retention / DAY_SECS end # Helper method for auto_prune # # @return [String] 'yes' or 'no' def auto_prune_human auto_prune == 1 ? 'yes' : 'no' end def last_job_date jobs.maximum(:EndTime) end # Shows the total file size of the jobs that run for a specific client # # @return [Integer] Size in Bytes def backup_jobs_size jobs.backup_type.map(&:job_bytes).sum end # Shows the total files' count for the jobs that run for a specific client # # @return [Integer] File count def files_count jobs.map(&:job_files).sum end # Fetches the client's jobs that are running at the moment # # @return [Integer] def running_jobs jobs.running.count end end diff --git a/app/models/host.rb b/app/models/host.rb index 641c1ae..8deec6e 100644 --- a/app/models/host.rb +++ b/app/models/host.rb @@ -1,51 +1,62 @@ class Host < ActiveRecord::Base FILE_RETENTION_DAYS = 60 JOB_RETENTION_DAYS = 180 CATALOG = 'MyCatalog' + AUTOPRUNE = 1 establish_connection :local_development + enum status: { draft: 0, pending: 1, config: 2, ready: 3 } + + belongs_to :client, class_name: :Client, foreign_key: :name, primary_key: :name + validates :file_retention, :job_retention, :port, :password, presence: true validates :port, numericality: true validates :name, presence: true, uniqueness: true validate :fqdn_format - before_validation :set_retention, :unset_baculized + scope :not_baculized, -> { where(baculized: false) } + + before_validation :set_retention, :unset_baculized, :sanitize_name def to_bacula_config_array [ "Client {", " Name = #{name}", " Address = #{fqdn}", " FDPort = #{port}", " Catalog = #{CATALOG}", " Password = \"#{password}\"", " File Retention = #{file_retention} days", " Job Retention = #{job_retention} days", " AutoPrune = yes", "}" ] end private + def sanitize_name + self.name = fqdn + end + def set_retention self.file_retention = FILE_RETENTION_DAYS self.job_retention = JOB_RETENTION_DAYS end def unset_baculized - self.baculized = false + self.baculized = false if new_record? true end def fqdn_format regex = /(?=^.{4,253}$)(^((?!-)[a-zA-Z0-9-]{1,63}(?128}, using: :btree create_table "Counters", id: false, force: true do |t| t.binary "Counter", limit: 255, null: false t.integer "MinValue", default: 0 t.integer "MaxValue", default: 0 t.integer "CurrentValue", default: 0 t.binary "WrapCounter", limit: 255, null: false end create_table "Device", primary_key: "DeviceId", force: true do |t| t.binary "Name", limit: 255, null: false t.integer "MediaTypeId", default: 0 t.integer "StorageId", default: 0 t.integer "DevMounts", default: 0 t.integer "DevReadBytes", limit: 8, default: 0 t.integer "DevWriteBytes", limit: 8, default: 0 t.integer "DevReadBytesSinceCleaning", limit: 8, default: 0 t.integer "DevWriteBytesSinceCleaning", limit: 8, default: 0 t.integer "DevReadTime", limit: 8, default: 0 t.integer "DevWriteTime", limit: 8, default: 0 t.integer "DevReadTimeSinceCleaning", limit: 8, default: 0 t.integer "DevWriteTimeSinceCleaning", limit: 8, default: 0 t.datetime "CleaningDate" t.integer "CleaningPeriod", limit: 8, default: 0 end create_table "File", primary_key: "FileId", force: true do |t| t.integer "FileIndex", default: 0 t.integer "JobId", null: false t.integer "PathId", null: false t.integer "FilenameId", null: false t.integer "DeltaSeq", limit: 2, default: 0 t.integer "MarkId", default: 0 t.binary "LStat", limit: 255, null: false t.binary "MD5", limit: 255 end add_index "File", ["JobId", "PathId", "FilenameId"], name: "JobId_2", using: :btree add_index "File", ["JobId"], name: "JobId", using: :btree create_table "FileSet", primary_key: "FileSetId", force: true do |t| t.binary "FileSet", limit: 255, null: false t.binary "MD5", limit: 255 t.datetime "CreateTime" end create_table "Filename", primary_key: "FilenameId", force: true do |t| t.binary "Name", null: false end add_index "Filename", ["Name"], name: "Name", length: {"Name"=>255}, using: :btree create_table "Job", primary_key: "JobId", force: true do |t| t.binary "Job", limit: 255, null: false t.binary "Name", limit: 255, null: false t.binary "Type", limit: 1, null: false t.binary "Level", limit: 1, null: false t.integer "ClientId", default: 0 t.binary "JobStatus", limit: 1, null: false t.datetime "SchedTime" t.datetime "StartTime" t.datetime "EndTime" t.datetime "RealEndTime" t.integer "JobTDate", limit: 8, default: 0 t.integer "VolSessionId", default: 0 t.integer "VolSessionTime", default: 0 t.integer "JobFiles", default: 0 t.integer "JobBytes", limit: 8, default: 0 t.integer "ReadBytes", limit: 8, default: 0 t.integer "JobErrors", default: 0 t.integer "JobMissingFiles", default: 0 t.integer "PoolId", default: 0 t.integer "FileSetId", default: 0 t.integer "PriorJobId", default: 0 t.integer "PurgedFiles", limit: 1, default: 0 t.integer "HasBase", limit: 1, default: 0 t.integer "HasCache", limit: 1, default: 0 t.integer "Reviewed", limit: 1, default: 0 t.binary "Comment" end add_index "Job", ["Name"], name: "Name", length: {"Name"=>128}, using: :btree create_table "JobHisto", id: false, force: true do |t| t.integer "JobId", null: false t.binary "Job", limit: 255, null: false t.binary "Name", limit: 255, null: false t.binary "Type", limit: 1, null: false t.binary "Level", limit: 1, null: false t.integer "ClientId", default: 0 t.binary "JobStatus", limit: 1, null: false t.datetime "SchedTime" t.datetime "StartTime" t.datetime "EndTime" t.datetime "RealEndTime" t.integer "JobTDate", limit: 8, default: 0 t.integer "VolSessionId", default: 0 t.integer "VolSessionTime", default: 0 t.integer "JobFiles", default: 0 t.integer "JobBytes", limit: 8, default: 0 t.integer "ReadBytes", limit: 8, default: 0 t.integer "JobErrors", default: 0 t.integer "JobMissingFiles", default: 0 t.integer "PoolId", default: 0 t.integer "FileSetId", default: 0 t.integer "PriorJobId", default: 0 t.integer "PurgedFiles", limit: 1, default: 0 t.integer "HasBase", limit: 1, default: 0 t.integer "HasCache", limit: 1, default: 0 t.integer "Reviewed", limit: 1, default: 0 t.binary "Comment" end add_index "JobHisto", ["JobId"], name: "JobId", using: :btree add_index "JobHisto", ["StartTime"], name: "StartTime", using: :btree create_table "JobMedia", primary_key: "JobMediaId", force: true do |t| t.integer "JobId", null: false t.integer "MediaId", null: false t.integer "FirstIndex", default: 0 t.integer "LastIndex", default: 0 t.integer "StartFile", default: 0 t.integer "EndFile", default: 0 t.integer "StartBlock", default: 0 t.integer "EndBlock", default: 0 t.integer "VolIndex", default: 0 end add_index "JobMedia", ["JobId", "MediaId"], name: "JobId", using: :btree create_table "Location", primary_key: "LocationId", force: true do |t| t.binary "Location", limit: 255, null: false t.integer "Cost", default: 0 t.integer "Enabled", limit: 1 end create_table "LocationLog", primary_key: "LocLogId", force: true do |t| t.datetime "Date" t.binary "Comment", null: false t.integer "MediaId", default: 0 t.integer "LocationId", default: 0 t.string "NewVolStatus", limit: 9, null: false t.integer "NewEnabled", limit: 1 end create_table "Log", primary_key: "LogId", force: true do |t| t.integer "JobId", default: 0 t.datetime "Time" t.binary "LogText", null: false end add_index "Log", ["JobId"], name: "JobId", using: :btree create_table "Media", primary_key: "MediaId", force: true do |t| t.binary "VolumeName", limit: 255, null: false t.integer "Slot", default: 0 t.integer "PoolId", default: 0 t.binary "MediaType", limit: 255, null: false t.integer "MediaTypeId", default: 0 t.integer "LabelType", limit: 1, default: 0 t.datetime "FirstWritten" t.datetime "LastWritten" t.datetime "LabelDate" t.integer "VolJobs", default: 0 t.integer "VolFiles", default: 0 t.integer "VolBlocks", default: 0 t.integer "VolMounts", default: 0 t.integer "VolBytes", limit: 8, default: 0 t.integer "VolParts", default: 0 t.integer "VolErrors", default: 0 t.integer "VolWrites", default: 0 t.integer "VolCapacityBytes", limit: 8, default: 0 t.string "VolStatus", limit: 9, null: false t.integer "Enabled", limit: 1, default: 1 t.integer "Recycle", limit: 1, default: 0 t.integer "ActionOnPurge", limit: 1, default: 0 t.integer "VolRetention", limit: 8, default: 0 t.integer "VolUseDuration", limit: 8, default: 0 t.integer "MaxVolJobs", default: 0 t.integer "MaxVolFiles", default: 0 t.integer "MaxVolBytes", limit: 8, default: 0 t.integer "InChanger", limit: 1, default: 0 t.integer "StorageId", default: 0 t.integer "DeviceId", default: 0 t.integer "MediaAddressing", limit: 1, default: 0 t.integer "VolReadTime", limit: 8, default: 0 t.integer "VolWriteTime", limit: 8, default: 0 t.integer "EndFile", default: 0 t.integer "EndBlock", default: 0 t.integer "LocationId", default: 0 t.integer "RecycleCount", default: 0 t.datetime "InitialWrite" t.integer "ScratchPoolId", default: 0 t.integer "RecyclePoolId", default: 0 t.binary "Comment" end add_index "Media", ["PoolId"], name: "PoolId", using: :btree add_index "Media", ["VolumeName"], name: "VolumeName", unique: true, length: {"VolumeName"=>128}, using: :btree create_table "MediaType", primary_key: "MediaTypeId", force: true do |t| t.binary "MediaType", limit: 255, null: false t.integer "ReadOnly", limit: 1, default: 0 end create_table "Path", primary_key: "PathId", force: true do |t| t.binary "Path", null: false end add_index "Path", ["Path"], name: "Path", length: {"Path"=>255}, using: :btree create_table "PathHierarchy", primary_key: "PathId", force: true do |t| t.integer "PPathId", null: false end add_index "PathHierarchy", ["PPathId"], name: "pathhierarchy_ppathid", using: :btree create_table "PathVisibility", id: false, force: true do |t| t.integer "PathId", null: false t.integer "JobId", null: false t.integer "Size", limit: 8, default: 0 t.integer "Files", default: 0 end add_index "PathVisibility", ["JobId"], name: "pathvisibility_jobid", using: :btree create_table "Pool", primary_key: "PoolId", force: true do |t| t.binary "Name", limit: 255, null: false t.integer "NumVols", default: 0 t.integer "MaxVols", default: 0 t.integer "UseOnce", limit: 1, default: 0 t.integer "UseCatalog", limit: 1, default: 0 t.integer "AcceptAnyVolume", limit: 1, default: 0 t.integer "VolRetention", limit: 8, default: 0 t.integer "VolUseDuration", limit: 8, default: 0 t.integer "MaxVolJobs", default: 0 t.integer "MaxVolFiles", default: 0 t.integer "MaxVolBytes", limit: 8, default: 0 t.integer "AutoPrune", limit: 1, default: 0 t.integer "Recycle", limit: 1, default: 0 t.integer "ActionOnPurge", limit: 1, default: 0 t.string "PoolType", limit: 9, null: false t.integer "LabelType", limit: 1, default: 0 t.binary "LabelFormat", limit: 255 t.integer "Enabled", limit: 1, default: 1 t.integer "ScratchPoolId", default: 0 t.integer "RecyclePoolId", default: 0 t.integer "NextPoolId", default: 0 t.integer "MigrationHighBytes", limit: 8, default: 0 t.integer "MigrationLowBytes", limit: 8, default: 0 t.integer "MigrationTime", limit: 8, default: 0 end add_index "Pool", ["Name"], name: "Name", unique: true, length: {"Name"=>128}, using: :btree create_table "RestoreObject", primary_key: "RestoreObjectId", force: true do |t| t.binary "ObjectName", null: false t.binary "RestoreObject", limit: 2147483647, null: false t.binary "PluginName", limit: 255, null: false t.integer "ObjectLength", default: 0 t.integer "ObjectFullLength", default: 0 t.integer "ObjectIndex", default: 0 t.integer "ObjectType", default: 0 t.integer "FileIndex", default: 0 t.integer "JobId", null: false t.integer "ObjectCompression", default: 0 end add_index "RestoreObject", ["JobId"], name: "JobId", using: :btree create_table "Status", primary_key: "JobStatus", force: true do |t| t.binary "JobStatusLong" t.integer "Severity" end create_table "Storage", primary_key: "StorageId", force: true do |t| t.binary "Name", limit: 255, null: false t.integer "AutoChanger", limit: 1, default: 0 end create_table "UnsavedFiles", primary_key: "UnsavedId", force: true do |t| t.integer "JobId", null: false t.integer "PathId", null: false t.integer "FilenameId", null: false end create_table "Version", id: false, force: true do |t| t.integer "VersionId", null: false end create_table "hosts", force: true do |t| t.binary "name", limit: 255, null: false t.binary "fqdn", limit: 255, null: false t.integer "port", null: false t.integer "file_retention", null: false t.integer "job_retention", null: false t.datetime "created_at" t.datetime "updated_at" t.string "password" t.boolean "baculized", default: false, null: false t.datetime "baculized_at" + t.integer "status", limit: 1, default: 0 + t.integer "client_id" end add_index "hosts", ["name"], name: "index_hosts_on_name", unique: true, length: {"name"=>128}, using: :btree create_table "users", force: true do |t| t.string "username", null: false t.string "email" t.integer "user_type", limit: 1, null: false t.boolean "enabled", default: false t.datetime "created_at" t.datetime "updated_at" end end diff --git a/spec/models/host_spec.rb b/spec/models/host_spec.rb index 3292fc0..8de5d9a 100644 --- a/spec/models/host_spec.rb +++ b/spec/models/host_spec.rb @@ -1,70 +1,75 @@ require 'spec_helper' describe Host do context 'validates' do - [:name, :password].each do |field| - it "presence of #{field}" do - expect(Host.new).to have(1).errors_on(field) - end + it "presence of Password" do + expect(Host.new).to have(1).errors_on(:password) end it 'numericality of :port' do expect(Host.new(port: :lala)).to have(2).errors_on(:port) end - [:file_retention, :job_retention].each do |field| + [:file_retention, :job_retention, :name].each do |field| it "#{field} is set automatically" do - host = Host.new + host = Host.new(fqdn: 'test') host.valid? expect(host.send(field)).to be_present end end end context 'when fqdn is invalid' do let(:host) { FactoryGirl.build(:host, fqdn: :lala) } it 'has errors' do expect(host).to have(1).errors_on(:fqdn) end end + context 'name field' do + let(:host) { FactoryGirl.create(:host, name: nil) } + it 'is generated by the system' do + expect(host.name).to be + end + end + describe '#to_bacula_config_array' do let(:host) { FactoryGirl.create(:host) } it "is a valid client directive" do expect(host.to_bacula_config_array).to include('Client {') expect(host.to_bacula_config_array).to include('}') end it "contains Address directive" do expect(host.to_bacula_config_array).to include(" Address = #{host.fqdn}") end it "contains FDPort directive" do expect(host.to_bacula_config_array).to include(" FDPort = #{host.port}") end it "contains Catalog directive" do expect(host.to_bacula_config_array).to include(" Catalog = #{Host::CATALOG}") end it "contains Password directive" do expect(host.to_bacula_config_array).to include(" Password = \"#{host.password}\"") end it "contains File Retention directive" do expect(host.to_bacula_config_array). to include(" File Retention = #{host.file_retention} days") end it "contains Job Retention directive" do expect(host.to_bacula_config_array). to include(" Job Retention = #{host.job_retention} days") end it "contains AutoPrune directive" do expect(host.to_bacula_config_array).to include(" AutoPrune = yes") end end end