Class: Backup::S3
- Inherits:
-
Object
- Object
- Backup::S3
- Includes:
- S3Helper
- Defined in:
- lib/capistrano/ops/rails/lib/backup/s3.rb
Instance Attribute Summary collapse
-
#access_key_id ⇒ Object
Returns the value of attribute access_key_id.
-
#endpoint ⇒ Object
Returns the value of attribute endpoint.
-
#region ⇒ Object
Returns the value of attribute region.
-
#s3_client ⇒ Object
Returns the value of attribute s3_client.
-
#s3_resource ⇒ Object
Returns the value of attribute s3_resource.
-
#secret_access_key ⇒ Object
Returns the value of attribute secret_access_key.
Instance Method Summary collapse
-
#initialize(endpoint: ENV['S3_BACKUP_ENDPOINT'], region: ENV['S3_BACKUP_REGION'], access_key_id: ENV['S3_BACKUP_KEY'], secret_access_key: ENV['S3_BACKUP_SECRET']) ⇒ S3
constructor
A new instance of S3.
- #remove_old_backups(basename, keep: 5) ⇒ Object
-
#start_writer_thread(folder_path, write_io) ⇒ Object
rubocop:enable Metrics/MethodLength.
- #upload(backup_file, key) ⇒ Object
-
#upload_file_as_stream(file_path, key) ⇒ Object
rubocop:disable Metrics/MethodLength.
- #upload_folder_as_tar_gz_stream(folder_path, key) ⇒ Object
- #upload_stream(backup_file, key, type) ⇒ Object
Methods included from S3Helper
#calculate_chunk_size, #calculate_total_size, #get_current_month, #get_delete_items, #get_items_by_month, #get_months, #get_old_months
Constructor Details
#initialize(endpoint: ENV['S3_BACKUP_ENDPOINT'], region: ENV['S3_BACKUP_REGION'], access_key_id: ENV['S3_BACKUP_KEY'], secret_access_key: ENV['S3_BACKUP_SECRET']) ⇒ S3
Returns a new instance of S3.
15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 |
# File 'lib/capistrano/ops/rails/lib/backup/s3.rb', line 15 def initialize(endpoint: ENV['S3_BACKUP_ENDPOINT'], region: ENV['S3_BACKUP_REGION'], access_key_id: ENV['S3_BACKUP_KEY'], secret_access_key: ENV['S3_BACKUP_SECRET']) self.endpoint = endpoint self.region = region self.access_key_id = access_key_id self.secret_access_key = secret_access_key configuration = { region: region, access_key_id: access_key_id, secret_access_key: secret_access_key, force_path_style: true } configuration[:endpoint] = endpoint unless endpoint.nil? self.s3_resource = Aws::S3::Resource.new(configuration) self.s3_client = Aws::S3::Client.new(configuration) end |
Instance Attribute Details
#access_key_id ⇒ Object
Returns the value of attribute access_key_id.
13 14 15 |
# File 'lib/capistrano/ops/rails/lib/backup/s3.rb', line 13 def access_key_id @access_key_id end |
#endpoint ⇒ Object
Returns the value of attribute endpoint.
13 14 15 |
# File 'lib/capistrano/ops/rails/lib/backup/s3.rb', line 13 def endpoint @endpoint end |
#region ⇒ Object
Returns the value of attribute region.
13 14 15 |
# File 'lib/capistrano/ops/rails/lib/backup/s3.rb', line 13 def region @region end |
#s3_client ⇒ Object
Returns the value of attribute s3_client.
13 14 15 |
# File 'lib/capistrano/ops/rails/lib/backup/s3.rb', line 13 def s3_client @s3_client end |
#s3_resource ⇒ Object
Returns the value of attribute s3_resource.
13 14 15 |
# File 'lib/capistrano/ops/rails/lib/backup/s3.rb', line 13 def s3_resource @s3_resource end |
#secret_access_key ⇒ Object
Returns the value of attribute secret_access_key.
13 14 15 |
# File 'lib/capistrano/ops/rails/lib/backup/s3.rb', line 13 def secret_access_key @secret_access_key end |
Instance Method Details
#remove_old_backups(basename, keep: 5) ⇒ Object
258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 |
# File 'lib/capistrano/ops/rails/lib/backup/s3.rb', line 258 def remove_old_backups(basename, keep: 5) all_items = s3_resource.bucket(ENV['S3_BACKUP_BUCKET']).objects(prefix: basename).map do |item| { key: item.key, last_modified: item.last_modified } end all_items = all_items.sort_by { |hsh| hsh[:last_modified] }.reverse months = get_months(all_items) old_months = get_old_months(months) if old_months&.empty? || old_months.nil? puts 'No old months to remove' else old_months.each do |month| items = get_items_by_month(all_items, month) delete_items = get_delete_items(items, 1) puts "Removing #{month} from S3" delete_items.each do |item_obj| puts "Removing #{item_obj[:key]} from S3" s3_resource.bucket(ENV['S3_BACKUP_BUCKET']).object(item_obj[:key]).delete end end puts 'Old months removed from S3' end current_month = get_current_month(all_items) current_month_delete_items = get_delete_items(current_month, keep) if current_month_delete_items&.empty? || current_month_delete_items.nil? puts 'No old backups to remove' exit(0) end current_month_delete_items.each do |item_obj| puts "Removing #{item_obj[:key]} from S3" s3_resource.bucket(ENV['S3_BACKUP_BUCKET']).object(item_obj[:key]).delete end puts 'Old backups removed from S3' rescue Backup::Error => e puts "Remove failed: #{e.}" raise e end |
#start_writer_thread(folder_path, write_io) ⇒ Object
rubocop:enable Metrics/MethodLength
226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 |
# File 'lib/capistrano/ops/rails/lib/backup/s3.rb', line 226 def start_writer_thread(folder_path, write_io) Thread.new do parent_folder = File.dirname(folder_path) folder_name = File.basename(folder_path) Zlib::GzipWriter.wrap(write_io) do |gz| Gem::Package::TarWriter.new(gz) do |tar| Dir.chdir(parent_folder) do Find.find(folder_name) do |file_path| if File.directory?(file_path) tar.mkdir(file_path, File.stat(file_path).mode) else mode = File.stat(file_path).mode tar.add_file_simple(file_path, mode, File.size(file_path)) do |tar_file| File.open(file_path, 'rb') do |f| while (chunk = f.read(1024 * 1024)) # Read in 1MB chunks tar_file.write(chunk) end end end end end end end end rescue StandardError => e puts "Error writing tar.gz data: #{e.}" ensure write_io.close unless write_io.closed? end end |
#upload(backup_file, key) ⇒ Object
32 33 34 35 36 37 38 39 40 |
# File 'lib/capistrano/ops/rails/lib/backup/s3.rb', line 32 def upload(backup_file, key) begin s3_resource.bucket(ENV['S3_BACKUP_BUCKET']).object(key).upload_file(backup_file) rescue Backup::Error => e puts "Upload failed: #{e.}" raise e end 'File uploaded to S3' end |
#upload_file_as_stream(file_path, key) ⇒ Object
rubocop:disable Metrics/MethodLength
51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 |
# File 'lib/capistrano/ops/rails/lib/backup/s3.rb', line 51 def upload_file_as_stream(file_path, key) bucket = ENV['S3_BACKUP_BUCKET'] # Calculate total size of the file to be uploaded total_size = File.size(file_path) chunk_size = calculate_chunk_size(total_size) uploaded_size = 0 # Upload the tar.gz data from the file in parts part_number = 1 parts = [] last_logged_progress = 0 max_retry_time = 300 # 5 minutes in seconds total_wait_time = 0 begin multipart_upload ||= s3_client.create_multipart_upload(bucket: bucket, key: key) File.open(file_path, 'rb') do |file| while (part = file.read(chunk_size)) # Read calculated chunk size retry_count = 0 begin # Initiate multipart upload part_upload = s3_client.upload_part( bucket: bucket, key: key, upload_id: multipart_upload.upload_id, part_number: part_number, body: part ) parts << { part_number: part_number, etag: part_upload.etag } uploaded_size += part.size part_number += 1 progress = (uploaded_size.to_f / total_size * 100).round if progress >= last_logged_progress + 10 puts "Upload progress: #{progress}% complete" last_logged_progress = progress end rescue StandardError => e retry_count += 1 wait_time = 2**retry_count total_wait_time += wait_time if total_wait_time > max_retry_time puts "Exceeded maximum retry time of #{max_retry_time / 60} minutes. Aborting upload." raise e end puts "Error uploading part #{part_number}: #{e..split("\n").first} (Attempt #{retry_count})" puts "Retry in #{wait_time} seconds" sleep(wait_time) # Exponential backoff puts 'Retrying upload part...' retry end end end # Complete multipart upload s3_client.complete_multipart_upload( bucket: bucket, key: key, upload_id: multipart_upload.upload_id, multipart_upload: { parts: parts } ) puts 'Completed multipart upload' rescue StandardError => e # Abort multipart upload in case of error s3_client.abort_multipart_upload( bucket: bucket, key: key, upload_id: multipart_upload.upload_id ) puts "Aborted multipart upload due to error: #{e.}" raise e end 'File uploaded to S3 as tar.gz' rescue StandardError => e puts "Upload failed: #{e.}" raise e end |
#upload_folder_as_tar_gz_stream(folder_path, key) ⇒ Object
132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 |
# File 'lib/capistrano/ops/rails/lib/backup/s3.rb', line 132 def upload_folder_as_tar_gz_stream(folder_path, key) bucket = ENV['S3_BACKUP_BUCKET'] # Calculate total size of the files to be uploaded total_size = calculate_total_size(folder_path) chunk_size = calculate_chunk_size(total_size) # Create a pipe to stream data read_io, write_io = IO.pipe read_io.binmode write_io.binmode uploaded_size = 0 # Start a thread to write the tar.gz data to the pipe writer_thread = start_writer_thread(folder_path, write_io) # Upload the tar.gz data from the pipe in parts part_number = 1 parts = [] last_logged_progress = 0 max_retry_time = 300 # 5 minutes in seconds total_wait_time = 0 begin # Initiate multipart upload multipart_upload ||= s3_client.create_multipart_upload(bucket: bucket, key: key) while (part = read_io.read(chunk_size)) # Read calculated chunk size retry_count = 0 begin part_upload = s3_client.upload_part( bucket: bucket, key: key, upload_id: multipart_upload.upload_id, part_number: part_number, body: part ) parts << { part_number: part_number, etag: part_upload.etag } uploaded_size += part.size part_number += 1 progress = (uploaded_size.to_f / total_size * 100).round if progress >= last_logged_progress + 10 puts "Upload progress: #{progress}% complete" last_logged_progress = progress end rescue StandardError => e retry_count += 1 wait_time = 2**retry_count total_wait_time += wait_time if total_wait_time > max_retry_time puts "Exceeded maximum retry time of #{max_retry_time / 60} minutes. Aborting upload." raise e end puts "Error uploading part #{part_number}: #{e..split("\n").first} (Attempt #{retry_count})" puts "Retry in #{wait_time} seconds" sleep(wait_time) # Exponential backoff puts 'Retrying upload part...' retry end end # Complete multipart upload s3_client.complete_multipart_upload( bucket: bucket, key: key, upload_id: multipart_upload.upload_id, multipart_upload: { parts: parts } ) puts 'Completed multipart upload' rescue StandardError => e # Abort multipart upload in case of error if multipart_upload s3_client.abort_multipart_upload( bucket: bucket, key: key, upload_id: multipart_upload.upload_id ) end puts "Aborted multipart upload due to error: #{e.}" raise e ensure read_io.close unless read_io.closed? writer_thread.join end 'Folder uploaded to S3 as tar.gz' rescue StandardError => e puts "Upload failed: #{e.}" raise e end |
#upload_stream(backup_file, key, type) ⇒ Object
42 43 44 45 46 47 48 |
# File 'lib/capistrano/ops/rails/lib/backup/s3.rb', line 42 def upload_stream(backup_file, key, type) if type == 'folder' upload_folder_as_tar_gz_stream(backup_file, key) else upload_file_as_stream(backup_file, key) end end |