2021-07-27 17:42:25 -05:00
|
|
|
# frozen_string_literal: true
|
|
|
|
|
|
|
|
require "digest/sha1"
|
|
|
|
|
|
|
|
class ExternalUploadStub < ActiveRecord::Base
|
|
|
|
CREATED_EXPIRY_HOURS = 1
|
|
|
|
UPLOADED_EXPIRY_HOURS = 24
|
FEATURE: Uppy direct S3 multipart uploads in composer (#14051)
This pull request introduces the endpoints required, and the JavaScript functionality in the `ComposerUppyUpload` mixin, for direct S3 multipart uploads. There are four new endpoints in the uploads controller:
* `create-multipart.json` - Creates the multipart upload in S3 along with an `ExternalUploadStub` record, storing information about the file in the same way as `generate-presigned-put.json` does for regular direct S3 uploads
* `batch-presign-multipart-parts.json` - Takes a list of part numbers and the unique identifier for an `ExternalUploadStub` record, and generates the presigned URLs for those parts if the multipart upload still exists and if the user has permission to access that upload
* `complete-multipart.json` - Completes the multipart upload in S3. Needs the full list of part numbers and their associated ETags which are returned when the part is uploaded to the presigned URL above. Only works if the user has permission to access the associated `ExternalUploadStub` record and the multipart upload still exists.
After we confirm the upload is complete in S3, we go through the regular `UploadCreator` flow, the same as `complete-external-upload.json`, and promote the temporary upload S3 into a full `Upload` record, moving it to its final destination.
* `abort-multipart.json` - Aborts the multipart upload on S3 and destroys the `ExternalUploadStub` record if the user has permission to access that upload.
Also added are a few new columns to `ExternalUploadStub`:
* multipart - Whether or not this is a multipart upload
* external_upload_identifier - The "upload ID" for an S3 multipart upload
* filesize - The size of the file when the `create-multipart.json` or `generate-presigned-put.json` is called. This is used for validation.
When the user completes a direct S3 upload, either regular or multipart, we take the `filesize` that was captured when the `ExternalUploadStub` was first created and compare it with the final `Content-Length` size of the file where it is stored in S3. Then, if the two do not match, we throw an error, delete the file on S3, and ban the user from uploading files for N (default 5) minutes. This would only happen if the user uploads a different file than what they first specified, or in the case of multipart uploads uploaded larger chunks than needed. This is done to prevent abuse of S3 storage by bad actors.
Also included in this PR is an update to vendor/uppy.js. This has been built locally from the latest uppy source at https://github.com/transloadit/uppy/commit/d613b849a6591083f8a0968aa8d66537e231bbcd. This must be done so that I can get my multipart upload changes into Discourse. When the Uppy team cuts a proper release, we can bump the package.json versions instead.
2021-08-24 17:46:54 -05:00
|
|
|
FAILED_EXPIRY_HOURS = 48
|
2021-07-27 17:42:25 -05:00
|
|
|
|
2023-01-09 06:20:10 -06:00
|
|
|
belongs_to :created_by, class_name: "User"
|
|
|
|
|
|
|
|
validates :filesize,
|
|
|
|
numericality: {
|
|
|
|
allow_nil: false,
|
|
|
|
only_integer: true,
|
|
|
|
greater_than_or_equal_to: 1,
|
|
|
|
}
|
|
|
|
|
|
|
|
scope :expired_created,
|
|
|
|
-> {
|
|
|
|
where(
|
|
|
|
"status = ? AND created_at <= ?",
|
|
|
|
ExternalUploadStub.statuses[:created],
|
|
|
|
CREATED_EXPIRY_HOURS.hours.ago,
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
|
|
|
scope :expired_uploaded,
|
|
|
|
-> {
|
|
|
|
where(
|
|
|
|
"status = ? AND created_at <= ?",
|
|
|
|
ExternalUploadStub.statuses[:uploaded],
|
|
|
|
UPLOADED_EXPIRY_HOURS.hours.ago,
|
|
|
|
)
|
|
|
|
}
|
2021-07-27 17:42:25 -05:00
|
|
|
|
|
|
|
before_create do
|
|
|
|
self.unique_identifier = SecureRandom.uuid
|
|
|
|
self.status = ExternalUploadStub.statuses[:created] if self.status.blank?
|
|
|
|
end
|
|
|
|
|
|
|
|
def self.statuses
|
2023-01-09 06:20:10 -06:00
|
|
|
@statuses ||= Enum.new(created: 1, uploaded: 2, failed: 3)
|
2021-07-27 17:42:25 -05:00
|
|
|
end
|
|
|
|
|
|
|
|
def self.cleanup!
|
|
|
|
expired_created.delete_all
|
|
|
|
expired_uploaded.delete_all
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
# == Schema Information
|
|
|
|
#
|
|
|
|
# Table name: external_upload_stubs
|
|
|
|
#
|
FEATURE: Uppy direct S3 multipart uploads in composer (#14051)
This pull request introduces the endpoints required, and the JavaScript functionality in the `ComposerUppyUpload` mixin, for direct S3 multipart uploads. There are four new endpoints in the uploads controller:
* `create-multipart.json` - Creates the multipart upload in S3 along with an `ExternalUploadStub` record, storing information about the file in the same way as `generate-presigned-put.json` does for regular direct S3 uploads
* `batch-presign-multipart-parts.json` - Takes a list of part numbers and the unique identifier for an `ExternalUploadStub` record, and generates the presigned URLs for those parts if the multipart upload still exists and if the user has permission to access that upload
* `complete-multipart.json` - Completes the multipart upload in S3. Needs the full list of part numbers and their associated ETags which are returned when the part is uploaded to the presigned URL above. Only works if the user has permission to access the associated `ExternalUploadStub` record and the multipart upload still exists.
After we confirm the upload is complete in S3, we go through the regular `UploadCreator` flow, the same as `complete-external-upload.json`, and promote the temporary upload S3 into a full `Upload` record, moving it to its final destination.
* `abort-multipart.json` - Aborts the multipart upload on S3 and destroys the `ExternalUploadStub` record if the user has permission to access that upload.
Also added are a few new columns to `ExternalUploadStub`:
* multipart - Whether or not this is a multipart upload
* external_upload_identifier - The "upload ID" for an S3 multipart upload
* filesize - The size of the file when the `create-multipart.json` or `generate-presigned-put.json` is called. This is used for validation.
When the user completes a direct S3 upload, either regular or multipart, we take the `filesize` that was captured when the `ExternalUploadStub` was first created and compare it with the final `Content-Length` size of the file where it is stored in S3. Then, if the two do not match, we throw an error, delete the file on S3, and ban the user from uploading files for N (default 5) minutes. This would only happen if the user uploads a different file than what they first specified, or in the case of multipart uploads uploaded larger chunks than needed. This is done to prevent abuse of S3 storage by bad actors.
Also included in this PR is an update to vendor/uppy.js. This has been built locally from the latest uppy source at https://github.com/transloadit/uppy/commit/d613b849a6591083f8a0968aa8d66537e231bbcd. This must be done so that I can get my multipart upload changes into Discourse. When the Uppy team cuts a proper release, we can bump the package.json versions instead.
2021-08-24 17:46:54 -05:00
|
|
|
# id :bigint not null, primary key
|
|
|
|
# key :string not null
|
|
|
|
# original_filename :string not null
|
|
|
|
# status :integer default(1), not null
|
|
|
|
# unique_identifier :uuid not null
|
|
|
|
# created_by_id :integer not null
|
|
|
|
# upload_type :string not null
|
|
|
|
# created_at :datetime not null
|
|
|
|
# updated_at :datetime not null
|
|
|
|
# multipart :boolean default(FALSE), not null
|
|
|
|
# external_upload_identifier :string
|
|
|
|
# filesize :bigint not null
|
2021-07-27 17:42:25 -05:00
|
|
|
#
|
|
|
|
# Indexes
|
|
|
|
#
|
FEATURE: Uppy direct S3 multipart uploads in composer (#14051)
This pull request introduces the endpoints required, and the JavaScript functionality in the `ComposerUppyUpload` mixin, for direct S3 multipart uploads. There are four new endpoints in the uploads controller:
* `create-multipart.json` - Creates the multipart upload in S3 along with an `ExternalUploadStub` record, storing information about the file in the same way as `generate-presigned-put.json` does for regular direct S3 uploads
* `batch-presign-multipart-parts.json` - Takes a list of part numbers and the unique identifier for an `ExternalUploadStub` record, and generates the presigned URLs for those parts if the multipart upload still exists and if the user has permission to access that upload
* `complete-multipart.json` - Completes the multipart upload in S3. Needs the full list of part numbers and their associated ETags which are returned when the part is uploaded to the presigned URL above. Only works if the user has permission to access the associated `ExternalUploadStub` record and the multipart upload still exists.
After we confirm the upload is complete in S3, we go through the regular `UploadCreator` flow, the same as `complete-external-upload.json`, and promote the temporary upload S3 into a full `Upload` record, moving it to its final destination.
* `abort-multipart.json` - Aborts the multipart upload on S3 and destroys the `ExternalUploadStub` record if the user has permission to access that upload.
Also added are a few new columns to `ExternalUploadStub`:
* multipart - Whether or not this is a multipart upload
* external_upload_identifier - The "upload ID" for an S3 multipart upload
* filesize - The size of the file when the `create-multipart.json` or `generate-presigned-put.json` is called. This is used for validation.
When the user completes a direct S3 upload, either regular or multipart, we take the `filesize` that was captured when the `ExternalUploadStub` was first created and compare it with the final `Content-Length` size of the file where it is stored in S3. Then, if the two do not match, we throw an error, delete the file on S3, and ban the user from uploading files for N (default 5) minutes. This would only happen if the user uploads a different file than what they first specified, or in the case of multipart uploads uploaded larger chunks than needed. This is done to prevent abuse of S3 storage by bad actors.
Also included in this PR is an update to vendor/uppy.js. This has been built locally from the latest uppy source at https://github.com/transloadit/uppy/commit/d613b849a6591083f8a0968aa8d66537e231bbcd. This must be done so that I can get my multipart upload changes into Discourse. When the Uppy team cuts a proper release, we can bump the package.json versions instead.
2021-08-24 17:46:54 -05:00
|
|
|
# index_external_upload_stubs_on_created_by_id (created_by_id)
|
|
|
|
# index_external_upload_stubs_on_external_upload_identifier (external_upload_identifier)
|
|
|
|
# index_external_upload_stubs_on_key (key) UNIQUE
|
|
|
|
# index_external_upload_stubs_on_status (status)
|
2021-07-27 17:42:25 -05:00
|
|
|
#
|