mirror of
https://github.com/mattermost/mattermost.git
synced 2025-02-25 18:55:24 -06:00
* Implement AppendFile for FileBackend * Split test into subtests * [MM-26812] Add support for resumable file uploads (#15252) * Implement UploadSession * Implement UploadSessionStore * Add error strings * Implement resumable file uploads * Add UploadType * Fix retry layer tests * Regenerate store layers * Fix store error handling * Use base for filename * Prevent concurrent uploads on the same upload session * Fix erroneus error string * Improve error handling Co-authored-by: Mattermod <mattermod@users.noreply.github.com> * Fix translations Co-authored-by: Mattermod <mattermod@users.noreply.github.com>
309 lines
7.4 KiB
Go
309 lines
7.4 KiB
Go
// Copyright (c) 2015-present Mattermost, Inc. All Rights Reserved.
|
|
// See LICENSE.txt for license information.
|
|
|
|
package app
|
|
|
|
import (
|
|
"bytes"
|
|
"io"
|
|
"io/ioutil"
|
|
"math/rand"
|
|
"path/filepath"
|
|
"sync"
|
|
"sync/atomic"
|
|
"testing"
|
|
|
|
"github.com/mattermost/mattermost-server/v5/model"
|
|
"github.com/mattermost/mattermost-server/v5/utils/fileutils"
|
|
|
|
"github.com/stretchr/testify/require"
|
|
)
|
|
|
|
func TestCreateUploadSession(t *testing.T) {
|
|
th := Setup(t).InitBasic()
|
|
defer th.TearDown()
|
|
|
|
us := &model.UploadSession{
|
|
Type: model.UploadTypeAttachment,
|
|
UserId: th.BasicUser.Id,
|
|
ChannelId: th.BasicChannel.Id,
|
|
Filename: "upload",
|
|
FileSize: 8 * 1024 * 1024,
|
|
}
|
|
|
|
t.Run("FileSize over limit", func(t *testing.T) {
|
|
maxFileSize := *th.App.Config().FileSettings.MaxFileSize
|
|
th.App.UpdateConfig(func(cfg *model.Config) { *cfg.FileSettings.MaxFileSize = us.FileSize - 1 })
|
|
defer th.App.UpdateConfig(func(cfg *model.Config) { *cfg.FileSettings.MaxFileSize = maxFileSize })
|
|
u, err := th.App.CreateUploadSession(us)
|
|
require.NotNil(t, err)
|
|
require.Equal(t, "app.upload.create.upload_too_large.app_error", err.Id)
|
|
require.Nil(t, u)
|
|
})
|
|
|
|
t.Run("invalid Id", func(t *testing.T) {
|
|
u, err := th.App.CreateUploadSession(us)
|
|
require.NotNil(t, err)
|
|
require.Equal(t, "model.upload_session.is_valid.id.app_error", err.Id)
|
|
require.Nil(t, u)
|
|
})
|
|
|
|
t.Run("invalid UserId", func(t *testing.T) {
|
|
us.Id = model.NewId()
|
|
us.UserId = ""
|
|
u, err := th.App.CreateUploadSession(us)
|
|
require.NotNil(t, err)
|
|
require.Equal(t, "model.upload_session.is_valid.user_id.app_error", err.Id)
|
|
require.Nil(t, u)
|
|
})
|
|
|
|
t.Run("invalid ChannelId", func(t *testing.T) {
|
|
us.UserId = th.BasicUser.Id
|
|
us.ChannelId = ""
|
|
u, err := th.App.CreateUploadSession(us)
|
|
require.NotNil(t, err)
|
|
require.Equal(t, "model.upload_session.is_valid.channel_id.app_error", err.Id)
|
|
require.Nil(t, u)
|
|
})
|
|
|
|
t.Run("non-existing channel", func(t *testing.T) {
|
|
us.ChannelId = model.NewId()
|
|
u, err := th.App.CreateUploadSession(us)
|
|
require.NotNil(t, err)
|
|
require.Equal(t, "app.upload.create.incorrect_channel_id.app_error", err.Id)
|
|
require.Nil(t, u)
|
|
})
|
|
|
|
t.Run("deleted channel", func(t *testing.T) {
|
|
ch := th.CreateChannel(th.BasicTeam)
|
|
th.App.DeleteChannel(ch, th.BasicUser.Id)
|
|
us.ChannelId = ch.Id
|
|
u, err := th.App.CreateUploadSession(us)
|
|
require.NotNil(t, err)
|
|
require.Equal(t, "app.upload.create.cannot_upload_to_deleted_channel.app_error", err.Id)
|
|
require.Nil(t, u)
|
|
})
|
|
|
|
t.Run("success", func(t *testing.T) {
|
|
us.ChannelId = th.BasicChannel.Id
|
|
u, err := th.App.CreateUploadSession(us)
|
|
require.Nil(t, err)
|
|
require.NotEmpty(t, u)
|
|
})
|
|
}
|
|
|
|
func TestUploadData(t *testing.T) {
|
|
th := Setup(t).InitBasic()
|
|
defer th.TearDown()
|
|
|
|
us := &model.UploadSession{
|
|
Id: model.NewId(),
|
|
Type: model.UploadTypeAttachment,
|
|
UserId: th.BasicUser.Id,
|
|
ChannelId: th.BasicChannel.Id,
|
|
Filename: "upload",
|
|
FileSize: 8 * 1024 * 1024,
|
|
}
|
|
|
|
var err error
|
|
us, err = th.App.CreateUploadSession(us)
|
|
require.Nil(t, err)
|
|
require.NotEmpty(t, us)
|
|
|
|
data := make([]byte, us.FileSize)
|
|
_, err2 := rand.Read(data)
|
|
require.NoError(t, err2)
|
|
|
|
t.Run("write error", func(t *testing.T) {
|
|
rd := &io.LimitedReader{
|
|
R: bytes.NewReader(data),
|
|
N: 1024 * 1024,
|
|
}
|
|
|
|
ok, err := th.App.FileExists(us.Path)
|
|
require.False(t, ok)
|
|
require.Nil(t, err)
|
|
|
|
u := *us
|
|
u.Path = ""
|
|
info, err := th.App.UploadData(&u, rd)
|
|
require.Nil(t, info)
|
|
require.NotNil(t, err)
|
|
require.NotEqual(t, "app.upload.upload_data.first_part_too_small.app_error", err.Id)
|
|
})
|
|
|
|
t.Run("first part too small", func(t *testing.T) {
|
|
rd := &io.LimitedReader{
|
|
R: bytes.NewReader(data),
|
|
N: 1024 * 1024,
|
|
}
|
|
|
|
ok, err := th.App.FileExists(us.Path)
|
|
require.False(t, ok)
|
|
require.Nil(t, err)
|
|
|
|
info, err := th.App.UploadData(us, rd)
|
|
require.Nil(t, info)
|
|
require.NotNil(t, err)
|
|
require.Equal(t, "app.upload.upload_data.first_part_too_small.app_error", err.Id)
|
|
|
|
ok, err = th.App.FileExists(us.Path)
|
|
require.False(t, ok)
|
|
require.Nil(t, err)
|
|
})
|
|
|
|
t.Run("resume success", func(t *testing.T) {
|
|
rd := &io.LimitedReader{
|
|
R: bytes.NewReader(data),
|
|
N: 5 * 1024 * 1024,
|
|
}
|
|
info, err := th.App.UploadData(us, rd)
|
|
require.Nil(t, info)
|
|
require.Nil(t, err)
|
|
|
|
rd = &io.LimitedReader{
|
|
R: bytes.NewReader(data[5*1024*1024:]),
|
|
N: 3 * 1024 * 1024,
|
|
}
|
|
info, err = th.App.UploadData(us, rd)
|
|
require.Nil(t, err)
|
|
require.NotEmpty(t, info)
|
|
|
|
d, err := th.App.ReadFile(us.Path)
|
|
require.Nil(t, err)
|
|
require.Equal(t, data, d)
|
|
})
|
|
|
|
t.Run("all at once success", func(t *testing.T) {
|
|
us.Id = model.NewId()
|
|
us, err = th.App.CreateUploadSession(us)
|
|
require.Nil(t, err)
|
|
require.NotEmpty(t, us)
|
|
|
|
info, err := th.App.UploadData(us, bytes.NewReader(data))
|
|
require.Nil(t, err)
|
|
require.NotEmpty(t, info)
|
|
|
|
d, err := th.App.ReadFile(us.Path)
|
|
require.Nil(t, err)
|
|
require.Equal(t, data, d)
|
|
})
|
|
|
|
t.Run("small file success", func(t *testing.T) {
|
|
us.Id = model.NewId()
|
|
us.FileSize = 1024 * 1024
|
|
us, err = th.App.CreateUploadSession(us)
|
|
require.Nil(t, err)
|
|
require.NotEmpty(t, us)
|
|
|
|
rd := &io.LimitedReader{
|
|
R: bytes.NewReader(data),
|
|
N: 1024 * 1024,
|
|
}
|
|
info, err := th.App.UploadData(us, rd)
|
|
require.Nil(t, err)
|
|
require.NotEmpty(t, info)
|
|
|
|
d, err := th.App.ReadFile(us.Path)
|
|
require.Nil(t, err)
|
|
require.Equal(t, data[:1024*1024], d)
|
|
})
|
|
|
|
t.Run("image processing", func(t *testing.T) {
|
|
testDir, _ := fileutils.FindDir("tests")
|
|
data, err := ioutil.ReadFile(filepath.Join(testDir, "test.png"))
|
|
require.Nil(t, err)
|
|
require.NotEmpty(t, data)
|
|
|
|
us.Id = model.NewId()
|
|
us.Filename = "test.png"
|
|
us.FileSize = int64(len(data))
|
|
us, err = th.App.CreateUploadSession(us)
|
|
require.Nil(t, err)
|
|
require.NotEmpty(t, us)
|
|
|
|
info, err := th.App.UploadData(us, bytes.NewReader(data))
|
|
require.Nil(t, err)
|
|
require.NotEmpty(t, info)
|
|
require.NotZero(t, info.Width)
|
|
require.NotZero(t, info.Height)
|
|
require.NotEmpty(t, info.ThumbnailPath)
|
|
require.NotEmpty(t, info.PreviewPath)
|
|
})
|
|
}
|
|
|
|
func TestUploadDataConcurrent(t *testing.T) {
|
|
th := Setup(t).InitBasic()
|
|
defer th.TearDown()
|
|
|
|
us := &model.UploadSession{
|
|
Id: model.NewId(),
|
|
Type: model.UploadTypeAttachment,
|
|
UserId: th.BasicUser.Id,
|
|
ChannelId: th.BasicChannel.Id,
|
|
Filename: "upload",
|
|
FileSize: 8 * 1024 * 1024,
|
|
}
|
|
|
|
var err error
|
|
us, err = th.App.CreateUploadSession(us)
|
|
require.Nil(t, err)
|
|
require.NotEmpty(t, us)
|
|
|
|
data := make([]byte, us.FileSize)
|
|
_, err2 := rand.Read(data)
|
|
require.NoError(t, err2)
|
|
|
|
var nErrs int32
|
|
var wg sync.WaitGroup
|
|
n := 8
|
|
wg.Add(n)
|
|
|
|
for i := 0; i < n; i++ {
|
|
go func() {
|
|
defer wg.Done()
|
|
rd := &io.LimitedReader{
|
|
R: bytes.NewReader(data),
|
|
N: 5 * 1024 * 1024,
|
|
}
|
|
u := *us
|
|
_, err := th.App.UploadData(&u, rd)
|
|
if err != nil && err.Id == "app.upload.upload_data.concurrent.app_error" {
|
|
atomic.AddInt32(&nErrs, 1)
|
|
}
|
|
}()
|
|
}
|
|
|
|
wg.Wait()
|
|
|
|
// Verify that only 1 request was able to perform the upload.
|
|
require.Equal(t, int32(n-1), nErrs)
|
|
|
|
wg.Add(n)
|
|
|
|
for i := 0; i < n; i++ {
|
|
go func() {
|
|
defer wg.Done()
|
|
rd := &io.LimitedReader{
|
|
R: bytes.NewReader(data[5*1024*1024:]),
|
|
N: 3 * 1024 * 1024,
|
|
}
|
|
u := *us
|
|
u.FileOffset = 5 * 1024 * 1024
|
|
_, err := th.App.UploadData(&u, rd)
|
|
if err != nil && err.Id == "app.upload.upload_data.concurrent.app_error" {
|
|
atomic.AddInt32(&nErrs, 1)
|
|
}
|
|
}()
|
|
}
|
|
|
|
wg.Wait()
|
|
|
|
// Verify that only 1 request was able to finish the upload.
|
|
require.Equal(t, int32(n*2-2), nErrs)
|
|
|
|
d, err := th.App.ReadFile(us.Path)
|
|
require.Nil(t, err)
|
|
require.Equal(t, data, d)
|
|
}
|