mirror of https://github.com/Xhofe/alist
254 lines
6.3 KiB
Go
254 lines
6.3 KiB
Go
package stream
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
"os"
|
|
|
|
"github.com/alist-org/alist/v3/internal/errs"
|
|
"github.com/alist-org/alist/v3/internal/model"
|
|
"github.com/alist-org/alist/v3/pkg/http_range"
|
|
"github.com/alist-org/alist/v3/pkg/utils"
|
|
)
|
|
|
|
type FileStream struct {
|
|
Ctx context.Context
|
|
model.Obj
|
|
io.Reader
|
|
Mimetype string
|
|
WebPutAsTask bool
|
|
ForceStreamUpload bool
|
|
Exist model.Obj //the file existed in the destination, we can reuse some info since we wil overwrite it
|
|
utils.Closers
|
|
tmpFile *os.File //if present, tmpFile has full content, it will be deleted at last
|
|
peekBuff *bytes.Reader
|
|
}
|
|
|
|
func (f *FileStream) GetSize() int64 {
|
|
if f.tmpFile != nil {
|
|
info, err := f.tmpFile.Stat()
|
|
if err == nil {
|
|
return info.Size()
|
|
}
|
|
}
|
|
return f.Obj.GetSize()
|
|
}
|
|
|
|
func (f *FileStream) GetMimetype() string {
|
|
return f.Mimetype
|
|
}
|
|
|
|
func (f *FileStream) NeedStore() bool {
|
|
return f.WebPutAsTask
|
|
}
|
|
|
|
func (f *FileStream) IsForceStreamUpload() bool {
|
|
return f.ForceStreamUpload
|
|
}
|
|
|
|
func (f *FileStream) Close() error {
|
|
var err1, err2 error
|
|
|
|
err1 = f.Closers.Close()
|
|
if errors.Is(err1, os.ErrClosed) {
|
|
err1 = nil
|
|
}
|
|
if f.tmpFile != nil {
|
|
err2 = os.RemoveAll(f.tmpFile.Name())
|
|
if err2 != nil {
|
|
err2 = errs.NewErr(err2, "failed to remove tmpFile [%s]", f.tmpFile.Name())
|
|
}
|
|
}
|
|
|
|
return errors.Join(err1, err2)
|
|
}
|
|
|
|
func (f *FileStream) GetExist() model.Obj {
|
|
return f.Exist
|
|
}
|
|
func (f *FileStream) SetExist(obj model.Obj) {
|
|
f.Exist = obj
|
|
}
|
|
|
|
// CacheFullInTempFile save all data into tmpFile. Not recommended since it wears disk,
|
|
// and can't start upload until the file is written. It's not thread-safe!
|
|
func (f *FileStream) CacheFullInTempFile() (model.File, error) {
|
|
if f.tmpFile != nil {
|
|
return f.tmpFile, nil
|
|
}
|
|
if file, ok := f.Reader.(model.File); ok {
|
|
return file, nil
|
|
}
|
|
tmpF, err := utils.CreateTempFile(f.Reader, f.GetSize())
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
f.Add(tmpF)
|
|
f.tmpFile = tmpF
|
|
f.Reader = tmpF
|
|
return f.tmpFile, nil
|
|
}
|
|
|
|
const InMemoryBufMaxSize = 10 // Megabytes
|
|
const InMemoryBufMaxSizeBytes = InMemoryBufMaxSize * 1024 * 1024
|
|
|
|
// RangeRead have to cache all data first since only Reader is provided.
|
|
// also support a peeking RangeRead at very start, but won't buffer more than 10MB data in memory
|
|
func (f *FileStream) RangeRead(httpRange http_range.Range) (io.Reader, error) {
|
|
if httpRange.Length == -1 {
|
|
httpRange.Length = f.GetSize()
|
|
}
|
|
if f.peekBuff != nil && httpRange.Start < int64(f.peekBuff.Len()) && httpRange.Start+httpRange.Length-1 < int64(f.peekBuff.Len()) {
|
|
return io.NewSectionReader(f.peekBuff, httpRange.Start, httpRange.Length), nil
|
|
}
|
|
if f.tmpFile == nil {
|
|
if httpRange.Start == 0 && httpRange.Length <= InMemoryBufMaxSizeBytes && f.peekBuff == nil {
|
|
bufSize := utils.Min(httpRange.Length, f.GetSize())
|
|
newBuf := bytes.NewBuffer(make([]byte, 0, bufSize))
|
|
n, err := utils.CopyWithBufferN(newBuf, f.Reader, bufSize)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if n != bufSize {
|
|
return nil, fmt.Errorf("stream RangeRead did not get all data in peek, expect =%d ,actual =%d", bufSize, n)
|
|
}
|
|
f.peekBuff = bytes.NewReader(newBuf.Bytes())
|
|
f.Reader = io.MultiReader(f.peekBuff, f.Reader)
|
|
return io.NewSectionReader(f.peekBuff, httpRange.Start, httpRange.Length), nil
|
|
} else {
|
|
_, err := f.CacheFullInTempFile()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
}
|
|
return io.NewSectionReader(f.tmpFile, httpRange.Start, httpRange.Length), nil
|
|
}
|
|
|
|
var _ model.FileStreamer = (*SeekableStream)(nil)
|
|
var _ model.FileStreamer = (*FileStream)(nil)
|
|
|
|
//var _ seekableStream = (*FileStream)(nil)
|
|
|
|
// for most internal stream, which is either RangeReadCloser or MFile
|
|
type SeekableStream struct {
|
|
FileStream
|
|
Link *model.Link
|
|
// should have one of belows to support rangeRead
|
|
rangeReadCloser model.RangeReadCloserIF
|
|
mFile model.File
|
|
}
|
|
|
|
func NewSeekableStream(fs FileStream, link *model.Link) (*SeekableStream, error) {
|
|
if len(fs.Mimetype) == 0 {
|
|
fs.Mimetype = utils.GetMimeType(fs.Obj.GetName())
|
|
}
|
|
ss := SeekableStream{FileStream: fs, Link: link}
|
|
if ss.Reader != nil {
|
|
result, ok := ss.Reader.(model.File)
|
|
if ok {
|
|
ss.mFile = result
|
|
ss.Closers.Add(result)
|
|
return &ss, nil
|
|
}
|
|
}
|
|
if ss.Link != nil {
|
|
if ss.Link.MFile != nil {
|
|
ss.mFile = ss.Link.MFile
|
|
ss.Reader = ss.Link.MFile
|
|
ss.Closers.Add(ss.Link.MFile)
|
|
return &ss, nil
|
|
}
|
|
|
|
if ss.Link.RangeReadCloser != nil {
|
|
ss.rangeReadCloser = ss.Link.RangeReadCloser
|
|
return &ss, nil
|
|
}
|
|
if len(ss.Link.URL) > 0 {
|
|
rrc, err := GetRangeReadCloserFromLink(ss.GetSize(), link)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
ss.rangeReadCloser = rrc
|
|
return &ss, nil
|
|
}
|
|
}
|
|
|
|
return nil, fmt.Errorf("illegal seekableStream")
|
|
}
|
|
|
|
//func (ss *SeekableStream) Peek(length int) {
|
|
//
|
|
//}
|
|
|
|
// RangeRead is not thread-safe, pls use it in single thread only.
|
|
func (ss *SeekableStream) RangeRead(httpRange http_range.Range) (io.Reader, error) {
|
|
if httpRange.Length == -1 {
|
|
httpRange.Length = ss.GetSize()
|
|
}
|
|
if ss.mFile != nil {
|
|
return io.NewSectionReader(ss.mFile, httpRange.Start, httpRange.Length), nil
|
|
}
|
|
if ss.tmpFile != nil {
|
|
return io.NewSectionReader(ss.tmpFile, httpRange.Start, httpRange.Length), nil
|
|
}
|
|
if ss.rangeReadCloser != nil {
|
|
rc, err := ss.rangeReadCloser.RangeRead(ss.Ctx, httpRange)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return rc, nil
|
|
}
|
|
return nil, fmt.Errorf("can't find mFile or rangeReadCloser")
|
|
}
|
|
|
|
//func (f *FileStream) GetReader() io.Reader {
|
|
// return f.Reader
|
|
//}
|
|
|
|
// only provide Reader as full stream when it's demanded. in rapid-upload, we can skip this to save memory
|
|
func (ss *SeekableStream) Read(p []byte) (n int, err error) {
|
|
//f.mu.Lock()
|
|
|
|
//f.peekedOnce = true
|
|
//defer f.mu.Unlock()
|
|
if ss.Reader == nil {
|
|
if ss.rangeReadCloser == nil {
|
|
return 0, fmt.Errorf("illegal seekableStream")
|
|
}
|
|
rc, err := ss.rangeReadCloser.RangeRead(ss.Ctx, http_range.Range{Length: -1})
|
|
if err != nil {
|
|
return 0, nil
|
|
}
|
|
ss.Reader = io.NopCloser(rc)
|
|
ss.Closers.Add(rc)
|
|
|
|
}
|
|
return ss.Reader.Read(p)
|
|
}
|
|
|
|
func (ss *SeekableStream) CacheFullInTempFile() (model.File, error) {
|
|
if ss.tmpFile != nil {
|
|
return ss.tmpFile, nil
|
|
}
|
|
if ss.mFile != nil {
|
|
return ss.mFile, nil
|
|
}
|
|
tmpF, err := utils.CreateTempFile(ss, ss.GetSize())
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
ss.Add(tmpF)
|
|
ss.tmpFile = tmpF
|
|
ss.Reader = tmpF
|
|
return ss.tmpFile, nil
|
|
}
|
|
|
|
func (f *FileStream) SetTmpFile(r *os.File) {
|
|
f.Reader = r
|
|
f.tmpFile = r
|
|
}
|