bfs:对FileHeader的压缩和解压使用Pool管理

This commit is contained in:
GoEdgeLab
2024-04-28 10:06:29 +08:00
parent 33c8e4b201
commit b3aa8dfeed
8 changed files with 470 additions and 65 deletions

View File

@@ -3,6 +3,8 @@
package bfs
import (
"encoding/json"
"github.com/TeaOSLab/EdgeNode/internal/utils"
"sort"
)
@@ -20,6 +22,35 @@ type FileHeader struct {
IsWriting bool `json:"11,omitempty"`
}
func (this *FileHeader) BlockAt(offset int64) (blockInfo BlockInfo, ok bool) {
var l = len(this.BodyBlocks)
if l == 1 {
if this.BodyBlocks[0].Contains(offset) {
return this.BodyBlocks[0], true
}
return
}
sort.Search(l, func(i int) bool {
if this.BodyBlocks[i].Contains(offset) {
blockInfo = this.BodyBlocks[i]
ok = true
return true
}
return this.BodyBlocks[i].OriginOffsetFrom > offset
})
return
}
func (this *FileHeader) MaxOffset() int64 {
var l = len(this.BodyBlocks)
if l > 0 {
return this.BodyBlocks[l-1].OriginOffsetTo
}
return 0
}
func (this *FileHeader) Compact() {
// TODO 合并相邻的headerBlocks和bodyBlocks必须是对应的BFile offset也要相邻
@@ -65,31 +96,37 @@ func (this *FileHeader) Clone() *FileHeader {
}
}
func (this *FileHeader) BlockAt(offset int64) (blockInfo BlockInfo, ok bool) {
var l = len(this.BodyBlocks)
if l == 1 {
if this.BodyBlocks[0].Contains(offset) {
return this.BodyBlocks[0], true
}
return
func (this *FileHeader) Encode(hash string) ([]byte, error) {
headerJSON, err := json.Marshal(this)
if err != nil {
return nil, err
}
sort.Search(l, func(i int) bool {
if this.BodyBlocks[i].Contains(offset) {
blockInfo = this.BodyBlocks[i]
ok = true
return true
}
return this.BodyBlocks[i].OriginOffsetFrom > offset
})
return
}
func (this *FileHeader) MaxOffset() int64 {
var l = len(this.BodyBlocks)
if l > 0 {
return this.BodyBlocks[l-1].OriginOffsetTo
// we do not compress data which size is less than 100 bytes
if len(headerJSON) < 100 {
return EncodeMetaBlock(MetaActionNew, hash, append([]byte("json:"), headerJSON...))
}
return 0
var buf = utils.SharedBufferPool.Get()
defer utils.SharedBufferPool.Put(buf)
compressor, err := SharedCompressPool.Get(buf)
if err != nil {
return nil, err
}
_, err = compressor.Write(headerJSON)
if err != nil {
_ = compressor.Close()
SharedCompressPool.Put(compressor)
return nil, err
}
err = compressor.Close()
SharedCompressPool.Put(compressor)
if err != nil {
return nil, err
}
return EncodeMetaBlock(MetaActionNew, hash, buf.Bytes())
}

View File

@@ -5,7 +5,6 @@ package bfs
import (
"bytes"
"encoding/json"
"github.com/klauspost/compress/gzip"
)
// LazyFileHeader load file header lazily to save memory
@@ -31,18 +30,30 @@ func (this *LazyFileHeader) FileHeaderUnsafe() (*FileHeader, error) {
return this.fileHeader, nil
}
// TODO 使用pool管理gzip
gzReader, err := gzip.NewReader(bytes.NewBuffer(this.rawData))
var jsonPrefix = []byte("json:")
var header = &FileHeader{}
// json
if bytes.HasPrefix(this.rawData, jsonPrefix) {
err := json.Unmarshal(this.rawData[len(jsonPrefix):], header)
if err != nil {
return nil, err
}
return header, nil
}
decompressor, err := SharedDecompressPool.Get(bytes.NewBuffer(this.rawData))
if err != nil {
return nil, err
}
defer func() {
_ = gzReader.Close()
_ = decompressor.Close()
SharedDecompressPool.Put(decompressor)
}()
var header = &FileHeader{}
err = json.NewDecoder(gzReader).Decode(header)
err = json.NewDecoder(decompressor).Decode(header)
if err != nil {
return nil, err
}

View File

@@ -0,0 +1,87 @@
// Copyright 2024 GoEdge CDN goedge.cdn@gmail.com. All rights reserved. Official site: https://goedge.cn .
package bfs_test
import (
"github.com/TeaOSLab/EdgeNode/internal/utils/bfs"
"runtime"
"testing"
)
func TestNewLazyFileHeaderFromData(t *testing.T) {
var header = &bfs.FileHeader{
Version: 1,
Status: 200,
BodyBlocks: []bfs.BlockInfo{
{
BFileOffsetFrom: 0,
BFileOffsetTo: 1 << 20,
},
},
}
blockBytes, err := header.Encode(bfs.Hash("123456"))
if err != nil {
t.Fatal(err)
}
_, _, rawData, err := bfs.DecodeMetaBlock(blockBytes)
if err != nil {
t.Fatal(err)
}
var lazyHeader = bfs.NewLazyFileHeaderFromData(rawData)
newHeader, err := lazyHeader.FileHeaderUnsafe()
if err != nil {
t.Fatal(err)
}
t.Log(newHeader)
}
func BenchmarkLazyFileHeader_Decode(b *testing.B) {
runtime.GOMAXPROCS(12)
var header = &bfs.FileHeader{
Version: 1,
Status: 200,
BodyBlocks: []bfs.BlockInfo{},
}
var offset int64
for {
var end = offset + 16<<10
if end > 1<<20 {
break
}
header.BodyBlocks = append(header.BodyBlocks, bfs.BlockInfo{
BFileOffsetFrom: offset,
BFileOffsetTo: end,
})
offset = end
}
var hash = bfs.Hash("123456")
blockBytes, err := header.Encode(hash)
if err != nil {
b.Fatal(err)
}
b.ResetTimer()
b.RunParallel(func(pb *testing.PB) {
for pb.Next() {
_, _, rawData, decodeErr := bfs.DecodeMetaBlock(blockBytes)
if decodeErr != nil {
b.Fatal(decodeErr)
}
var lazyHeader = bfs.NewLazyFileHeaderFromData(rawData)
_, decodeErr = lazyHeader.FileHeaderUnsafe()
if decodeErr != nil {
b.Fatal(decodeErr)
}
}
})
}

View File

@@ -3,9 +3,13 @@
package bfs_test
import (
"encoding/json"
"github.com/TeaOSLab/EdgeNode/internal/utils/bfs"
"github.com/TeaOSLab/EdgeNode/internal/utils/fasttime"
"github.com/iwind/TeaGo/assert"
"github.com/iwind/TeaGo/logs"
"math/rand"
"runtime"
"testing"
)
@@ -176,6 +180,120 @@ func TestFileHeader_Clone(t *testing.T) {
a.IsTrue(header.BodyBlocks[0].OriginOffsetFrom != clonedHeader.BodyBlocks[0].OriginOffsetFrom)
}
func TestFileHeader_Encode(t *testing.T) {
{
var header = &bfs.FileHeader{
Version: 1,
Status: 200,
ModifiedAt: fasttime.Now().Unix(),
ExpiresAt: fasttime.Now().Unix() + 3600,
BodySize: 1 << 20,
HeaderSize: 1 << 10,
BodyBlocks: []bfs.BlockInfo{
{
BFileOffsetFrom: 1 << 10,
BFileOffsetTo: 1 << 20,
},
},
}
data, err := header.Encode(bfs.Hash("123456"))
if err != nil {
t.Fatal(err)
}
jsonBytes, _ := json.Marshal(header)
t.Log(len(header.BodyBlocks), "blocks", len(data), "bytes", "json:", len(jsonBytes), "bytes")
_, _, _, err = bfs.DecodeMetaBlock(data)
if err != nil {
t.Fatal(err)
}
}
{
var header = &bfs.FileHeader{
Version: 1,
Status: 200,
BodyBlocks: []bfs.BlockInfo{},
}
var offset int64
for {
var end = offset + 16<<10
if end > 256<<10 {
break
}
header.BodyBlocks = append(header.BodyBlocks, bfs.BlockInfo{
BFileOffsetFrom: offset,
BFileOffsetTo: end,
})
offset = end
}
data, err := header.Encode(bfs.Hash("123456"))
if err != nil {
t.Fatal(err)
}
jsonBytes, _ := json.Marshal(header)
t.Log(len(header.BodyBlocks), "blocks", len(data), "bytes", "json:", len(jsonBytes), "bytes")
}
{
var header = &bfs.FileHeader{
Version: 1,
Status: 200,
BodyBlocks: []bfs.BlockInfo{},
}
var offset int64
for {
var end = offset + 16<<10
if end > 512<<10 {
break
}
header.BodyBlocks = append(header.BodyBlocks, bfs.BlockInfo{
BFileOffsetFrom: offset,
BFileOffsetTo: end,
})
offset = end
}
data, err := header.Encode(bfs.Hash("123456"))
if err != nil {
t.Fatal(err)
}
jsonBytes, _ := json.Marshal(header)
t.Log(len(header.BodyBlocks), "blocks", len(data), "bytes", "json:", len(jsonBytes), "bytes")
}
{
var header = &bfs.FileHeader{
Version: 1,
Status: 200,
BodyBlocks: []bfs.BlockInfo{},
}
var offset int64
for {
var end = offset + 16<<10
if end > 1<<20 {
break
}
header.BodyBlocks = append(header.BodyBlocks, bfs.BlockInfo{
BFileOffsetFrom: offset,
BFileOffsetTo: end,
})
offset = end
}
data, err := header.Encode(bfs.Hash("123456"))
if err != nil {
t.Fatal(err)
}
jsonBytes, _ := json.Marshal(header)
t.Log(len(header.BodyBlocks), "blocks", len(data), "bytes", "json:", len(jsonBytes), "bytes")
}
}
func BenchmarkFileHeader_Compact(b *testing.B) {
for i := 0; i < b.N; i++ {
var header = &bfs.FileHeader{
@@ -197,3 +315,40 @@ func BenchmarkFileHeader_Compact(b *testing.B) {
header.Compact()
}
}
func BenchmarkFileHeader_Encode(b *testing.B) {
runtime.GOMAXPROCS(12)
b.RunParallel(func(pb *testing.PB) {
for pb.Next() {
var header = &bfs.FileHeader{
Version: 1,
Status: 200,
ModifiedAt: rand.Int63(),
BodySize: rand.Int63(),
BodyBlocks: []bfs.BlockInfo{},
}
var offset int64
for {
var end = offset + 16<<10
if end > 2<<20 {
break
}
header.BodyBlocks = append(header.BodyBlocks, bfs.BlockInfo{
BFileOffsetFrom: offset + int64(rand.Int()%1000000),
BFileOffsetTo: end + int64(rand.Int()%1000000),
})
offset = end
}
var hash = bfs.Hash("123456")
_, err := header.Encode(hash)
if err != nil {
b.Fatal(err)
}
}
})
}

View File

@@ -0,0 +1,66 @@
// Copyright 2024 GoEdge CDN goedge.cdn@gmail.com. All rights reserved. Official site: https://goedge.cn .
package bfs
import (
"github.com/TeaOSLab/EdgeNode/internal/utils/percpu"
"github.com/klauspost/compress/gzip"
"io"
"runtime"
)
var SharedDecompressPool = NewGzipReaderPool()
type GzipReaderPool struct {
c chan *gzip.Reader
cList []chan *gzip.Reader
}
func NewGzipReaderPool() *GzipReaderPool {
const poolSize = 16
var countProcs = runtime.GOMAXPROCS(0)
if countProcs <= 0 {
countProcs = runtime.NumCPU()
}
countProcs *= 4
var cList []chan *gzip.Reader
for i := 0; i < countProcs; i++ {
cList = append(cList, make(chan *gzip.Reader, poolSize))
}
return &GzipReaderPool{
c: make(chan *gzip.Reader, poolSize),
cList: cList,
}
}
func (this *GzipReaderPool) Get(rawReader io.Reader) (*gzip.Reader, error) {
select {
case w := <-this.getC():
err := w.Reset(rawReader)
if err != nil {
return nil, err
}
return w, nil
default:
return gzip.NewReader(rawReader)
}
}
func (this *GzipReaderPool) Put(reader *gzip.Reader) {
select {
case this.getC() <- reader:
default:
// 不需要close因为已经在使用的时候调用了
}
}
func (this *GzipReaderPool) getC() chan *gzip.Reader {
var procId = percpu.GetProcId()
if procId < len(this.cList) {
return this.cList[procId]
}
return this.c
}

View File

@@ -0,0 +1,63 @@
// Copyright 2024 GoEdge CDN goedge.cdn@gmail.com. All rights reserved. Official site: https://goedge.cn .
package bfs
import (
"github.com/TeaOSLab/EdgeNode/internal/utils/percpu"
"github.com/klauspost/compress/gzip"
"io"
"runtime"
)
var SharedCompressPool = NewGzipWriterPool()
type GzipWriterPool struct {
c chan *gzip.Writer
cList []chan *gzip.Writer
}
func NewGzipWriterPool() *GzipWriterPool {
const poolSize = 16
var countProcs = runtime.GOMAXPROCS(0)
if countProcs <= 0 {
countProcs = runtime.NumCPU()
}
countProcs *= 4
var cList []chan *gzip.Writer
for i := 0; i < countProcs; i++ {
cList = append(cList, make(chan *gzip.Writer, poolSize))
}
return &GzipWriterPool{
c: make(chan *gzip.Writer, poolSize),
cList: cList,
}
}
func (this *GzipWriterPool) Get(rawWriter io.Writer) (*gzip.Writer, error) {
select {
case w := <-this.getC():
w.Reset(rawWriter)
return w, nil
default:
return gzip.NewWriterLevel(rawWriter, gzip.BestSpeed)
}
}
func (this *GzipWriterPool) Put(writer *gzip.Writer) {
select {
case this.getC() <- writer:
default:
// 不需要close因为已经在使用的时候调用了
}
}
func (this *GzipWriterPool) getC() chan *gzip.Writer {
var procId = percpu.GetProcId()
if procId < len(this.cList) {
return this.cList[procId]
}
return this.c
}

View File

@@ -5,11 +5,8 @@ package bfs
import (
"bytes"
"encoding/binary"
"encoding/json"
"github.com/TeaOSLab/EdgeNode/internal/utils"
"github.com/TeaOSLab/EdgeNode/internal/utils/fasttime"
"github.com/TeaOSLab/EdgeNode/internal/zero"
"github.com/klauspost/compress/gzip"
"io"
"os"
"sync"
@@ -189,7 +186,7 @@ func (this *MetaFile) WriteClose(hash string, headerSize int64, bodySize int64)
header.BodySize = bodySize
header.Compact()
blockBytes, err := this.encodeFileHeader(hash, header)
blockBytes, err := header.Encode(hash)
if err != nil {
return err
}
@@ -315,7 +312,7 @@ func (this *MetaFile) Compact() error {
return err
}
blockBytes, err := this.encodeFileHeader(hash, header)
blockBytes, err := header.Encode(hash)
if err != nil {
return err
}
@@ -381,33 +378,3 @@ func (this *MetaFile) RemoveAll() error {
_ = this.fp.Close()
return os.Remove(this.fp.Name())
}
// encode file header to data bytes
func (this *MetaFile) encodeFileHeader(hash string, header *FileHeader) ([]byte, error) {
headerJSON, err := json.Marshal(header)
if err != nil {
return nil, err
}
var buf = utils.SharedBufferPool.Get()
defer utils.SharedBufferPool.Put(buf)
// TODO 考虑使用gzip pool
gzWriter, err := gzip.NewWriterLevel(buf, gzip.BestSpeed)
if err != nil {
return nil, err
}
_, err = gzWriter.Write(headerJSON)
if err != nil {
_ = gzWriter.Close()
return nil, err
}
err = gzWriter.Close()
if err != nil {
return nil, err
}
return EncodeMetaBlock(MetaActionNew, hash, buf.Bytes())
}

View File

@@ -0,0 +1,19 @@
// Copyright 2024 GoEdge CDN goedge.cdn@gmail.com. All rights reserved. Official site: https://goedge.cn .
package percpu
import (
_ "unsafe"
)
//go:linkname runtime_procPin runtime.procPin
func runtime_procPin() int
//go:linkname runtime_procUnpin runtime.procUnpin
func runtime_procUnpin() int
func GetProcId() int {
var pid = runtime_procPin()
runtime_procUnpin()
return pid
}