mirror of
				https://gitee.com/gitea/gitea
				synced 2025-11-04 08:30:25 +08:00 
			
		
		
		
	* Vendor: update gitea.com/macaron/session to a177a270 * make vendor * Vendor: update gitea.com/macaron/macaron to 0db5d458 * make vendor * Vendor: update gitea.com/macaron/cache to 905232fb * make vendor * Vendor: update gitea.com/macaron/i18n to 4ca3dd0c * make vendor * Vendor: update gitea.com/macaron/gzip to efa5e847 * make vendor * Vendor: update gitea.com/macaron/captcha to e8597820 * make vendor
		
			
				
	
	
		
			510 lines
		
	
	
		
			9.0 KiB
		
	
	
	
		
			Go
		
	
	
	
		
			Vendored
		
	
	
	
			
		
		
	
	
			510 lines
		
	
	
		
			9.0 KiB
		
	
	
	
		
			Go
		
	
	
	
		
			Vendored
		
	
	
	
package nodb
 | 
						|
 | 
						|
import (
 | 
						|
	"encoding/binary"
 | 
						|
	"errors"
 | 
						|
	"time"
 | 
						|
 | 
						|
	"gitea.com/lunny/nodb/store"
 | 
						|
)
 | 
						|
 | 
						|
type FVPair struct {
 | 
						|
	Field []byte
 | 
						|
	Value []byte
 | 
						|
}
 | 
						|
 | 
						|
var errHashKey = errors.New("invalid hash key")
 | 
						|
var errHSizeKey = errors.New("invalid hsize key")
 | 
						|
 | 
						|
const (
 | 
						|
	hashStartSep byte = ':'
 | 
						|
	hashStopSep  byte = hashStartSep + 1
 | 
						|
)
 | 
						|
 | 
						|
func checkHashKFSize(key []byte, field []byte) error {
 | 
						|
	if len(key) > MaxKeySize || len(key) == 0 {
 | 
						|
		return errKeySize
 | 
						|
	} else if len(field) > MaxHashFieldSize || len(field) == 0 {
 | 
						|
		return errHashFieldSize
 | 
						|
	}
 | 
						|
	return nil
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) hEncodeSizeKey(key []byte) []byte {
 | 
						|
	buf := make([]byte, len(key)+2)
 | 
						|
 | 
						|
	buf[0] = db.index
 | 
						|
	buf[1] = HSizeType
 | 
						|
 | 
						|
	copy(buf[2:], key)
 | 
						|
	return buf
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) hDecodeSizeKey(ek []byte) ([]byte, error) {
 | 
						|
	if len(ek) < 2 || ek[0] != db.index || ek[1] != HSizeType {
 | 
						|
		return nil, errHSizeKey
 | 
						|
	}
 | 
						|
 | 
						|
	return ek[2:], nil
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) hEncodeHashKey(key []byte, field []byte) []byte {
 | 
						|
	buf := make([]byte, len(key)+len(field)+1+1+2+1)
 | 
						|
 | 
						|
	pos := 0
 | 
						|
	buf[pos] = db.index
 | 
						|
	pos++
 | 
						|
	buf[pos] = HashType
 | 
						|
	pos++
 | 
						|
 | 
						|
	binary.BigEndian.PutUint16(buf[pos:], uint16(len(key)))
 | 
						|
	pos += 2
 | 
						|
 | 
						|
	copy(buf[pos:], key)
 | 
						|
	pos += len(key)
 | 
						|
 | 
						|
	buf[pos] = hashStartSep
 | 
						|
	pos++
 | 
						|
	copy(buf[pos:], field)
 | 
						|
 | 
						|
	return buf
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) hDecodeHashKey(ek []byte) ([]byte, []byte, error) {
 | 
						|
	if len(ek) < 5 || ek[0] != db.index || ek[1] != HashType {
 | 
						|
		return nil, nil, errHashKey
 | 
						|
	}
 | 
						|
 | 
						|
	pos := 2
 | 
						|
	keyLen := int(binary.BigEndian.Uint16(ek[pos:]))
 | 
						|
	pos += 2
 | 
						|
 | 
						|
	if keyLen+5 > len(ek) {
 | 
						|
		return nil, nil, errHashKey
 | 
						|
	}
 | 
						|
 | 
						|
	key := ek[pos : pos+keyLen]
 | 
						|
	pos += keyLen
 | 
						|
 | 
						|
	if ek[pos] != hashStartSep {
 | 
						|
		return nil, nil, errHashKey
 | 
						|
	}
 | 
						|
 | 
						|
	pos++
 | 
						|
	field := ek[pos:]
 | 
						|
	return key, field, nil
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) hEncodeStartKey(key []byte) []byte {
 | 
						|
	return db.hEncodeHashKey(key, nil)
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) hEncodeStopKey(key []byte) []byte {
 | 
						|
	k := db.hEncodeHashKey(key, nil)
 | 
						|
 | 
						|
	k[len(k)-1] = hashStopSep
 | 
						|
 | 
						|
	return k
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) hSetItem(key []byte, field []byte, value []byte) (int64, error) {
 | 
						|
	t := db.hashBatch
 | 
						|
 | 
						|
	ek := db.hEncodeHashKey(key, field)
 | 
						|
 | 
						|
	var n int64 = 1
 | 
						|
	if v, _ := db.bucket.Get(ek); v != nil {
 | 
						|
		n = 0
 | 
						|
	} else {
 | 
						|
		if _, err := db.hIncrSize(key, 1); err != nil {
 | 
						|
			return 0, err
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	t.Put(ek, value)
 | 
						|
	return n, nil
 | 
						|
}
 | 
						|
 | 
						|
//	ps : here just focus on deleting the hash data,
 | 
						|
//		 any other likes expire is ignore.
 | 
						|
func (db *DB) hDelete(t *batch, key []byte) int64 {
 | 
						|
	sk := db.hEncodeSizeKey(key)
 | 
						|
	start := db.hEncodeStartKey(key)
 | 
						|
	stop := db.hEncodeStopKey(key)
 | 
						|
 | 
						|
	var num int64 = 0
 | 
						|
	it := db.bucket.RangeLimitIterator(start, stop, store.RangeROpen, 0, -1)
 | 
						|
	for ; it.Valid(); it.Next() {
 | 
						|
		t.Delete(it.Key())
 | 
						|
		num++
 | 
						|
	}
 | 
						|
	it.Close()
 | 
						|
 | 
						|
	t.Delete(sk)
 | 
						|
	return num
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) hExpireAt(key []byte, when int64) (int64, error) {
 | 
						|
	t := db.hashBatch
 | 
						|
	t.Lock()
 | 
						|
	defer t.Unlock()
 | 
						|
 | 
						|
	if hlen, err := db.HLen(key); err != nil || hlen == 0 {
 | 
						|
		return 0, err
 | 
						|
	} else {
 | 
						|
		db.expireAt(t, HashType, key, when)
 | 
						|
		if err := t.Commit(); err != nil {
 | 
						|
			return 0, err
 | 
						|
		}
 | 
						|
	}
 | 
						|
	return 1, nil
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HLen(key []byte) (int64, error) {
 | 
						|
	if err := checkKeySize(key); err != nil {
 | 
						|
		return 0, err
 | 
						|
	}
 | 
						|
 | 
						|
	return Int64(db.bucket.Get(db.hEncodeSizeKey(key)))
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HSet(key []byte, field []byte, value []byte) (int64, error) {
 | 
						|
	if err := checkHashKFSize(key, field); err != nil {
 | 
						|
		return 0, err
 | 
						|
	} else if err := checkValueSize(value); err != nil {
 | 
						|
		return 0, err
 | 
						|
	}
 | 
						|
 | 
						|
	t := db.hashBatch
 | 
						|
	t.Lock()
 | 
						|
	defer t.Unlock()
 | 
						|
 | 
						|
	n, err := db.hSetItem(key, field, value)
 | 
						|
	if err != nil {
 | 
						|
		return 0, err
 | 
						|
	}
 | 
						|
 | 
						|
	//todo add binlog
 | 
						|
 | 
						|
	err = t.Commit()
 | 
						|
	return n, err
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HGet(key []byte, field []byte) ([]byte, error) {
 | 
						|
	if err := checkHashKFSize(key, field); err != nil {
 | 
						|
		return nil, err
 | 
						|
	}
 | 
						|
 | 
						|
	return db.bucket.Get(db.hEncodeHashKey(key, field))
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HMset(key []byte, args ...FVPair) error {
 | 
						|
	t := db.hashBatch
 | 
						|
	t.Lock()
 | 
						|
	defer t.Unlock()
 | 
						|
 | 
						|
	var err error
 | 
						|
	var ek []byte
 | 
						|
	var num int64 = 0
 | 
						|
	for i := 0; i < len(args); i++ {
 | 
						|
		if err := checkHashKFSize(key, args[i].Field); err != nil {
 | 
						|
			return err
 | 
						|
		} else if err := checkValueSize(args[i].Value); err != nil {
 | 
						|
			return err
 | 
						|
		}
 | 
						|
 | 
						|
		ek = db.hEncodeHashKey(key, args[i].Field)
 | 
						|
 | 
						|
		if v, err := db.bucket.Get(ek); err != nil {
 | 
						|
			return err
 | 
						|
		} else if v == nil {
 | 
						|
			num++
 | 
						|
		}
 | 
						|
 | 
						|
		t.Put(ek, args[i].Value)
 | 
						|
	}
 | 
						|
 | 
						|
	if _, err = db.hIncrSize(key, num); err != nil {
 | 
						|
		return err
 | 
						|
	}
 | 
						|
 | 
						|
	//todo add binglog
 | 
						|
	err = t.Commit()
 | 
						|
	return err
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HMget(key []byte, args ...[]byte) ([][]byte, error) {
 | 
						|
	var ek []byte
 | 
						|
 | 
						|
	it := db.bucket.NewIterator()
 | 
						|
	defer it.Close()
 | 
						|
 | 
						|
	r := make([][]byte, len(args))
 | 
						|
	for i := 0; i < len(args); i++ {
 | 
						|
		if err := checkHashKFSize(key, args[i]); err != nil {
 | 
						|
			return nil, err
 | 
						|
		}
 | 
						|
 | 
						|
		ek = db.hEncodeHashKey(key, args[i])
 | 
						|
 | 
						|
		r[i] = it.Find(ek)
 | 
						|
	}
 | 
						|
 | 
						|
	return r, nil
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HDel(key []byte, args ...[]byte) (int64, error) {
 | 
						|
	t := db.hashBatch
 | 
						|
 | 
						|
	var ek []byte
 | 
						|
	var v []byte
 | 
						|
	var err error
 | 
						|
 | 
						|
	t.Lock()
 | 
						|
	defer t.Unlock()
 | 
						|
 | 
						|
	it := db.bucket.NewIterator()
 | 
						|
	defer it.Close()
 | 
						|
 | 
						|
	var num int64 = 0
 | 
						|
	for i := 0; i < len(args); i++ {
 | 
						|
		if err := checkHashKFSize(key, args[i]); err != nil {
 | 
						|
			return 0, err
 | 
						|
		}
 | 
						|
 | 
						|
		ek = db.hEncodeHashKey(key, args[i])
 | 
						|
 | 
						|
		v = it.RawFind(ek)
 | 
						|
		if v == nil {
 | 
						|
			continue
 | 
						|
		} else {
 | 
						|
			num++
 | 
						|
			t.Delete(ek)
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	if _, err = db.hIncrSize(key, -num); err != nil {
 | 
						|
		return 0, err
 | 
						|
	}
 | 
						|
 | 
						|
	err = t.Commit()
 | 
						|
 | 
						|
	return num, err
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) hIncrSize(key []byte, delta int64) (int64, error) {
 | 
						|
	t := db.hashBatch
 | 
						|
	sk := db.hEncodeSizeKey(key)
 | 
						|
 | 
						|
	var err error
 | 
						|
	var size int64 = 0
 | 
						|
	if size, err = Int64(db.bucket.Get(sk)); err != nil {
 | 
						|
		return 0, err
 | 
						|
	} else {
 | 
						|
		size += delta
 | 
						|
		if size <= 0 {
 | 
						|
			size = 0
 | 
						|
			t.Delete(sk)
 | 
						|
			db.rmExpire(t, HashType, key)
 | 
						|
		} else {
 | 
						|
			t.Put(sk, PutInt64(size))
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	return size, nil
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HIncrBy(key []byte, field []byte, delta int64) (int64, error) {
 | 
						|
	if err := checkHashKFSize(key, field); err != nil {
 | 
						|
		return 0, err
 | 
						|
	}
 | 
						|
 | 
						|
	t := db.hashBatch
 | 
						|
	var ek []byte
 | 
						|
	var err error
 | 
						|
 | 
						|
	t.Lock()
 | 
						|
	defer t.Unlock()
 | 
						|
 | 
						|
	ek = db.hEncodeHashKey(key, field)
 | 
						|
 | 
						|
	var n int64 = 0
 | 
						|
	if n, err = StrInt64(db.bucket.Get(ek)); err != nil {
 | 
						|
		return 0, err
 | 
						|
	}
 | 
						|
 | 
						|
	n += delta
 | 
						|
 | 
						|
	_, err = db.hSetItem(key, field, StrPutInt64(n))
 | 
						|
	if err != nil {
 | 
						|
		return 0, err
 | 
						|
	}
 | 
						|
 | 
						|
	err = t.Commit()
 | 
						|
 | 
						|
	return n, err
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HGetAll(key []byte) ([]FVPair, error) {
 | 
						|
	if err := checkKeySize(key); err != nil {
 | 
						|
		return nil, err
 | 
						|
	}
 | 
						|
 | 
						|
	start := db.hEncodeStartKey(key)
 | 
						|
	stop := db.hEncodeStopKey(key)
 | 
						|
 | 
						|
	v := make([]FVPair, 0, 16)
 | 
						|
 | 
						|
	it := db.bucket.RangeLimitIterator(start, stop, store.RangeROpen, 0, -1)
 | 
						|
	for ; it.Valid(); it.Next() {
 | 
						|
		_, f, err := db.hDecodeHashKey(it.Key())
 | 
						|
		if err != nil {
 | 
						|
			return nil, err
 | 
						|
		}
 | 
						|
 | 
						|
		v = append(v, FVPair{Field: f, Value: it.Value()})
 | 
						|
	}
 | 
						|
 | 
						|
	it.Close()
 | 
						|
 | 
						|
	return v, nil
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HKeys(key []byte) ([][]byte, error) {
 | 
						|
	if err := checkKeySize(key); err != nil {
 | 
						|
		return nil, err
 | 
						|
	}
 | 
						|
 | 
						|
	start := db.hEncodeStartKey(key)
 | 
						|
	stop := db.hEncodeStopKey(key)
 | 
						|
 | 
						|
	v := make([][]byte, 0, 16)
 | 
						|
 | 
						|
	it := db.bucket.RangeLimitIterator(start, stop, store.RangeROpen, 0, -1)
 | 
						|
	for ; it.Valid(); it.Next() {
 | 
						|
		_, f, err := db.hDecodeHashKey(it.Key())
 | 
						|
		if err != nil {
 | 
						|
			return nil, err
 | 
						|
		}
 | 
						|
		v = append(v, f)
 | 
						|
	}
 | 
						|
 | 
						|
	it.Close()
 | 
						|
 | 
						|
	return v, nil
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HValues(key []byte) ([][]byte, error) {
 | 
						|
	if err := checkKeySize(key); err != nil {
 | 
						|
		return nil, err
 | 
						|
	}
 | 
						|
 | 
						|
	start := db.hEncodeStartKey(key)
 | 
						|
	stop := db.hEncodeStopKey(key)
 | 
						|
 | 
						|
	v := make([][]byte, 0, 16)
 | 
						|
 | 
						|
	it := db.bucket.RangeLimitIterator(start, stop, store.RangeROpen, 0, -1)
 | 
						|
	for ; it.Valid(); it.Next() {
 | 
						|
		_, _, err := db.hDecodeHashKey(it.Key())
 | 
						|
		if err != nil {
 | 
						|
			return nil, err
 | 
						|
		}
 | 
						|
 | 
						|
		v = append(v, it.Value())
 | 
						|
	}
 | 
						|
 | 
						|
	it.Close()
 | 
						|
 | 
						|
	return v, nil
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HClear(key []byte) (int64, error) {
 | 
						|
	if err := checkKeySize(key); err != nil {
 | 
						|
		return 0, err
 | 
						|
	}
 | 
						|
 | 
						|
	t := db.hashBatch
 | 
						|
	t.Lock()
 | 
						|
	defer t.Unlock()
 | 
						|
 | 
						|
	num := db.hDelete(t, key)
 | 
						|
	db.rmExpire(t, HashType, key)
 | 
						|
 | 
						|
	err := t.Commit()
 | 
						|
	return num, err
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HMclear(keys ...[]byte) (int64, error) {
 | 
						|
	t := db.hashBatch
 | 
						|
	t.Lock()
 | 
						|
	defer t.Unlock()
 | 
						|
 | 
						|
	for _, key := range keys {
 | 
						|
		if err := checkKeySize(key); err != nil {
 | 
						|
			return 0, err
 | 
						|
		}
 | 
						|
 | 
						|
		db.hDelete(t, key)
 | 
						|
		db.rmExpire(t, HashType, key)
 | 
						|
	}
 | 
						|
 | 
						|
	err := t.Commit()
 | 
						|
	return int64(len(keys)), err
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) hFlush() (drop int64, err error) {
 | 
						|
	t := db.hashBatch
 | 
						|
 | 
						|
	t.Lock()
 | 
						|
	defer t.Unlock()
 | 
						|
 | 
						|
	return db.flushType(t, HashType)
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HScan(key []byte, count int, inclusive bool, match string) ([][]byte, error) {
 | 
						|
	return db.scan(HSizeType, key, count, inclusive, match)
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HExpire(key []byte, duration int64) (int64, error) {
 | 
						|
	if duration <= 0 {
 | 
						|
		return 0, errExpireValue
 | 
						|
	}
 | 
						|
 | 
						|
	return db.hExpireAt(key, time.Now().Unix()+duration)
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HExpireAt(key []byte, when int64) (int64, error) {
 | 
						|
	if when <= time.Now().Unix() {
 | 
						|
		return 0, errExpireValue
 | 
						|
	}
 | 
						|
 | 
						|
	return db.hExpireAt(key, when)
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HTTL(key []byte) (int64, error) {
 | 
						|
	if err := checkKeySize(key); err != nil {
 | 
						|
		return -1, err
 | 
						|
	}
 | 
						|
 | 
						|
	return db.ttl(HashType, key)
 | 
						|
}
 | 
						|
 | 
						|
func (db *DB) HPersist(key []byte) (int64, error) {
 | 
						|
	if err := checkKeySize(key); err != nil {
 | 
						|
		return 0, err
 | 
						|
	}
 | 
						|
 | 
						|
	t := db.hashBatch
 | 
						|
	t.Lock()
 | 
						|
	defer t.Unlock()
 | 
						|
 | 
						|
	n, err := db.rmExpire(t, HashType, key)
 | 
						|
	if err != nil {
 | 
						|
		return 0, err
 | 
						|
	}
 | 
						|
 | 
						|
	err = t.Commit()
 | 
						|
	return n, err
 | 
						|
}
 |