|
- package nodb
-
- import (
- "bytes"
- "encoding/binary"
- "errors"
- "time"
-
- "github.com/lunny/nodb/store"
- )
-
- const (
- MinScore int64 = -1<<63 + 1
- MaxScore int64 = 1<<63 - 1
- InvalidScore int64 = -1 << 63
-
- AggregateSum byte = 0
- AggregateMin byte = 1
- AggregateMax byte = 2
- )
-
- type ScorePair struct {
- Score int64
- Member []byte
- }
-
- var errZSizeKey = errors.New("invalid zsize key")
- var errZSetKey = errors.New("invalid zset key")
- var errZScoreKey = errors.New("invalid zscore key")
- var errScoreOverflow = errors.New("zset score overflow")
- var errInvalidAggregate = errors.New("invalid aggregate")
- var errInvalidWeightNum = errors.New("invalid weight number")
- var errInvalidSrcKeyNum = errors.New("invalid src key number")
-
- const (
- zsetNScoreSep byte = '<'
- zsetPScoreSep byte = zsetNScoreSep + 1
- zsetStopScoreSep byte = zsetPScoreSep + 1
-
- zsetStartMemSep byte = ':'
- zsetStopMemSep byte = zsetStartMemSep + 1
- )
-
- func checkZSetKMSize(key []byte, member []byte) error {
- if len(key) > MaxKeySize || len(key) == 0 {
- return errKeySize
- } else if len(member) > MaxZSetMemberSize || len(member) == 0 {
- return errZSetMemberSize
- }
- return nil
- }
-
- func (db *DB) zEncodeSizeKey(key []byte) []byte {
- buf := make([]byte, len(key)+2)
- buf[0] = db.index
- buf[1] = ZSizeType
-
- copy(buf[2:], key)
- return buf
- }
-
- func (db *DB) zDecodeSizeKey(ek []byte) ([]byte, error) {
- if len(ek) < 2 || ek[0] != db.index || ek[1] != ZSizeType {
- return nil, errZSizeKey
- }
-
- return ek[2:], nil
- }
-
- func (db *DB) zEncodeSetKey(key []byte, member []byte) []byte {
- buf := make([]byte, len(key)+len(member)+5)
-
- pos := 0
- buf[pos] = db.index
- pos++
-
- buf[pos] = ZSetType
- pos++
-
- binary.BigEndian.PutUint16(buf[pos:], uint16(len(key)))
- pos += 2
-
- copy(buf[pos:], key)
- pos += len(key)
-
- buf[pos] = zsetStartMemSep
- pos++
-
- copy(buf[pos:], member)
-
- return buf
- }
-
- func (db *DB) zDecodeSetKey(ek []byte) ([]byte, []byte, error) {
- if len(ek) < 5 || ek[0] != db.index || ek[1] != ZSetType {
- return nil, nil, errZSetKey
- }
-
- keyLen := int(binary.BigEndian.Uint16(ek[2:]))
- if keyLen+5 > len(ek) {
- return nil, nil, errZSetKey
- }
-
- key := ek[4 : 4+keyLen]
-
- if ek[4+keyLen] != zsetStartMemSep {
- return nil, nil, errZSetKey
- }
-
- member := ek[5+keyLen:]
- return key, member, nil
- }
-
- func (db *DB) zEncodeStartSetKey(key []byte) []byte {
- k := db.zEncodeSetKey(key, nil)
- return k
- }
-
- func (db *DB) zEncodeStopSetKey(key []byte) []byte {
- k := db.zEncodeSetKey(key, nil)
- k[len(k)-1] = zsetStartMemSep + 1
- return k
- }
-
- func (db *DB) zEncodeScoreKey(key []byte, member []byte, score int64) []byte {
- buf := make([]byte, len(key)+len(member)+14)
-
- pos := 0
- buf[pos] = db.index
- pos++
-
- buf[pos] = ZScoreType
- pos++
-
- binary.BigEndian.PutUint16(buf[pos:], uint16(len(key)))
- pos += 2
-
- copy(buf[pos:], key)
- pos += len(key)
-
- if score < 0 {
- buf[pos] = zsetNScoreSep
- } else {
- buf[pos] = zsetPScoreSep
- }
-
- pos++
- binary.BigEndian.PutUint64(buf[pos:], uint64(score))
- pos += 8
-
- buf[pos] = zsetStartMemSep
- pos++
-
- copy(buf[pos:], member)
- return buf
- }
-
- func (db *DB) zEncodeStartScoreKey(key []byte, score int64) []byte {
- return db.zEncodeScoreKey(key, nil, score)
- }
-
- func (db *DB) zEncodeStopScoreKey(key []byte, score int64) []byte {
- k := db.zEncodeScoreKey(key, nil, score)
- k[len(k)-1] = zsetStopMemSep
- return k
- }
-
- func (db *DB) zDecodeScoreKey(ek []byte) (key []byte, member []byte, score int64, err error) {
- if len(ek) < 14 || ek[0] != db.index || ek[1] != ZScoreType {
- err = errZScoreKey
- return
- }
-
- keyLen := int(binary.BigEndian.Uint16(ek[2:]))
- if keyLen+14 > len(ek) {
- err = errZScoreKey
- return
- }
-
- key = ek[4 : 4+keyLen]
- pos := 4 + keyLen
-
- if (ek[pos] != zsetNScoreSep) && (ek[pos] != zsetPScoreSep) {
- err = errZScoreKey
- return
- }
- pos++
-
- score = int64(binary.BigEndian.Uint64(ek[pos:]))
- pos += 8
-
- if ek[pos] != zsetStartMemSep {
- err = errZScoreKey
- return
- }
-
- pos++
-
- member = ek[pos:]
- return
- }
-
- func (db *DB) zSetItem(t *batch, key []byte, score int64, member []byte) (int64, error) {
- if score <= MinScore || score >= MaxScore {
- return 0, errScoreOverflow
- }
-
- var exists int64 = 0
- ek := db.zEncodeSetKey(key, member)
-
- if v, err := db.bucket.Get(ek); err != nil {
- return 0, err
- } else if v != nil {
- exists = 1
-
- if s, err := Int64(v, err); err != nil {
- return 0, err
- } else {
- sk := db.zEncodeScoreKey(key, member, s)
- t.Delete(sk)
- }
- }
-
- t.Put(ek, PutInt64(score))
-
- sk := db.zEncodeScoreKey(key, member, score)
- t.Put(sk, []byte{})
-
- return exists, nil
- }
-
- func (db *DB) zDelItem(t *batch, key []byte, member []byte, skipDelScore bool) (int64, error) {
- ek := db.zEncodeSetKey(key, member)
- if v, err := db.bucket.Get(ek); err != nil {
- return 0, err
- } else if v == nil {
- //not exists
- return 0, nil
- } else {
- //exists
- if !skipDelScore {
- //we must del score
- if s, err := Int64(v, err); err != nil {
- return 0, err
- } else {
- sk := db.zEncodeScoreKey(key, member, s)
- t.Delete(sk)
- }
- }
- }
-
- t.Delete(ek)
-
- return 1, nil
- }
-
- func (db *DB) zDelete(t *batch, key []byte) int64 {
- delMembCnt, _ := db.zRemRange(t, key, MinScore, MaxScore, 0, -1)
- // todo : log err
- return delMembCnt
- }
-
- func (db *DB) zExpireAt(key []byte, when int64) (int64, error) {
- t := db.zsetBatch
- t.Lock()
- defer t.Unlock()
-
- if zcnt, err := db.ZCard(key); err != nil || zcnt == 0 {
- return 0, err
- } else {
- db.expireAt(t, ZSetType, key, when)
- if err := t.Commit(); err != nil {
- return 0, err
- }
- }
- return 1, nil
- }
-
- func (db *DB) ZAdd(key []byte, args ...ScorePair) (int64, error) {
- if len(args) == 0 {
- return 0, nil
- }
-
- t := db.zsetBatch
- t.Lock()
- defer t.Unlock()
-
- var num int64 = 0
- for i := 0; i < len(args); i++ {
- score := args[i].Score
- member := args[i].Member
-
- if err := checkZSetKMSize(key, member); err != nil {
- return 0, err
- }
-
- if n, err := db.zSetItem(t, key, score, member); err != nil {
- return 0, err
- } else if n == 0 {
- //add new
- num++
- }
- }
-
- if _, err := db.zIncrSize(t, key, num); err != nil {
- return 0, err
- }
-
- //todo add binlog
- err := t.Commit()
- return num, err
- }
-
- func (db *DB) zIncrSize(t *batch, key []byte, delta int64) (int64, error) {
- sk := db.zEncodeSizeKey(key)
-
- size, err := Int64(db.bucket.Get(sk))
- if err != nil {
- return 0, err
- } else {
- size += delta
- if size <= 0 {
- size = 0
- t.Delete(sk)
- db.rmExpire(t, ZSetType, key)
- } else {
- t.Put(sk, PutInt64(size))
- }
- }
-
- return size, nil
- }
-
- func (db *DB) ZCard(key []byte) (int64, error) {
- if err := checkKeySize(key); err != nil {
- return 0, err
- }
-
- sk := db.zEncodeSizeKey(key)
- return Int64(db.bucket.Get(sk))
- }
-
- func (db *DB) ZScore(key []byte, member []byte) (int64, error) {
- if err := checkZSetKMSize(key, member); err != nil {
- return InvalidScore, err
- }
-
- var score int64 = InvalidScore
-
- k := db.zEncodeSetKey(key, member)
- if v, err := db.bucket.Get(k); err != nil {
- return InvalidScore, err
- } else if v == nil {
- return InvalidScore, ErrScoreMiss
- } else {
- if score, err = Int64(v, nil); err != nil {
- return InvalidScore, err
- }
- }
-
- return score, nil
- }
-
- func (db *DB) ZRem(key []byte, members ...[]byte) (int64, error) {
- if len(members) == 0 {
- return 0, nil
- }
-
- t := db.zsetBatch
- t.Lock()
- defer t.Unlock()
-
- var num int64 = 0
- for i := 0; i < len(members); i++ {
- if err := checkZSetKMSize(key, members[i]); err != nil {
- return 0, err
- }
-
- if n, err := db.zDelItem(t, key, members[i], false); err != nil {
- return 0, err
- } else if n == 1 {
- num++
- }
- }
-
- if _, err := db.zIncrSize(t, key, -num); err != nil {
- return 0, err
- }
-
- err := t.Commit()
- return num, err
- }
-
- func (db *DB) ZIncrBy(key []byte, delta int64, member []byte) (int64, error) {
- if err := checkZSetKMSize(key, member); err != nil {
- return InvalidScore, err
- }
-
- t := db.zsetBatch
- t.Lock()
- defer t.Unlock()
-
- ek := db.zEncodeSetKey(key, member)
-
- var oldScore int64 = 0
- v, err := db.bucket.Get(ek)
- if err != nil {
- return InvalidScore, err
- } else if v == nil {
- db.zIncrSize(t, key, 1)
- } else {
- if oldScore, err = Int64(v, err); err != nil {
- return InvalidScore, err
- }
- }
-
- newScore := oldScore + delta
- if newScore >= MaxScore || newScore <= MinScore {
- return InvalidScore, errScoreOverflow
- }
-
- sk := db.zEncodeScoreKey(key, member, newScore)
- t.Put(sk, []byte{})
- t.Put(ek, PutInt64(newScore))
-
- if v != nil {
- // so as to update score, we must delete the old one
- oldSk := db.zEncodeScoreKey(key, member, oldScore)
- t.Delete(oldSk)
- }
-
- err = t.Commit()
- return newScore, err
- }
-
- func (db *DB) ZCount(key []byte, min int64, max int64) (int64, error) {
- if err := checkKeySize(key); err != nil {
- return 0, err
- }
- minKey := db.zEncodeStartScoreKey(key, min)
- maxKey := db.zEncodeStopScoreKey(key, max)
-
- rangeType := store.RangeROpen
-
- it := db.bucket.RangeLimitIterator(minKey, maxKey, rangeType, 0, -1)
- var n int64 = 0
- for ; it.Valid(); it.Next() {
- n++
- }
- it.Close()
-
- return n, nil
- }
-
- func (db *DB) zrank(key []byte, member []byte, reverse bool) (int64, error) {
- if err := checkZSetKMSize(key, member); err != nil {
- return 0, err
- }
-
- k := db.zEncodeSetKey(key, member)
-
- it := db.bucket.NewIterator()
- defer it.Close()
-
- if v := it.Find(k); v == nil {
- return -1, nil
- } else {
- if s, err := Int64(v, nil); err != nil {
- return 0, err
- } else {
- var rit *store.RangeLimitIterator
-
- sk := db.zEncodeScoreKey(key, member, s)
-
- if !reverse {
- minKey := db.zEncodeStartScoreKey(key, MinScore)
-
- rit = store.NewRangeIterator(it, &store.Range{minKey, sk, store.RangeClose})
- } else {
- maxKey := db.zEncodeStopScoreKey(key, MaxScore)
- rit = store.NewRevRangeIterator(it, &store.Range{sk, maxKey, store.RangeClose})
- }
-
- var lastKey []byte = nil
- var n int64 = 0
-
- for ; rit.Valid(); rit.Next() {
- n++
-
- lastKey = rit.BufKey(lastKey)
- }
-
- if _, m, _, err := db.zDecodeScoreKey(lastKey); err == nil && bytes.Equal(m, member) {
- n--
- return n, nil
- }
- }
- }
-
- return -1, nil
- }
-
- func (db *DB) zIterator(key []byte, min int64, max int64, offset int, count int, reverse bool) *store.RangeLimitIterator {
- minKey := db.zEncodeStartScoreKey(key, min)
- maxKey := db.zEncodeStopScoreKey(key, max)
-
- if !reverse {
- return db.bucket.RangeLimitIterator(minKey, maxKey, store.RangeClose, offset, count)
- } else {
- return db.bucket.RevRangeLimitIterator(minKey, maxKey, store.RangeClose, offset, count)
- }
- }
-
- func (db *DB) zRemRange(t *batch, key []byte, min int64, max int64, offset int, count int) (int64, error) {
- if len(key) > MaxKeySize {
- return 0, errKeySize
- }
-
- it := db.zIterator(key, min, max, offset, count, false)
- var num int64 = 0
- for ; it.Valid(); it.Next() {
- sk := it.RawKey()
- _, m, _, err := db.zDecodeScoreKey(sk)
- if err != nil {
- continue
- }
-
- if n, err := db.zDelItem(t, key, m, true); err != nil {
- return 0, err
- } else if n == 1 {
- num++
- }
-
- t.Delete(sk)
- }
- it.Close()
-
- if _, err := db.zIncrSize(t, key, -num); err != nil {
- return 0, err
- }
-
- return num, nil
- }
-
- func (db *DB) zRange(key []byte, min int64, max int64, offset int, count int, reverse bool) ([]ScorePair, error) {
- if len(key) > MaxKeySize {
- return nil, errKeySize
- }
-
- if offset < 0 {
- return []ScorePair{}, nil
- }
-
- nv := 64
- if count > 0 {
- nv = count
- }
-
- v := make([]ScorePair, 0, nv)
-
- var it *store.RangeLimitIterator
-
- //if reverse and offset is 0, count < 0, we may use forward iterator then reverse
- //because store iterator prev is slower than next
- if !reverse || (offset == 0 && count < 0) {
- it = db.zIterator(key, min, max, offset, count, false)
- } else {
- it = db.zIterator(key, min, max, offset, count, true)
- }
-
- for ; it.Valid(); it.Next() {
- _, m, s, err := db.zDecodeScoreKey(it.Key())
- //may be we will check key equal?
- if err != nil {
- continue
- }
-
- v = append(v, ScorePair{Member: m, Score: s})
- }
- it.Close()
-
- if reverse && (offset == 0 && count < 0) {
- for i, j := 0, len(v)-1; i < j; i, j = i+1, j-1 {
- v[i], v[j] = v[j], v[i]
- }
- }
-
- return v, nil
- }
-
- func (db *DB) zParseLimit(key []byte, start int, stop int) (offset int, count int, err error) {
- if start < 0 || stop < 0 {
- //refer redis implementation
- var size int64
- size, err = db.ZCard(key)
- if err != nil {
- return
- }
-
- llen := int(size)
-
- if start < 0 {
- start = llen + start
- }
- if stop < 0 {
- stop = llen + stop
- }
-
- if start < 0 {
- start = 0
- }
-
- if start >= llen {
- offset = -1
- return
- }
- }
-
- if start > stop {
- offset = -1
- return
- }
-
- offset = start
- count = (stop - start) + 1
- return
- }
-
- func (db *DB) ZClear(key []byte) (int64, error) {
- t := db.zsetBatch
- t.Lock()
- defer t.Unlock()
-
- rmCnt, err := db.zRemRange(t, key, MinScore, MaxScore, 0, -1)
- if err == nil {
- err = t.Commit()
- }
-
- return rmCnt, err
- }
-
- func (db *DB) ZMclear(keys ...[]byte) (int64, error) {
- t := db.zsetBatch
- t.Lock()
- defer t.Unlock()
-
- for _, key := range keys {
- if _, err := db.zRemRange(t, key, MinScore, MaxScore, 0, -1); err != nil {
- return 0, err
- }
- }
-
- err := t.Commit()
-
- return int64(len(keys)), err
- }
-
- func (db *DB) ZRange(key []byte, start int, stop int) ([]ScorePair, error) {
- return db.ZRangeGeneric(key, start, stop, false)
- }
-
- //min and max must be inclusive
- //if no limit, set offset = 0 and count = -1
- func (db *DB) ZRangeByScore(key []byte, min int64, max int64,
- offset int, count int) ([]ScorePair, error) {
- return db.ZRangeByScoreGeneric(key, min, max, offset, count, false)
- }
-
- func (db *DB) ZRank(key []byte, member []byte) (int64, error) {
- return db.zrank(key, member, false)
- }
-
- func (db *DB) ZRemRangeByRank(key []byte, start int, stop int) (int64, error) {
- offset, count, err := db.zParseLimit(key, start, stop)
- if err != nil {
- return 0, err
- }
-
- var rmCnt int64
-
- t := db.zsetBatch
- t.Lock()
- defer t.Unlock()
-
- rmCnt, err = db.zRemRange(t, key, MinScore, MaxScore, offset, count)
- if err == nil {
- err = t.Commit()
- }
-
- return rmCnt, err
- }
-
- //min and max must be inclusive
- func (db *DB) ZRemRangeByScore(key []byte, min int64, max int64) (int64, error) {
- t := db.zsetBatch
- t.Lock()
- defer t.Unlock()
-
- rmCnt, err := db.zRemRange(t, key, min, max, 0, -1)
- if err == nil {
- err = t.Commit()
- }
-
- return rmCnt, err
- }
-
- func (db *DB) ZRevRange(key []byte, start int, stop int) ([]ScorePair, error) {
- return db.ZRangeGeneric(key, start, stop, true)
- }
-
- func (db *DB) ZRevRank(key []byte, member []byte) (int64, error) {
- return db.zrank(key, member, true)
- }
-
- //min and max must be inclusive
- //if no limit, set offset = 0 and count = -1
- func (db *DB) ZRevRangeByScore(key []byte, min int64, max int64, offset int, count int) ([]ScorePair, error) {
- return db.ZRangeByScoreGeneric(key, min, max, offset, count, true)
- }
-
- func (db *DB) ZRangeGeneric(key []byte, start int, stop int, reverse bool) ([]ScorePair, error) {
- offset, count, err := db.zParseLimit(key, start, stop)
- if err != nil {
- return nil, err
- }
-
- return db.zRange(key, MinScore, MaxScore, offset, count, reverse)
- }
-
- //min and max must be inclusive
- //if no limit, set offset = 0 and count = -1
- func (db *DB) ZRangeByScoreGeneric(key []byte, min int64, max int64,
- offset int, count int, reverse bool) ([]ScorePair, error) {
-
- return db.zRange(key, min, max, offset, count, reverse)
- }
-
- func (db *DB) zFlush() (drop int64, err error) {
- t := db.zsetBatch
- t.Lock()
- defer t.Unlock()
- return db.flushType(t, ZSetType)
- }
-
- func (db *DB) ZExpire(key []byte, duration int64) (int64, error) {
- if duration <= 0 {
- return 0, errExpireValue
- }
-
- return db.zExpireAt(key, time.Now().Unix()+duration)
- }
-
- func (db *DB) ZExpireAt(key []byte, when int64) (int64, error) {
- if when <= time.Now().Unix() {
- return 0, errExpireValue
- }
-
- return db.zExpireAt(key, when)
- }
-
- func (db *DB) ZTTL(key []byte) (int64, error) {
- if err := checkKeySize(key); err != nil {
- return -1, err
- }
-
- return db.ttl(ZSetType, key)
- }
-
- func (db *DB) ZPersist(key []byte) (int64, error) {
- if err := checkKeySize(key); err != nil {
- return 0, err
- }
-
- t := db.zsetBatch
- t.Lock()
- defer t.Unlock()
-
- n, err := db.rmExpire(t, ZSetType, key)
- if err != nil {
- return 0, err
- }
-
- err = t.Commit()
- return n, err
- }
-
- func getAggregateFunc(aggregate byte) func(int64, int64) int64 {
- switch aggregate {
- case AggregateSum:
- return func(a int64, b int64) int64 {
- return a + b
- }
- case AggregateMax:
- return func(a int64, b int64) int64 {
- if a > b {
- return a
- }
- return b
- }
- case AggregateMin:
- return func(a int64, b int64) int64 {
- if a > b {
- return b
- }
- return a
- }
- }
- return nil
- }
-
- func (db *DB) ZUnionStore(destKey []byte, srcKeys [][]byte, weights []int64, aggregate byte) (int64, error) {
-
- var destMap = map[string]int64{}
- aggregateFunc := getAggregateFunc(aggregate)
- if aggregateFunc == nil {
- return 0, errInvalidAggregate
- }
- if len(srcKeys) < 1 {
- return 0, errInvalidSrcKeyNum
- }
- if weights != nil {
- if len(srcKeys) != len(weights) {
- return 0, errInvalidWeightNum
- }
- } else {
- weights = make([]int64, len(srcKeys))
- for i := 0; i < len(weights); i++ {
- weights[i] = 1
- }
- }
-
- for i, key := range srcKeys {
- scorePairs, err := db.ZRange(key, 0, -1)
- if err != nil {
- return 0, err
- }
- for _, pair := range scorePairs {
- if score, ok := destMap[String(pair.Member)]; !ok {
- destMap[String(pair.Member)] = pair.Score * weights[i]
- } else {
- destMap[String(pair.Member)] = aggregateFunc(score, pair.Score*weights[i])
- }
- }
- }
-
- t := db.zsetBatch
- t.Lock()
- defer t.Unlock()
-
- db.zDelete(t, destKey)
-
- for member, score := range destMap {
- if err := checkZSetKMSize(destKey, []byte(member)); err != nil {
- return 0, err
- }
-
- if _, err := db.zSetItem(t, destKey, score, []byte(member)); err != nil {
- return 0, err
- }
- }
-
- var num = int64(len(destMap))
- sk := db.zEncodeSizeKey(destKey)
- t.Put(sk, PutInt64(num))
-
- //todo add binlog
- if err := t.Commit(); err != nil {
- return 0, err
- }
- return num, nil
- }
-
- func (db *DB) ZInterStore(destKey []byte, srcKeys [][]byte, weights []int64, aggregate byte) (int64, error) {
-
- aggregateFunc := getAggregateFunc(aggregate)
- if aggregateFunc == nil {
- return 0, errInvalidAggregate
- }
- if len(srcKeys) < 1 {
- return 0, errInvalidSrcKeyNum
- }
- if weights != nil {
- if len(srcKeys) != len(weights) {
- return 0, errInvalidWeightNum
- }
- } else {
- weights = make([]int64, len(srcKeys))
- for i := 0; i < len(weights); i++ {
- weights[i] = 1
- }
- }
-
- var destMap = map[string]int64{}
- scorePairs, err := db.ZRange(srcKeys[0], 0, -1)
- if err != nil {
- return 0, err
- }
- for _, pair := range scorePairs {
- destMap[String(pair.Member)] = pair.Score * weights[0]
- }
-
- for i, key := range srcKeys[1:] {
- scorePairs, err := db.ZRange(key, 0, -1)
- if err != nil {
- return 0, err
- }
- tmpMap := map[string]int64{}
- for _, pair := range scorePairs {
- if score, ok := destMap[String(pair.Member)]; ok {
- tmpMap[String(pair.Member)] = aggregateFunc(score, pair.Score*weights[i+1])
- }
- }
- destMap = tmpMap
- }
-
- t := db.zsetBatch
- t.Lock()
- defer t.Unlock()
-
- db.zDelete(t, destKey)
-
- for member, score := range destMap {
- if err := checkZSetKMSize(destKey, []byte(member)); err != nil {
- return 0, err
- }
- if _, err := db.zSetItem(t, destKey, score, []byte(member)); err != nil {
- return 0, err
- }
- }
-
- var num int64 = int64(len(destMap))
- sk := db.zEncodeSizeKey(destKey)
- t.Put(sk, PutInt64(num))
- //todo add binlog
- if err := t.Commit(); err != nil {
- return 0, err
- }
- return num, nil
- }
-
- func (db *DB) ZScan(key []byte, count int, inclusive bool, match string) ([][]byte, error) {
- return db.scan(ZSizeType, key, count, inclusive, match)
- }
|