ledisdb/ledis/t_bin.go

833 lines
17 KiB
Go
Raw Normal View History

2014-06-30 06:50:42 +04:00
package ledis
import (
"encoding/binary"
"errors"
"github.com/siddontang/ledisdb/leveldb"
"sort"
2014-07-03 07:38:52 +04:00
"time"
2014-06-30 06:50:42 +04:00
)
const (
OPand uint8 = iota + 1
OPor
OPxor
OPnot
)
2014-07-03 07:38:52 +04:00
type BitPair struct {
Pos int32
Val uint8
}
type segBitInfo struct {
Seq uint32
Off uint32
Val uint8
}
type segBitInfoArray []segBitInfo
2014-06-30 06:50:42 +04:00
const (
// byte
segByteWidth uint32 = 9
segByteSize uint32 = 1 << segByteWidth
// bit
segBitWidth uint32 = segByteWidth + 3
segBitSize uint32 = segByteSize << 3
maxByteSize uint32 = 8 << 20
maxSegCount uint32 = maxByteSize / segByteSize
minSeq uint32 = 0
maxSeq uint32 = uint32((maxByteSize << 3) - 1)
)
var bitsInByte = [256]int32{0, 1, 1, 2, 1, 2, 2, 3, 1, 2, 2, 3, 2, 3, 3,
4, 1, 2, 2, 3, 2, 3, 3, 4, 2, 3, 3, 4, 3, 4, 4, 5, 1, 2, 2, 3, 2, 3,
3, 4, 2, 3, 3, 4, 3, 4, 4, 5, 2, 3, 3, 4, 3, 4, 4, 5, 3, 4, 4, 5, 4,
5, 5, 6, 1, 2, 2, 3, 2, 3, 3, 4, 2, 3, 3, 4, 3, 4, 4, 5, 2, 3, 3, 4,
3, 4, 4, 5, 3, 4, 4, 5, 4, 5, 5, 6, 2, 3, 3, 4, 3, 4, 4, 5, 3, 4, 4,
5, 4, 5, 5, 6, 3, 4, 4, 5, 4, 5, 5, 6, 4, 5, 5, 6, 5, 6, 6, 7, 1, 2,
2, 3, 2, 3, 3, 4, 2, 3, 3, 4, 3, 4, 4, 5, 2, 3, 3, 4, 3, 4, 4, 5, 3,
4, 4, 5, 4, 5, 5, 6, 2, 3, 3, 4, 3, 4, 4, 5, 3, 4, 4, 5, 4, 5, 5, 6,
3, 4, 4, 5, 4, 5, 5, 6, 4, 5, 5, 6, 5, 6, 6, 7, 2, 3, 3, 4, 3, 4, 4,
5, 3, 4, 4, 5, 4, 5, 5, 6, 3, 4, 4, 5, 4, 5, 5, 6, 4, 5, 5, 6, 5, 6,
6, 7, 3, 4, 4, 5, 4, 5, 5, 6, 4, 5, 5, 6, 5, 6, 6, 7, 4, 5, 5, 6, 5,
6, 6, 7, 5, 6, 6, 7, 6, 7, 7, 8}
var fillBits = [...]uint8{1, 3, 7, 15, 31, 63, 127, 255}
var emptySegment []byte = make([]byte, segByteSize, segByteSize)
2014-06-30 06:50:42 +04:00
var fillSegment []byte = func() []byte {
data := make([]byte, segByteSize, segByteSize)
for i := uint32(0); i < segByteSize; i++ {
data[i] = 0xff
}
return data
}()
var errBinKey = errors.New("invalid bin key")
var errOffset = errors.New("invalid offset")
var errDuplicatePos = errors.New("duplicate bit pos")
2014-06-30 06:50:42 +04:00
func getBit(sz []byte, offset uint32) uint8 {
index := offset >> 3
if index >= uint32(len(sz)) {
return 0 // error("overflow")
}
offset -= index << 3
return sz[index] >> offset & 1
}
func setBit(sz []byte, offset uint32, val uint8) bool {
if val != 1 && val != 0 {
return false // error("invalid val")
}
index := offset >> 3
if index >= uint32(len(sz)) {
return false // error("overflow")
}
offset -= index << 3
if sz[index]>>offset&1 != val {
sz[index] ^= (1 << offset)
}
return true
}
2014-07-08 13:10:44 +04:00
func (datas segBitInfoArray) Len() int {
return len(datas)
}
2014-07-08 13:10:44 +04:00
func (datas segBitInfoArray) Less(i, j int) bool {
res := (datas)[i].Seq < (datas)[j].Seq
if !res && (datas)[i].Seq == (datas)[j].Seq {
res = (datas)[i].Off < (datas)[j].Off
}
return res
}
2014-07-08 13:10:44 +04:00
func (datas segBitInfoArray) Swap(i, j int) {
datas[i], datas[j] = datas[j], datas[i]
}
2014-06-30 06:50:42 +04:00
func (db *DB) bEncodeMetaKey(key []byte) []byte {
mk := make([]byte, len(key)+2)
mk[0] = db.index
mk[1] = binMetaType
copy(mk, key)
return mk
}
func (db *DB) bEncodeBinKey(key []byte, seq uint32) []byte {
bk := make([]byte, len(key)+8)
pos := 0
bk[pos] = db.index
pos++
bk[pos] = binType
pos++
binary.BigEndian.PutUint16(bk[pos:], uint16(len(key)))
pos += 2
copy(bk[pos:], key)
pos += len(key)
binary.BigEndian.PutUint32(bk[pos:], seq)
return bk
}
func (db *DB) bDecodeBinKey(bkey []byte) (key []byte, seq uint32, err error) {
if len(bkey) < 8 || bkey[0] != db.index {
err = errBinKey
return
}
keyLen := binary.BigEndian.Uint16(bkey[2:4])
if int(keyLen+8) != len(bkey) {
err = errBinKey
return
}
key = bkey[4 : 4+keyLen]
seq = uint32(binary.BigEndian.Uint32(bkey[4+keyLen:]))
return
}
func (db *DB) bCapByteSize(seq uint32, off uint32) uint32 {
var offByteSize uint32 = (off >> 3) + 1
if offByteSize > segByteSize {
offByteSize = segByteSize
}
return seq<<segByteWidth + offByteSize
}
func (db *DB) bParseOffset(key []byte, offset int32) (seq uint32, off uint32, err error) {
if offset < 0 {
if tailSeq, tailOff, e := db.bGetMeta(key); e != nil {
err = e
return
} else if tailSeq >= 0 {
offset += int32((uint32(tailSeq)<<segBitWidth | uint32(tailOff)) + 1)
2014-06-30 06:50:42 +04:00
if offset < 0 {
err = errOffset
return
}
}
}
off = uint32(offset)
seq = off >> segBitWidth
off &= (segBitSize - 1)
return
}
func (db *DB) bGetMeta(key []byte) (tailSeq int32, tailOff int32, err error) {
var v []byte
mk := db.bEncodeMetaKey(key)
v, err = db.db.Get(mk)
if err != nil {
return
}
if v != nil {
tailSeq = int32(binary.LittleEndian.Uint32(v[0:4]))
tailOff = int32(binary.LittleEndian.Uint32(v[4:8]))
} else {
tailSeq = -1
tailOff = -1
}
return
}
func (db *DB) bSetMeta(t *tx, key []byte, tailSeq uint32, tailOff uint32) {
ek := db.bEncodeMetaKey(key)
buf := make([]byte, 8)
binary.LittleEndian.PutUint32(buf[0:4], tailSeq)
binary.LittleEndian.PutUint32(buf[4:8], tailOff)
t.Put(ek, buf)
return
}
func (db *DB) bUpdateMeta(t *tx, key []byte, seq uint32, off uint32) (tailSeq uint32, tailOff uint32, err error) {
var ts, to int32
if ts, to, err = db.bGetMeta(key); err != nil {
return
} else {
tailSeq = uint32(MaxInt32(ts, 0))
tailOff = uint32(MaxInt32(to, 0))
}
if seq > tailSeq || (seq == tailSeq && off > tailOff) {
db.bSetMeta(t, key, seq, off)
tailSeq = seq
tailOff = off
}
return
}
func (db *DB) bDelete(t *tx, key []byte) (drop int64) {
mk := db.bEncodeMetaKey(key)
t.Delete(mk)
minKey := db.bEncodeBinKey(key, minSeq)
maxKey := db.bEncodeBinKey(key, maxSeq)
it := db.db.RangeIterator(minKey, maxKey, leveldb.RangeClose)
for ; it.Valid(); it.Next() {
t.Delete(it.Key())
drop++
2014-06-30 06:50:42 +04:00
}
it.Close()
return drop
}
2014-07-03 07:38:52 +04:00
func (db *DB) bGetSegment(key []byte, seq uint32) ([]byte, []byte, error) {
bk := db.bEncodeBinKey(key, seq)
segment, err := db.db.Get(bk)
if err != nil {
return bk, nil, err
}
return bk, segment, nil
}
func (db *DB) bAllocateSegment(key []byte, seq uint32) ([]byte, []byte, error) {
bk, segment, err := db.bGetSegment(key, seq)
if err == nil && segment == nil {
segment = make([]byte, segByteSize, segByteSize)
}
return bk, segment, err
}
func (db *DB) bIterator(key []byte) *leveldb.RangeLimitIterator {
sk := db.bEncodeBinKey(key, minSeq)
ek := db.bEncodeBinKey(key, maxSeq)
return db.db.RangeIterator(sk, ek, leveldb.RangeClose)
}
func (db *DB) bSegAnd(a []byte, b []byte, res *[]byte) {
if a == nil || b == nil {
*res = nil
return
}
data := *res
if data == nil {
data = make([]byte, segByteSize, segByteSize)
*res = data
}
for i := uint32(0); i < segByteSize; i++ {
data[i] = a[i] & b[i]
}
return
}
func (db *DB) bSegOr(a []byte, b []byte, res *[]byte) {
if a == nil || b == nil {
if a == nil && b == nil {
*res = nil
} else if a == nil {
*res = b
} else {
*res = a
}
return
}
data := *res
if data == nil {
data = make([]byte, segByteSize, segByteSize)
*res = data
}
for i := uint32(0); i < segByteSize; i++ {
data[i] = a[i] | b[i]
}
return
}
func (db *DB) bSegXor(a []byte, b []byte, res *[]byte) {
if a == nil && b == nil {
*res = fillSegment
return
}
if a == nil {
a = emptySegment
}
if b == nil {
b = emptySegment
}
data := *res
if data == nil {
data = make([]byte, segByteSize, segByteSize)
*res = data
}
for i := uint32(0); i < segByteSize; i++ {
data[i] = a[i] ^ b[i]
}
return
}
func (db *DB) bExpireAt(key []byte, when int64) (int64, error) {
t := db.binTx
t.Lock()
defer t.Unlock()
if seq, _, err := db.bGetMeta(key); err != nil || seq < 0 {
return 0, err
} else {
db.expireAt(t, bExpType, key, when)
if err := t.Commit(); err != nil {
return 0, err
}
}
return 1, nil
}
2014-06-30 06:50:42 +04:00
func (db *DB) BGet(key []byte) (data []byte, err error) {
if err = checkKeySize(key); err != nil {
return
}
var ts, to int32
if ts, to, err = db.bGetMeta(key); err != nil || ts < 0 {
return
}
var tailSeq, tailOff = uint32(ts), uint32(to)
var capByteSize uint32 = db.bCapByteSize(tailSeq, tailOff)
data = make([]byte, capByteSize, capByteSize)
minKey := db.bEncodeBinKey(key, minSeq)
maxKey := db.bEncodeBinKey(key, tailSeq)
it := db.db.RangeIterator(minKey, maxKey, leveldb.RangeClose)
var seq, s, e uint32
for ; it.Valid(); it.Next() {
if _, seq, err = db.bDecodeBinKey(it.Key()); err != nil {
data = nil
break
}
s = seq << segByteWidth
e = MinUInt32(s+segByteSize, capByteSize)
copy(data[s:e], it.Value())
}
it.Close()
return
}
func (db *DB) BDelete(key []byte) (drop int64, err error) {
if err = checkKeySize(key); err != nil {
return
}
t := db.binTx
t.Lock()
defer t.Unlock()
drop = db.bDelete(t, key)
db.rmExpire(t, bExpType, key)
err = t.Commit()
return
}
func (db *DB) BSetBit(key []byte, offset int32, val uint8) (ori uint8, err error) {
if err = checkKeySize(key); err != nil {
return
}
// todo : check offset
var seq, off uint32
if seq, off, err = db.bParseOffset(key, offset); err != nil {
return 0, err
}
var bk, segment []byte
2014-07-03 07:38:52 +04:00
if bk, segment, err = db.bAllocateSegment(key, seq); err != nil {
2014-06-30 06:50:42 +04:00
return 0, err
}
if segment != nil {
ori = getBit(segment, off)
if setBit(segment, off, val) {
t := db.binTx
t.Lock()
2014-06-30 06:50:42 +04:00
t.Put(bk, segment)
if _, _, e := db.bUpdateMeta(t, key, seq, off); e != nil {
err = e
return
}
err = t.Commit()
t.Unlock()
2014-06-30 06:50:42 +04:00
}
}
return
}
2014-07-08 13:10:44 +04:00
func (db *DB) BMSetBit(key []byte, args ...BitPair) (place int64, err error) {
if err = checkKeySize(key); err != nil {
return
}
2014-07-03 07:38:52 +04:00
2014-07-08 13:10:44 +04:00
// (ps : so as to aviod wasting memory copy while calling db.Get() and batch.Put(),
// here we sequence the params by pos, so that we can merge the execution of
// diff pos setting which targets on the same segment respectively. )
2014-07-03 07:38:52 +04:00
2014-07-08 13:10:44 +04:00
// #1 : sequence request data
var argCnt = len(args)
2014-07-08 13:10:44 +04:00
var bitInfos segBitInfoArray = make(segBitInfoArray, argCnt)
var seq, off uint32
2014-07-03 07:38:52 +04:00
for i, info := range args {
if seq, off, err = db.bParseOffset(key, info.Pos); err != nil {
return
}
2014-07-03 07:38:52 +04:00
2014-07-08 13:10:44 +04:00
bitInfos[i].Seq = seq
bitInfos[i].Off = off
bitInfos[i].Val = info.Val
}
2014-07-03 07:38:52 +04:00
2014-07-08 13:10:44 +04:00
sort.Sort(bitInfos)
2014-07-03 07:38:52 +04:00
for i := 1; i < argCnt; i++ {
2014-07-08 13:10:44 +04:00
if bitInfos[i].Seq == bitInfos[i-1].Seq && bitInfos[i].Off == bitInfos[i-1].Off {
return 0, errDuplicatePos
}
}
2014-07-08 13:10:44 +04:00
// #2 : execute bit set in order
t := db.binTx
t.Lock()
defer t.Unlock()
2014-07-08 13:10:44 +04:00
var curBinKey, curSeg []byte
var curSeq, maxSeq, maxOff uint32
2014-07-08 13:10:44 +04:00
for _, info := range bitInfos {
if curSeg != nil && info.Seq != curSeq {
t.Put(curBinKey, curSeg)
curSeg = nil
}
2014-07-08 13:10:44 +04:00
if curSeg == nil {
curSeq = info.Seq
if curBinKey, curSeg, err = db.bAllocateSegment(key, info.Seq); err != nil {
return
}
2014-07-08 13:10:44 +04:00
if curSeg == nil {
continue
}
}
2014-07-08 13:10:44 +04:00
if setBit(curSeg, info.Off, info.Val) {
maxSeq = info.Seq
maxOff = info.Off
place++
}
}
2014-07-08 13:10:44 +04:00
if curSeg != nil {
t.Put(curBinKey, curSeg)
}
// finally, update meta
if place > 0 {
if _, _, err = db.bUpdateMeta(t, key, maxSeq, maxOff); err != nil {
return
}
err = t.Commit()
}
return
}
2014-07-03 07:38:52 +04:00
2014-06-30 06:50:42 +04:00
func (db *DB) BGetBit(key []byte, offset int32) (uint8, error) {
if seq, off, err := db.bParseOffset(key, offset); err != nil {
return 0, err
} else {
2014-07-03 07:38:52 +04:00
_, segment, err := db.bGetSegment(key, seq)
2014-06-30 06:50:42 +04:00
if err != nil {
return 0, err
}
if segment == nil {
return 0, nil
} else {
return getBit(segment, off), nil
}
}
}
// func (db *DB) BGetRange(key []byte, start int32, end int32) ([]byte, error) {
// section := make([]byte)
// return
// }
func (db *DB) BCount(key []byte, start int32, end int32) (cnt int32, err error) {
var sseq uint32
if sseq, _, err = db.bParseOffset(key, start); err != nil {
return
}
var eseq uint32
if eseq, _, err = db.bParseOffset(key, end); err != nil {
return
}
var segment []byte
skey := db.bEncodeBinKey(key, sseq)
ekey := db.bEncodeBinKey(key, eseq)
it := db.db.RangeIterator(skey, ekey, leveldb.RangeClose)
for ; it.Valid(); it.Next() {
segment = it.Value()
for _, bit := range segment {
cnt += bitsInByte[bit]
}
}
it.Close()
return
}
func (db *DB) BTail(key []byte) (int32, error) {
// effective length of data, the highest bit-pos set in history
tailSeq, tailOff, err := db.bGetMeta(key)
if err != nil {
return 0, err
}
tail := int32(-1)
if tailSeq >= 0 {
tail = int32(uint32(tailSeq)<<segBitWidth | uint32(tailOff))
}
return tail, nil
}
func (db *DB) BOperation(op uint8, dstkey []byte, srckeys ...[]byte) (blen int32, err error) {
2014-07-08 13:10:44 +04:00
// blen -
// the size of the string stored in the destination key,
2014-06-30 06:50:42 +04:00
// that is equal to the size of the longest input string.
2014-06-30 06:50:42 +04:00
var exeOp func([]byte, []byte, *[]byte)
switch op {
case OPand:
exeOp = db.bSegAnd
case OPor:
exeOp = db.bSegOr
case OPxor, OPnot:
exeOp = db.bSegXor
2014-06-30 06:50:42 +04:00
default:
return
}
if dstkey == nil || srckeys == nil {
2014-06-30 06:50:42 +04:00
return
}
t := db.binTx
t.Lock()
defer t.Unlock()
2014-07-08 13:10:44 +04:00
var srcKseq, srcKoff int32
var seq, off, maxDstSeq, maxDstOff uint32
2014-06-30 06:50:42 +04:00
var keyNum int = len(srckeys)
2014-07-08 13:10:44 +04:00
var validKeyNum int
for i := 0; i < keyNum; i++ {
if srcKseq, srcKoff, err = db.bGetMeta(srckeys[i]); err != nil {
2014-06-30 06:50:42 +04:00
return
2014-07-08 13:10:44 +04:00
} else if srcKseq < 0 {
srckeys[i] = nil
continue
2014-06-30 06:50:42 +04:00
}
2014-07-08 13:10:44 +04:00
validKeyNum++
seq = uint32(srcKseq)
off = uint32(srcKoff)
if seq > maxDstSeq || (seq == maxDstSeq && off > maxDstOff) {
maxDstSeq = seq
maxDstOff = off
2014-06-30 06:50:42 +04:00
}
}
2014-07-08 13:10:44 +04:00
if (op == OPnot && validKeyNum != 1) ||
(op != OPnot && validKeyNum < 2) {
return // with not enough existing source key
}
var srcIdx int
for srcIdx = 0; srcIdx < keyNum; srcIdx++ {
if srckeys[srcIdx] != nil {
break
}
2014-06-30 06:50:42 +04:00
}
// init - data
2014-07-08 13:10:44 +04:00
var segments = make([][]byte, maxDstSeq+1)
2014-06-30 06:50:42 +04:00
if op == OPnot {
2014-07-08 13:10:44 +04:00
// ps :
// ( ~num == num ^ 0x11111111 )
// we init the result segments with all bit set,
// then we can calculate through the way of 'xor'.
// ahead segments bin format : 1111 ... 1111
for i := uint32(0); i < maxDstSeq; i++ {
segments[i] = fillSegment
}
2014-07-08 13:10:44 +04:00
// last segment bin format : 1111..1100..0000
var tailSeg = make([]byte, segByteSize, segByteSize)
var fillByte = fillBits[7]
2014-07-08 13:10:44 +04:00
var tailSegLen = db.bCapByteSize(uint32(0), maxDstOff)
for i := uint32(0); i < tailSegLen-1; i++ {
tailSeg[i] = fillByte
2014-06-30 06:50:42 +04:00
}
2014-07-08 13:10:44 +04:00
tailSeg[tailSegLen-1] = fillBits[maxDstOff-(tailSegLen-1)<<3]
segments[maxDstSeq] = tailSeg
} else {
2014-07-08 13:10:44 +04:00
// ps : init segments by data corresponding to the 1st valid source key
it := db.bIterator(srckeys[srcIdx])
for ; it.Valid(); it.Next() {
if _, seq, err = db.bDecodeBinKey(it.Key()); err != nil {
// to do ...
it.Close()
return
}
segments[seq] = it.Value()
}
it.Close()
srcIdx++
2014-06-30 06:50:42 +04:00
}
// operation with following keys
var res []byte
for i := srcIdx; i < keyNum; i++ {
2014-07-08 13:10:44 +04:00
if srckeys[i] == nil {
2014-06-30 06:50:42 +04:00
continue
}
it := db.bIterator(srckeys[i])
for idx, end := uint32(0), false; !end; it.Next() {
2014-06-30 06:50:42 +04:00
end = !it.Valid()
if !end {
if _, seq, err = db.bDecodeBinKey(it.Key()); err != nil {
// to do ...
it.Close()
return
}
} else {
2014-07-08 13:10:44 +04:00
seq = maxDstSeq + 1
2014-06-30 06:50:42 +04:00
}
// todo :
// operation 'and' can be optimize here :
// if seq > max_segments_idx, this loop can be break,
2014-07-08 13:10:44 +04:00
// which can avoid cost from Key() and bDecodeBinKey()
for ; idx < seq; idx++ {
res = nil
exeOp(segments[idx], nil, &res)
segments[idx] = res
}
2014-06-30 06:50:42 +04:00
if !end {
res = it.Value()
exeOp(segments[seq], res, &res)
segments[seq] = res
idx++
2014-06-30 06:50:42 +04:00
}
}
it.Close()
}
// clear the old data in case
db.bDelete(t, dstkey)
db.rmExpire(t, bExpType, dstkey)
2014-07-08 13:10:44 +04:00
// set data
2014-06-30 06:50:42 +04:00
db.bSetMeta(t, dstkey, maxDstSeq, maxDstOff)
var bk []byte
2014-07-08 13:10:44 +04:00
for seq, segt := range segments {
if segt != nil {
2014-06-30 06:50:42 +04:00
bk = db.bEncodeBinKey(dstkey, uint32(seq))
2014-07-08 13:10:44 +04:00
t.Put(bk, segt)
2014-06-30 06:50:42 +04:00
}
}
err = t.Commit()
2014-07-08 13:10:44 +04:00
if err == nil {
2014-06-30 06:50:42 +04:00
blen = int32(maxDstSeq<<segBitWidth | maxDstOff)
}
return
}
2014-07-03 07:38:52 +04:00
func (db *DB) BExpire(key []byte, duration int64) (int64, error) {
if duration <= 0 {
return 0, errExpireValue
}
2014-06-30 06:50:42 +04:00
2014-07-03 07:38:52 +04:00
if err := checkKeySize(key); err != nil {
return -1, err
}
2014-06-30 06:50:42 +04:00
2014-07-03 07:38:52 +04:00
return db.bExpireAt(key, time.Now().Unix()+duration)
}
2014-06-30 06:50:42 +04:00
2014-07-03 07:38:52 +04:00
func (db *DB) BExpireAt(key []byte, when int64) (int64, error) {
if when <= time.Now().Unix() {
return 0, errExpireValue
}
if err := checkKeySize(key); err != nil {
return -1, err
}
2014-06-30 06:50:42 +04:00
2014-07-03 07:38:52 +04:00
return db.bExpireAt(key, when)
}
2014-06-30 06:50:42 +04:00
2014-07-03 07:38:52 +04:00
func (db *DB) BTTL(key []byte) (int64, error) {
if err := checkKeySize(key); err != nil {
return -1, err
}
2014-06-30 06:50:42 +04:00
2014-07-03 07:38:52 +04:00
return db.ttl(bExpType, key)
}
2014-06-30 06:50:42 +04:00
2014-07-03 07:38:52 +04:00
func (db *DB) BPersist(key []byte) (int64, error) {
if err := checkKeySize(key); err != nil {
return 0, err
}
t := db.binTx
t.Lock()
defer t.Unlock()
n, err := db.rmExpire(t, bExpType, key)
if err != nil {
return 0, err
}
err = t.Commit()
return n, err
}
2014-06-30 06:50:42 +04:00
// func (db *DB) BScan(key []byte, count int, inclusive bool) ([]KVPair, error) {
// }
2014-07-03 07:38:52 +04:00
func (db *DB) bFlush() (drop int64, err error) {
t := db.binTx
t.Lock()
defer t.Unlock()
minKey := make([]byte, 2)
minKey[0] = db.index
minKey[1] = binType
2014-06-30 06:50:42 +04:00
2014-07-03 07:38:52 +04:00
maxKey := make([]byte, 2)
maxKey[0] = db.index
maxKey[1] = binMetaType + 1
drop, err = db.flushRegion(t, minKey, maxKey)
err = db.expFlush(t, bExpType)
err = t.Commit()
return
}