Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Set write batch readTs to the readMark.DoneUntil() value #778

Merged
merged 1 commit into from
May 2, 2019
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
12 changes: 10 additions & 2 deletions batch.go
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,14 @@ type WriteBatch struct {
// creating and committing transactions. Due to the nature of SSI guaratees provided by Badger,
// blind writes can never encounter transaction conflicts (ErrConflict).
func (db *DB) NewWriteBatch() *WriteBatch {
return &WriteBatch{db: db, txn: db.newTransaction(true, true)}
txn := db.newTransaction(true, true)
// If we let it stay at zero, compactions would not allow older key versions to be deleted,
// because the read timestamps of pending txns, would be zero. Therefore, we set it to the
// maximum read timestamp that's done. This allows compactions to discard key versions below
// this read timestamp, while also not blocking on pending txns to finish before starting this
// one.
txn.readTs = db.orc.readMark.DoneUntil()
return &WriteBatch{db: db, txn: txn}
}

// Cancel function must be called if there's a chance that Flush might not get
Expand Down Expand Up @@ -128,7 +135,8 @@ func (wb *WriteBatch) commit() error {
wb.wg.Add(1)
wb.txn.CommitWith(wb.callback)
wb.txn = wb.db.newTransaction(true, true)
wb.txn.readTs = 0 // We're not reading anything.
// See comment about readTs in NewWriteBatch.
wb.txn.readTs = wb.db.orc.readMark.DoneUntil()
return wb.err
}

Expand Down
58 changes: 58 additions & 0 deletions batch_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -18,6 +18,8 @@ package badger

import (
"fmt"
"io/ioutil"
"os"
"testing"
"time"

Expand Down Expand Up @@ -67,3 +69,59 @@ func TestWriteBatch(t *testing.T) {
require.NoError(t, err)
})
}

func TestWriteBatchCompaction(t *testing.T) {
dir, err := ioutil.TempDir(".", "badger-test")
require.NoError(t, err)
defer os.RemoveAll(dir)

opts := DefaultOptions
opts.ValueDir = dir
opts.Dir = dir

db, err := Open(opts)
require.NoError(t, err)

wb := db.NewWriteBatch()
entries := 10000
for i := 0; i < entries; i++ {
require.Nil(t, wb.Set([]byte(fmt.Sprintf("foo%d", i)), []byte("bar"), 0))
}
require.Nil(t, wb.Flush())

wb = db.NewWriteBatch()
// Delete 50% of the entries
for i := 0; i < entries/2; i++ {
require.Nil(t, wb.Delete([]byte(fmt.Sprintf("foo%d", i))))
}
require.Nil(t, wb.Flush())

// It is necessary that we call db.Update(..) before compaction so that the db.orc.readMark
// value is incremented. The transactions in WriteBatch call do not increment the
// db.orc.readMark value and hence compaction wouldn't discard any entries added by write batch
// if we do not increment the db.orc.readMark value
require.Nil(t, db.Update(func(txn *Txn) error {
txn.Set([]byte("key1"), []byte("val1"))
return nil
}))

// Close DB to force compaction
require.Nil(t, db.Close())

db, err = Open(opts)
require.NoError(t, err)
defer db.Close()

iopt := DefaultIteratorOptions
iopt.AllVersions = true
txn := db.NewTransaction(false)
defer txn.Discard()
it := txn.NewIterator(iopt)
defer it.Close()
countAfterCompaction := 0
for it.Rewind(); it.Valid(); it.Next() {
countAfterCompaction++
}
// We have deleted 50% of the keys
require.Less(t, countAfterCompaction, entries+entries/2)
}