Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

blockchain, utreexobackends: move nodesmapslice and cachedleavesmapslice to utreexobackends #168

Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
189 changes: 189 additions & 0 deletions blockchain/internal/utreexobackends/cachedleavesmap.go
Original file line number Diff line number Diff line change
@@ -0,0 +1,189 @@
package utreexobackends

import (
"sync"

"github.com/utreexo/utreexo"
"github.com/utreexo/utreexod/blockchain/internal/sizehelper"
"github.com/utreexo/utreexod/chaincfg/chainhash"
)

const (
// Bucket size for the cached leaves map.
cachedLeavesMapBucketSize = 16 + sizehelper.Uint64Size*chainhash.HashSize + sizehelper.Uint64Size*sizehelper.Uint64Size
)

// CachedLeavesMapSlice is a slice of maps for utxo entries. The slice of maps are needed to
// guarantee that the map will only take up N amount of bytes. As of v1.20, the
// go runtime will allocate 2^N + few extra buckets, meaning that for large N, we'll
// allocate a lot of extra memory if the amount of entries goes over the previously
// allocated buckets. A slice of maps allows us to have a better control of how much
// total memory gets allocated by all the maps.
type CachedLeavesMapSlice struct {
// mtx protects against concurrent access for the map slice.
mtx *sync.Mutex

// maps are the underlying maps in the slice of maps.
maps []map[utreexo.Hash]uint64

// maxEntries is the maximum amount of elemnts that the map is allocated for.
maxEntries []int

// maxTotalMemoryUsage is the maximum memory usage in bytes that the state
// should contain in normal circumstances.
maxTotalMemoryUsage uint64
}

// Length returns the length of all the maps in the map slice added together.
//
// This function is safe for concurrent access.
func (ms *CachedLeavesMapSlice) Length() int {
ms.mtx.Lock()
defer ms.mtx.Unlock()

var l int
for _, m := range ms.maps {
l += len(m)
}

return l
}

// Get looks for the outpoint in all the maps in the map slice and returns
// the entry. nil and false is returned if the outpoint is not found.
//
// This function is safe for concurrent access.
func (ms *CachedLeavesMapSlice) Get(k utreexo.Hash) (uint64, bool) {
ms.mtx.Lock()
defer ms.mtx.Unlock()

var v uint64
var found bool

for _, m := range ms.maps {
v, found = m[k]
if found {
return v, found
}
}

return 0, false
}

// Put puts the keys and the values into one of the maps in the map slice. If the
// existing maps are all full and it fails to put the entry in the cache, it will
// return false.
//
// This function is safe for concurrent access.
func (ms *CachedLeavesMapSlice) Put(k utreexo.Hash, v uint64) bool {
ms.mtx.Lock()
defer ms.mtx.Unlock()

for i := range ms.maxEntries {
m := ms.maps[i]
_, found := m[k]
if found {
m[k] = v
return true
}
}

for i, maxNum := range ms.maxEntries {
m := ms.maps[i]
if len(m) >= maxNum {
// Don't try to insert if the map already at max since
// that'll force the map to allocate double the memory it's
// currently taking up.
continue
}

m[k] = v
return true // Return as we were successful in adding the entry.
}

// We only reach this code if we've failed to insert into the map above as
// all the current maps were full.
return false
}

// Delete attempts to delete the given outpoint in all of the maps. No-op if the
// outpoint doesn't exist.
//
// This function is safe for concurrent access.
func (ms *CachedLeavesMapSlice) Delete(k utreexo.Hash) {
ms.mtx.Lock()
defer ms.mtx.Unlock()

for i := 0; i < len(ms.maps); i++ {
delete(ms.maps[i], k)
}
}

// DeleteMaps deletes all maps and allocate new ones with the maxEntries defined in
// ms.maxEntries.
//
// This function is safe for concurrent access.
func (ms *CachedLeavesMapSlice) DeleteMaps() {
ms.mtx.Lock()
defer ms.mtx.Unlock()

ms.maps = make([]map[utreexo.Hash]uint64, len(ms.maxEntries))
for i := range ms.maxEntries {
ms.maps[i] = make(map[utreexo.Hash]uint64, ms.maxEntries[i])
}
}

// ForEach loops through all the elements in the cachedleaves map slice and calls fn with the key-value pairs.
//
// This function is safe for concurrent access.
func (ms *CachedLeavesMapSlice) ForEach(fn func(utreexo.Hash, uint64)) {
ms.mtx.Lock()
defer ms.mtx.Unlock()

for _, m := range ms.maps {
for k, v := range m {
fn(k, v)
}
}
}

// createMaps creates a slice of maps and returns the total count that the maps
// can handle. maxEntries are also set along with the newly created maps.
func (ms *CachedLeavesMapSlice) createMaps(maxMemoryUsage int64) int64 {
ms.mtx.Lock()
defer ms.mtx.Unlock()

if maxMemoryUsage <= 0 {
return 0
}

// Get the entry count for the maps we'll allocate.
var totalElemCount int
ms.maxEntries, totalElemCount = sizehelper.CalcNumEntries(cachedLeavesMapBucketSize, maxMemoryUsage)

// maxMemoryUsage that's smaller than the minimum map size will return a totalElemCount
// that's equal to 0.
if totalElemCount <= 0 {
return 0
}

// Create the maps.
ms.maps = make([]map[utreexo.Hash]uint64, len(ms.maxEntries))
for i := range ms.maxEntries {
ms.maps[i] = make(map[utreexo.Hash]uint64, ms.maxEntries[i])
}

return int64(totalElemCount)
}

// NewCachedLeavesMapSlice returns a new CachedLeavesMapSlice and the total amount of elements
// that the map slice can accomodate.
func NewCachedLeavesMapSlice(maxTotalMemoryUsage int64) (CachedLeavesMapSlice, int64) {
ms := CachedLeavesMapSlice{
mtx: new(sync.Mutex),
maxTotalMemoryUsage: uint64(maxTotalMemoryUsage),
}

totalCacheElem := ms.createMaps(maxTotalMemoryUsage)
return ms, totalCacheElem
}
78 changes: 78 additions & 0 deletions blockchain/internal/utreexobackends/cachedleavesmap_test.go
Original file line number Diff line number Diff line change
@@ -0,0 +1,78 @@
package utreexobackends

import (
"crypto/sha256"
"encoding/binary"
"testing"

"github.com/utreexo/utreexo"
)

func TestCachedLeavesMapSliceMaxCacheElems(t *testing.T) {
_, maxCacheElems := NewCachedLeavesMapSlice(0)
if maxCacheElems != 0 {
t.Fatalf("expected %v got %v", 0, maxCacheElems)
}

_, maxCacheElems = NewCachedLeavesMapSlice(-1)
if maxCacheElems != 0 {
t.Fatalf("expected %v got %v", 0, maxCacheElems)
}

_, maxCacheElems = NewCachedLeavesMapSlice(8000)
if maxCacheElems <= 0 {
t.Fatalf("expected something bigger than 0 but got %v", maxCacheElems)
}
}

func uint64ToHash(v uint64) utreexo.Hash {
var buf [8]byte
binary.LittleEndian.PutUint64(buf[:], v)
return sha256.Sum256(buf[:])
}

func TestCachedLeaveMapSliceDuplicates(t *testing.T) {
m, maxElems := NewCachedLeavesMapSlice(8000)
for i := 0; i < 10; i++ {
for j := int64(0); j < maxElems; j++ {
if !m.Put(uint64ToHash(uint64(j)), 0) {
t.Fatalf("unexpected error on m.put")
}
}
}

if m.Length() != int(maxElems) {
t.Fatalf("expected length of %v but got %v",
maxElems, m.Length())
}

// Try inserting x which should be unique. Should fail as the map is full.
x := uint64(0)
x -= 1
if m.Put(uint64ToHash(x), 0) {
t.Fatalf("expected error but successfully called put")
}

// Remove the first element in the first map and then try inserting
// a duplicate element.
m.Delete(uint64ToHash(0))
x = uint64(maxElems) - 1
if !m.Put(uint64ToHash(x), 0) {
t.Fatalf("unexpected failure on put")
}

// Make sure the length of the map is 1 less than the max elems.
if m.Length() != int(maxElems)-1 {
t.Fatalf("expected length of %v but got %v",
maxElems-1, m.Length())
}

// Put 0 back in and then compare the map.
if !m.Put(uint64ToHash(0), 0) {
t.Fatalf("didn't expect error but unsuccessfully called put")
}
if m.Length() != int(maxElems) {
t.Fatalf("expected length of %v but got %v",
maxElems, m.Length())
}
}
Loading
Loading