aboutsummaryrefslogtreecommitdiffstats
path: root/swarm/storage/localstore_test.go
diff options
context:
space:
mode:
Diffstat (limited to 'swarm/storage/localstore_test.go')
-rw-r--r--swarm/storage/localstore_test.go244
1 files changed, 0 insertions, 244 deletions
diff --git a/swarm/storage/localstore_test.go b/swarm/storage/localstore_test.go
deleted file mode 100644
index fcadcefa0..000000000
--- a/swarm/storage/localstore_test.go
+++ /dev/null
@@ -1,244 +0,0 @@
-// Copyright 2018 The go-ethereum Authors
-// This file is part of the go-ethereum library.
-//
-// The go-ethereum library is free software: you can redistribute it and/or modify
-// it under the terms of the GNU Lesser General Public License as published by
-// the Free Software Foundation, either version 3 of the License, or
-// (at your option) any later version.
-//
-// The go-ethereum library is distributed in the hope that it will be useful,
-// but WITHOUT ANY WARRANTY; without even the implied warranty of
-// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-// GNU Lesser General Public License for more details.
-//
-// You should have received a copy of the GNU Lesser General Public License
-// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
-
-package storage
-
-import (
- "context"
- "io/ioutil"
- "os"
- "testing"
- "time"
-
- "github.com/ethereum/go-ethereum/swarm/chunk"
-)
-
-var (
- hashfunc = MakeHashFunc(DefaultHash)
-)
-
-// tests that the content address validator correctly checks the data
-// tests that feed update chunks are passed through content address validator
-// the test checking the resouce update validator internal correctness is found in storage/feeds/handler_test.go
-func TestValidator(t *testing.T) {
- // set up localstore
- datadir, err := ioutil.TempDir("", "storage-testvalidator")
- if err != nil {
- t.Fatal(err)
- }
- defer os.RemoveAll(datadir)
-
- params := NewDefaultLocalStoreParams()
- params.Init(datadir)
- store, err := NewLocalStore(params, nil)
- if err != nil {
- t.Fatal(err)
- }
-
- // check puts with no validators, both succeed
- chunks := GenerateRandomChunks(259, 2)
- goodChunk := chunks[0]
- badChunk := chunks[1]
- copy(badChunk.Data(), goodChunk.Data())
-
- errs := putChunks(store, goodChunk, badChunk)
- if errs[0] != nil {
- t.Fatalf("expected no error on good content address chunk in spite of no validation, but got: %s", err)
- }
- if errs[1] != nil {
- t.Fatalf("expected no error on bad content address chunk in spite of no validation, but got: %s", err)
- }
-
- // add content address validator and check puts
- // bad should fail, good should pass
- store.Validators = append(store.Validators, NewContentAddressValidator(hashfunc))
- chunks = GenerateRandomChunks(chunk.DefaultSize, 2)
- goodChunk = chunks[0]
- badChunk = chunks[1]
- copy(badChunk.Data(), goodChunk.Data())
-
- errs = putChunks(store, goodChunk, badChunk)
- if errs[0] != nil {
- t.Fatalf("expected no error on good content address chunk with content address validator only, but got: %s", err)
- }
- if errs[1] == nil {
- t.Fatal("expected error on bad content address chunk with content address validator only, but got nil")
- }
-
- // append a validator that always denies
- // bad should fail, good should pass,
- var negV boolTestValidator
- store.Validators = append(store.Validators, negV)
-
- chunks = GenerateRandomChunks(chunk.DefaultSize, 2)
- goodChunk = chunks[0]
- badChunk = chunks[1]
- copy(badChunk.Data(), goodChunk.Data())
-
- errs = putChunks(store, goodChunk, badChunk)
- if errs[0] != nil {
- t.Fatalf("expected no error on good content address chunk with content address validator only, but got: %s", err)
- }
- if errs[1] == nil {
- t.Fatal("expected error on bad content address chunk with content address validator only, but got nil")
- }
-
- // append a validator that always approves
- // all shall pass
- var posV boolTestValidator = true
- store.Validators = append(store.Validators, posV)
-
- chunks = GenerateRandomChunks(chunk.DefaultSize, 2)
- goodChunk = chunks[0]
- badChunk = chunks[1]
- copy(badChunk.Data(), goodChunk.Data())
-
- errs = putChunks(store, goodChunk, badChunk)
- if errs[0] != nil {
- t.Fatalf("expected no error on good content address chunk with content address validator only, but got: %s", err)
- }
- if errs[1] != nil {
- t.Fatalf("expected no error on bad content address chunk in spite of no validation, but got: %s", err)
- }
-
-}
-
-type boolTestValidator bool
-
-func (self boolTestValidator) Validate(chunk Chunk) bool {
- return bool(self)
-}
-
-// putChunks adds chunks to localstore
-// It waits for receive on the stored channel
-// It logs but does not fail on delivery error
-func putChunks(store *LocalStore, chunks ...Chunk) []error {
- i := 0
- f := func(n int64) Chunk {
- chunk := chunks[i]
- i++
- return chunk
- }
- _, errs := put(store, len(chunks), f)
- return errs
-}
-
-func put(store *LocalStore, n int, f func(i int64) Chunk) (hs []Address, errs []error) {
- for i := int64(0); i < int64(n); i++ {
- chunk := f(chunk.DefaultSize)
- err := store.Put(context.TODO(), chunk)
- errs = append(errs, err)
- hs = append(hs, chunk.Address())
- }
- return hs, errs
-}
-
-// TestGetFrequentlyAccessedChunkWontGetGarbageCollected tests that the most
-// frequently accessed chunk is not garbage collected from LDBStore, i.e.,
-// from disk when we are at the capacity and garbage collector runs. For that
-// we start putting random chunks into the DB while continuously accessing the
-// chunk we care about then check if we can still retrieve it from disk.
-func TestGetFrequentlyAccessedChunkWontGetGarbageCollected(t *testing.T) {
- ldbCap := defaultGCRatio
- store, cleanup := setupLocalStore(t, ldbCap)
- defer cleanup()
-
- var chunks []Chunk
- for i := 0; i < ldbCap; i++ {
- chunks = append(chunks, GenerateRandomChunk(chunk.DefaultSize))
- }
-
- mostAccessed := chunks[0].Address()
- for _, chunk := range chunks {
- if err := store.Put(context.Background(), chunk); err != nil {
- t.Fatal(err)
- }
-
- if _, err := store.Get(context.Background(), mostAccessed); err != nil {
- t.Fatal(err)
- }
- // Add time for MarkAccessed() to be able to finish in a separate Goroutine
- time.Sleep(1 * time.Millisecond)
- }
-
- store.DbStore.collectGarbage()
- if _, err := store.DbStore.Get(context.Background(), mostAccessed); err != nil {
- t.Logf("most frequntly accessed chunk not found on disk (key: %v)", mostAccessed)
- t.Fatal(err)
- }
-
-}
-
-func setupLocalStore(t *testing.T, ldbCap int) (ls *LocalStore, cleanup func()) {
- t.Helper()
-
- var err error
- datadir, err := ioutil.TempDir("", "storage")
- if err != nil {
- t.Fatal(err)
- }
-
- params := &LocalStoreParams{
- StoreParams: NewStoreParams(uint64(ldbCap), uint(ldbCap), nil, nil),
- }
- params.Init(datadir)
-
- store, err := NewLocalStore(params, nil)
- if err != nil {
- _ = os.RemoveAll(datadir)
- t.Fatal(err)
- }
-
- cleanup = func() {
- store.Close()
- _ = os.RemoveAll(datadir)
- }
-
- return store, cleanup
-}
-
-func TestHas(t *testing.T) {
- ldbCap := defaultGCRatio
- store, cleanup := setupLocalStore(t, ldbCap)
- defer cleanup()
-
- nonStoredAddr := GenerateRandomChunk(128).Address()
-
- has := store.Has(context.Background(), nonStoredAddr)
- if has {
- t.Fatal("Expected Has() to return false, but returned true!")
- }
-
- storeChunks := GenerateRandomChunks(128, 3)
- for _, ch := range storeChunks {
- err := store.Put(context.Background(), ch)
- if err != nil {
- t.Fatalf("Expected store to store chunk, but it failed: %v", err)
- }
-
- has := store.Has(context.Background(), ch.Address())
- if !has {
- t.Fatal("Expected Has() to return true, but returned false!")
- }
- }
-
- //let's be paranoic and test again that the non-existent chunk returns false
- has = store.Has(context.Background(), nonStoredAddr)
- if has {
- t.Fatal("Expected Has() to return false, but returned true!")
- }
-
-}