From 9e9fc87e70accf2b81be8772ab2ab0c914e95666 Mon Sep 17 00:00:00 2001 From: Anton Evangelatov Date: Sun, 23 Dec 2018 17:31:32 +0100 Subject: swarm: remove unused/dead code (#18351) --- swarm/storage/chunker.go | 17 +-------------- swarm/storage/database.go | 10 --------- swarm/storage/error.go | 12 ++--------- swarm/storage/ldbstore.go | 20 ++--------------- swarm/storage/ldbstore_test.go | 10 +-------- swarm/storage/mock/mock.go | 7 ------ swarm/storage/pyramid.go | 5 ----- swarm/storage/types.go | 49 ------------------------------------------ 8 files changed, 6 insertions(+), 124 deletions(-) (limited to 'swarm/storage') diff --git a/swarm/storage/chunker.go b/swarm/storage/chunker.go index cbe65372a..a8bfe2d1c 100644 --- a/swarm/storage/chunker.go +++ b/swarm/storage/chunker.go @@ -65,10 +65,6 @@ If all is well it is possible to implement this by simply composing readers so t The hashing itself does use extra copies and allocation though, since it does need it. */ -var ( - errAppendOppNotSuported = errors.New("Append operation not supported") -) - type ChunkerParams struct { chunkSize int64 hashSize int64 @@ -99,7 +95,6 @@ type TreeChunker struct { ctx context.Context branches int64 - hashFunc SwarmHasher dataSize int64 data io.Reader // calculated @@ -365,10 +360,6 @@ func (tc *TreeChunker) runWorker(ctx context.Context) { }() } -func (tc *TreeChunker) Append() (Address, func(), error) { - return nil, nil, errAppendOppNotSuported -} - // LazyChunkReader implements LazySectionReader type LazyChunkReader struct { ctx context.Context @@ -411,7 +402,6 @@ func (r *LazyChunkReader) Size(ctx context.Context, quitC chan bool) (n int64, e log.Debug("lazychunkreader.size", "addr", r.addr) if r.chunkData == nil { - startTime := time.Now() chunkData, err := r.getter.Get(cctx, Reference(r.addr)) if err != nil { @@ -420,13 +410,8 @@ func (r *LazyChunkReader) Size(ctx context.Context, quitC chan bool) (n int64, e } metrics.GetOrRegisterResettingTimer("lcr.getter.get", nil).UpdateSince(startTime) r.chunkData = chunkData - s := r.chunkData.Size() - log.Debug("lazychunkreader.size", "key", r.addr, "size", s) - if s < 0 { - return 0, errors.New("corrupt size") - } - return int64(s), nil } + s := r.chunkData.Size() log.Debug("lazychunkreader.size", "key", r.addr, "size", s) diff --git a/swarm/storage/database.go b/swarm/storage/database.go index e25fce31f..12367b905 100644 --- a/swarm/storage/database.go +++ b/swarm/storage/database.go @@ -64,16 +64,6 @@ func (db *LDBDatabase) Delete(key []byte) error { return db.db.Delete(key, nil) } -func (db *LDBDatabase) LastKnownTD() []byte { - data, _ := db.Get([]byte("LTD")) - - if len(data) == 0 { - data = []byte{0x0} - } - - return data -} - func (db *LDBDatabase) NewIterator() iterator.Iterator { metrics.GetOrRegisterCounter("ldbdatabase.newiterator", nil).Inc(1) diff --git a/swarm/storage/error.go b/swarm/storage/error.go index 44261c084..a9d0616fa 100644 --- a/swarm/storage/error.go +++ b/swarm/storage/error.go @@ -23,23 +23,15 @@ import ( const ( ErrInit = iota ErrNotFound - ErrIO ErrUnauthorized ErrInvalidValue ErrDataOverflow ErrNothingToReturn - ErrCorruptData ErrInvalidSignature ErrNotSynced - ErrPeriodDepth - ErrCnt ) var ( - ErrChunkNotFound = errors.New("chunk not found") - ErrFetching = errors.New("chunk still fetching") - ErrChunkInvalid = errors.New("invalid chunk") - ErrChunkForward = errors.New("cannot forward") - ErrChunkUnavailable = errors.New("chunk unavailable") - ErrChunkTimeout = errors.New("timeout") + ErrChunkNotFound = errors.New("chunk not found") + ErrChunkInvalid = errors.New("invalid chunk") ) diff --git a/swarm/storage/ldbstore.go b/swarm/storage/ldbstore.go index bd4f6b916..635d33429 100644 --- a/swarm/storage/ldbstore.go +++ b/swarm/storage/ldbstore.go @@ -248,10 +248,6 @@ func U64ToBytes(val uint64) []byte { return data } -func (s *LDBStore) updateIndexAccess(index *dpaDBIndex) { - index.Access = s.accessCnt -} - func getIndexKey(hash Address) []byte { hashSize := len(hash) key := make([]byte, hashSize+1) @@ -777,18 +773,6 @@ func (s *LDBStore) BinIndex(po uint8) uint64 { return s.bucketCnt[po] } -func (s *LDBStore) Size() uint64 { - s.lock.RLock() - defer s.lock.RUnlock() - return s.entryCnt -} - -func (s *LDBStore) CurrentStorageIndex() uint64 { - s.lock.RLock() - defer s.lock.RUnlock() - return s.dataIdx -} - // Put adds a chunk to the database, adding indices and incrementing global counters. // If it already exists, it merely increments the access count of the existing entry. // Is thread safe @@ -810,11 +794,11 @@ func (s *LDBStore) Put(ctx context.Context, chunk Chunk) error { batch := s.batch log.Trace("ldbstore.put: s.db.Get", "key", chunk.Address(), "ikey", fmt.Sprintf("%x", ikey)) - idata, err := s.db.Get(ikey) + _, err := s.db.Get(ikey) if err != nil { s.doPut(chunk, &index, po) } - idata = encodeIndex(&index) + idata := encodeIndex(&index) s.batch.Put(ikey, idata) // add the access-chunkindex index for garbage collection diff --git a/swarm/storage/ldbstore_test.go b/swarm/storage/ldbstore_test.go index e8b9ae39b..1fe466f93 100644 --- a/swarm/storage/ldbstore_test.go +++ b/swarm/storage/ldbstore_test.go @@ -79,14 +79,6 @@ func testPoFunc(k Address) (ret uint8) { return uint8(Proximity(basekey, k[:])) } -func (db *testDbStore) close() { - db.Close() - err := os.RemoveAll(db.dir) - if err != nil { - panic(err) - } -} - func testDbStoreRandom(n int, chunksize int64, mock bool, t *testing.T) { db, cleanup, err := newTestDbStore(mock, true) defer cleanup() @@ -453,7 +445,7 @@ func TestLDBStoreAddRemove(t *testing.T) { log.Info("ldbstore", "entrycnt", ldb.entryCnt, "accesscnt", ldb.accessCnt) for i := 0; i < n; i++ { - ret, err := ldb.Get(nil, chunks[i].Address()) + ret, err := ldb.Get(context.TODO(), chunks[i].Address()) if i%2 == 0 { // expect even chunks to be missing diff --git a/swarm/storage/mock/mock.go b/swarm/storage/mock/mock.go index 1fb71b70a..626ba3fe1 100644 --- a/swarm/storage/mock/mock.go +++ b/swarm/storage/mock/mock.go @@ -103,13 +103,6 @@ type Exporter interface { Export(w io.Writer) (n int, err error) } -// ImportExporter is an interface for importing and exporting -// mock store data to and from a tar archive. -type ImportExporter interface { - Importer - Exporter -} - // ExportedChunk is the structure that is saved in tar archive for // each chunk as JSON-encoded bytes. type ExportedChunk struct { diff --git a/swarm/storage/pyramid.go b/swarm/storage/pyramid.go index f74eef06b..e5bd7a76a 100644 --- a/swarm/storage/pyramid.go +++ b/swarm/storage/pyramid.go @@ -71,11 +71,6 @@ const ( splitTimeout = time.Minute * 5 ) -const ( - DataChunk = 0 - TreeChunk = 1 -) - type PyramidSplitterParams struct { SplitterParams getter Getter diff --git a/swarm/storage/types.go b/swarm/storage/types.go index 322d95c47..8bfb34815 100644 --- a/swarm/storage/types.go +++ b/swarm/storage/types.go @@ -23,7 +23,6 @@ import ( "crypto/rand" "encoding/binary" "fmt" - "hash" "io" "github.com/ethereum/go-ethereum/common" @@ -35,50 +34,10 @@ import ( const MaxPO = 16 const AddressLength = 32 -type Hasher func() hash.Hash type SwarmHasher func() SwarmHash -// Peer is the recorded as Source on the chunk -// should probably not be here? but network should wrap chunk object -type Peer interface{} - type Address []byte -func (a Address) Size() uint { - return uint(len(a)) -} - -func (a Address) isEqual(y Address) bool { - return bytes.Equal(a, y) -} - -func (a Address) bits(i, j uint) uint { - ii := i >> 3 - jj := i & 7 - if ii >= a.Size() { - return 0 - } - - if jj+j <= 8 { - return uint((a[ii] >> jj) & ((1 << j) - 1)) - } - - res := uint(a[ii] >> jj) - jj = 8 - jj - j -= jj - for j != 0 { - ii++ - if j < 8 { - res += uint(a[ii]&((1<