dht_bootstrap.go 6.25 KB
Newer Older
1
// Package dht implements a distributed hash table that satisfies the ipfs routing
2
// interface. This DHT is modeled after Kademlia with S/Kademlia modifications.
3 4 5
package dht

import (
Jeromy's avatar
Jeromy committed
6
	"context"
7 8 9 10
	"crypto/rand"
	"fmt"
	"time"

11 12 13
	u "github.com/ipfs/go-ipfs-util"
	goprocess "github.com/jbenet/goprocess"
	periodicproc "github.com/jbenet/goprocess/periodic"
14
	peer "github.com/libp2p/go-libp2p-peer"
George Antoniadis's avatar
George Antoniadis committed
15
	routing "github.com/libp2p/go-libp2p-routing"
16 17
)

18
// BootstrapConfig specifies parameters used bootstrapping the DHT.
19
//
20 21 22 23 24 25 26 27
// Note there is a tradeoff between the bootstrap period and the
// number of queries. We could support a higher period with less
// queries.
type BootstrapConfig struct {
	Queries int           // how many queries to run per period
	Period  time.Duration // how often to run periodi cbootstrap.
	Timeout time.Duration // how long to wait for a bootstrao query to run
}
28

29 30 31 32 33 34
var DefaultBootstrapConfig = BootstrapConfig{
	// For now, this is set to 1 query.
	// We are currently more interested in ensuring we have a properly formed
	// DHT than making sure our dht minimizes traffic. Once we are more certain
	// of our implementation's robustness, we should lower this down to 8 or 4.
	Queries: 1,
35

36
	// For now, this is set to 1 minute, which is a medium period. We are
37
	// We are currently more interested in ensuring we have a properly formed
38 39
	// DHT than making sure our dht minimizes traffic.
	Period: time.Duration(5 * time.Minute),
40

41
	Timeout: time.Duration(10 * time.Second),
42 43
}

44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66
// Bootstrap ensures the dht routing table remains healthy as peers come and go.
// it builds up a list of peers by requesting random peer IDs. The Bootstrap
// process will run a number of queries each time, and run every time signal fires.
// These parameters are configurable.
//
// As opposed to BootstrapWithConfig, Bootstrap satisfies the routing interface
func (dht *IpfsDHT) Bootstrap(ctx context.Context) error {
	proc, err := dht.BootstrapWithConfig(DefaultBootstrapConfig)
	if err != nil {
		return err
	}

	// wait till ctx or dht.Context exits.
	// we have to do it this way to satisfy the Routing interface (contexts)
	go func() {
		defer proc.Close()
		select {
		case <-ctx.Done():
		case <-dht.Context().Done():
		}
	}()

	return nil
67 68
}

69
// BootstrapWithConfig ensures the dht routing table remains healthy as peers come and go.
70 71 72 73
// it builds up a list of peers by requesting random peer IDs. The Bootstrap
// process will run a number of queries each time, and run every time signal fires.
// These parameters are configurable.
//
74
// BootstrapWithConfig returns a process, so the user can stop it.
vyzo's avatar
vyzo committed
75 76 77 78 79
func (dht *IpfsDHT) BootstrapWithConfig(cfg BootstrapConfig) (goprocess.Process, error) {
	if cfg.Queries <= 0 {
		return nil, fmt.Errorf("invalid number of queries: %d", cfg.Queries)
	}

80 81 82 83 84 85 86
	proc := dht.Process().Go(func(p goprocess.Process) {
		workerch := make(chan (<-chan struct{}))
		bootstrap := func() {
			proc := p.Go(dht.bootstrapWorker(cfg))
			workerch <- proc.Closed()
		}
		go bootstrap()
87 88 89
		for {
			select {
			case <-time.After(cfg.Period):
90
				ch := <-workerch
91
				select {
92 93 94 95
				case <-ch:
					go bootstrap()
				case <-p.Closing():
					return
96 97 98
				default:
					log.Warning("Previous bootstrapping attempt not completed within bootstrapping period")
				}
99
			case <-p.Closing():
100 101 102
				return
			}
		}
103
	})
vyzo's avatar
vyzo committed
104 105

	return proc, nil
106 107 108 109 110 111 112 113
}

// SignalBootstrap ensures the dht routing table remains healthy as peers come and go.
// it builds up a list of peers by requesting random peer IDs. The Bootstrap
// process will run a number of queries each time, and run every time signal fires.
// These parameters are configurable.
//
// SignalBootstrap returns a process, so the user can stop it.
114 115 116
func (dht *IpfsDHT) BootstrapOnSignal(cfg BootstrapConfig, signal <-chan time.Time) (goprocess.Process, error) {
	if cfg.Queries <= 0 {
		return nil, fmt.Errorf("invalid number of queries: %d", cfg.Queries)
117 118 119 120 121 122
	}

	if signal == nil {
		return nil, fmt.Errorf("invalid signal: %v", signal)
	}

123 124 125 126 127 128 129
	proc := periodicproc.Ticker(signal, dht.bootstrapWorker(cfg))

	return proc, nil
}

func (dht *IpfsDHT) bootstrapWorker(cfg BootstrapConfig) func(worker goprocess.Process) {
	return func(worker goprocess.Process) {
130 131 132 133
		// it would be useful to be able to send out signals of when we bootstrap, too...
		// maybe this is a good case for whole module event pub/sub?

		ctx := dht.Context()
134
		if err := dht.runBootstrap(ctx, cfg); err != nil {
Juan Batiz-Benet's avatar
Juan Batiz-Benet committed
135
			log.Warning(err)
136
			// A bootstrapping error is important to notice but not fatal.
137
		}
138
	}
139 140 141
}

// runBootstrap builds up list of peers by requesting random peer IDs
142
func (dht *IpfsDHT) runBootstrap(ctx context.Context, cfg BootstrapConfig) error {
Juan Batiz-Benet's avatar
Juan Batiz-Benet committed
143 144 145 146 147 148
	bslog := func(msg string) {
		log.Debugf("DHT %s dhtRunBootstrap %s -- routing table size: %d", dht.self, msg, dht.routingTable.Size())
	}
	bslog("start")
	defer bslog("end")
	defer log.EventBegin(ctx, "dhtRunBootstrap").Done()
149 150 151 152 153 154 155 156

	var merr u.MultiErr

	randomID := func() peer.ID {
		// 16 random bytes is not a valid peer id. it may be fine becuase
		// the dht will rehash to its own keyspace anyway.
		id := make([]byte, 16)
		rand.Read(id)
Juan Batiz-Benet's avatar
Juan Batiz-Benet committed
157
		id = u.Hash(id)
158 159 160 161 162
		return peer.ID(id)
	}

	// bootstrap sequentially, as results will compound
	runQuery := func(ctx context.Context, id peer.ID) {
163 164 165
		ctx, cancel := context.WithTimeout(ctx, cfg.Timeout)
		defer cancel()

166 167 168 169 170 171 172 173 174
		p, err := dht.FindPeer(ctx, id)
		if err == routing.ErrNotFound {
			// this isn't an error. this is precisely what we expect.
		} else if err != nil {
			merr = append(merr, err)
		} else {
			// woah, actually found a peer with that ID? this shouldn't happen normally
			// (as the ID we use is not a real ID). this is an odd error worth logging.
			err := fmt.Errorf("Bootstrap peer error: Actually FOUND peer. (%s, %s)", id, p)
Juan Batiz-Benet's avatar
Juan Batiz-Benet committed
175
			log.Warningf("%s", err)
176 177 178 179
			merr = append(merr, err)
		}
	}

180 181 182 183 184 185
	// these should be parallel normally. but can make them sequential for debugging.
	// note that the core/bootstrap context deadline should be extended too for that.
	for i := 0; i < cfg.Queries; i++ {
		id := randomID()
		log.Debugf("Bootstrapping query (%d/%d) to random ID: %s", i+1, cfg.Queries, id)
		runQuery(ctx, id)
186 187
	}

188 189 190 191 192
	// Find self to distribute peer info to our neighbors.
	// Do this after bootstrapping.
	log.Debugf("Bootstrapping query to self: %s", dht.self)
	runQuery(ctx, dht.self)

193 194 195 196 197
	if len(merr) > 0 {
		return merr
	}
	return nil
}