Skip to content
GitLab
Projects
Groups
Snippets
Help
Loading...
Help
What's new
10
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
Open sidebar
dms3
go-bitswap
Commits
2d3b3517
Commit
2d3b3517
authored
Sep 20, 2016
by
Jeromy
Browse files
Options
Browse Files
Download
Plain Diff
Merge branch 'version/0.4.3-rc4'
License: MIT Signed-off-by:
Jeromy
<
why@ipfs.io
>
parents
95b982a4
5575419b
Changes
11
Show whitespace changes
Inline
Side-by-side
Showing
11 changed files
with
194 additions
and
83 deletions
+194
-83
bitswap.go
bitswap.go
+43
-7
bitswap_test.go
bitswap_test.go
+73
-5
decision/bench_test.go
decision/bench_test.go
+1
-1
decision/engine.go
decision/engine.go
+2
-2
decision/ledger.go
decision/ledger.go
+1
-1
decision/peer_request_queue.go
decision/peer_request_queue.go
+3
-3
decision/peer_request_queue_test.go
decision/peer_request_queue_test.go
+5
-5
message/message.go
message/message.go
+2
-2
wantlist/wantlist.go
wantlist/wantlist.go
+29
-33
wantmanager.go
wantmanager.go
+13
-13
workers.go
workers.go
+22
-11
No files found.
bitswap.go
View file @
2d3b3517
...
...
@@ -17,7 +17,6 @@ import (
bsmsg
"github.com/ipfs/go-ipfs/exchange/bitswap/message"
bsnet
"github.com/ipfs/go-ipfs/exchange/bitswap/network"
notifications
"github.com/ipfs/go-ipfs/exchange/bitswap/notifications"
wantlist
"github.com/ipfs/go-ipfs/exchange/bitswap/wantlist"
flags
"github.com/ipfs/go-ipfs/flags"
"github.com/ipfs/go-ipfs/thirdparty/delay"
loggables
"gx/ipfs/QmYrv4LgCC8FhG2Ab4bwuq5DqBdwMtx3hMb3KKJDZcr2d7/go-libp2p-loggables"
...
...
@@ -89,7 +88,7 @@ func New(parent context.Context, p peer.ID, network bsnet.BitSwapNetwork,
notifications
:
notif
,
engine
:
decision
.
NewEngine
(
ctx
,
bstore
),
// TODO close the engine with Close() method
network
:
network
,
findKeys
:
make
(
chan
*
wantlist
.
Entry
,
sizeBatchRequestChan
),
findKeys
:
make
(
chan
*
blockRequest
,
sizeBatchRequestChan
),
process
:
px
,
newBlocks
:
make
(
chan
blocks
.
Block
,
HasBlockBufferSize
),
provideKeys
:
make
(
chan
key
.
Key
,
provideKeysBufferSize
),
...
...
@@ -132,7 +131,7 @@ type Bitswap struct {
notifications
notifications
.
PubSub
// send keys to a worker to find and connect to providers for them
findKeys
chan
*
wantlist
.
Entry
findKeys
chan
*
blockRequest
engine
*
decision
.
Engine
...
...
@@ -149,8 +148,8 @@ type Bitswap struct {
}
type
blockRequest
struct
{
k
ey
key
.
Key
c
tx
context
.
Context
K
ey
key
.
Key
C
tx
context
.
Context
}
// GetBlock attempts to retrieve a particular block from peers within the
...
...
@@ -240,13 +239,50 @@ func (bs *Bitswap) GetBlocks(ctx context.Context, keys []key.Key) (<-chan blocks
// NB: Optimization. Assumes that providers of key[0] are likely to
// be able to provide for all keys. This currently holds true in most
// every situation. Later, this assumption may not hold as true.
req
:=
&
wantlist
.
Entry
{
req
:=
&
blockRequest
{
Key
:
keys
[
0
],
Ctx
:
ctx
,
}
remaining
:=
make
(
map
[
key
.
Key
]
struct
{})
for
_
,
k
:=
range
keys
{
remaining
[
k
]
=
struct
{}{}
}
out
:=
make
(
chan
blocks
.
Block
)
go
func
()
{
ctx
,
cancel
:=
context
.
WithCancel
(
ctx
)
defer
cancel
()
defer
close
(
out
)
defer
func
()
{
var
toCancel
[]
key
.
Key
for
k
,
_
:=
range
remaining
{
toCancel
=
append
(
toCancel
,
k
)
}
bs
.
CancelWants
(
toCancel
)
}()
for
{
select
{
case
blk
,
ok
:=
<-
promise
:
if
!
ok
{
return
}
delete
(
remaining
,
blk
.
Key
())
select
{
case
out
<-
blk
:
case
<-
ctx
.
Done
()
:
return
}
case
<-
ctx
.
Done
()
:
return
}
}
}()
select
{
case
bs
.
findKeys
<-
req
:
return
promise
,
nil
return
out
,
nil
case
<-
ctx
.
Done
()
:
return
nil
,
ctx
.
Err
()
}
...
...
bitswap_test.go
View file @
2d3b3517
...
...
@@ -341,7 +341,6 @@ func TestDoubleGet(t *testing.T) {
blocks
:=
bg
.
Blocks
(
1
)
ctx1
,
cancel1
:=
context
.
WithCancel
(
context
.
Background
())
blkch1
,
err
:=
instances
[
1
]
.
Exchange
.
GetBlocks
(
ctx1
,
[]
key
.
Key
{
blocks
[
0
]
.
Key
()})
if
err
!=
nil
{
t
.
Fatal
(
err
)
...
...
@@ -369,11 +368,15 @@ func TestDoubleGet(t *testing.T) {
t
.
Fatal
(
err
)
}
blk
,
ok
:=
<-
blkch2
select
{
case
blk
,
ok
:=
<-
blkch2
:
if
!
ok
{
t
.
Fatal
(
"expected to get the block here"
)
}
t
.
Log
(
blk
)
case
<-
time
.
After
(
time
.
Second
*
5
)
:
t
.
Fatal
(
"timed out waiting on block"
)
}
for
_
,
inst
:=
range
instances
{
err
:=
inst
.
Exchange
.
Close
()
...
...
@@ -382,3 +385,68 @@ func TestDoubleGet(t *testing.T) {
}
}
}
func
TestWantlistCleanup
(
t
*
testing
.
T
)
{
net
:=
tn
.
VirtualNetwork
(
mockrouting
.
NewServer
(),
delay
.
Fixed
(
kNetworkDelay
))
sg
:=
NewTestSessionGenerator
(
net
)
defer
sg
.
Close
()
bg
:=
blocksutil
.
NewBlockGenerator
()
instances
:=
sg
.
Instances
(
1
)[
0
]
bswap
:=
instances
.
Exchange
blocks
:=
bg
.
Blocks
(
20
)
var
keys
[]
key
.
Key
for
_
,
b
:=
range
blocks
{
keys
=
append
(
keys
,
b
.
Key
())
}
ctx
,
cancel
:=
context
.
WithTimeout
(
context
.
Background
(),
time
.
Millisecond
*
50
)
defer
cancel
()
_
,
err
:=
bswap
.
GetBlock
(
ctx
,
keys
[
0
])
if
err
!=
context
.
DeadlineExceeded
{
t
.
Fatal
(
"shouldnt have fetched any blocks"
)
}
time
.
Sleep
(
time
.
Millisecond
*
50
)
if
len
(
bswap
.
GetWantlist
())
>
0
{
t
.
Fatal
(
"should not have anyting in wantlist"
)
}
ctx
,
cancel
=
context
.
WithTimeout
(
context
.
Background
(),
time
.
Millisecond
*
50
)
defer
cancel
()
_
,
err
=
bswap
.
GetBlocks
(
ctx
,
keys
[
:
10
])
if
err
!=
nil
{
t
.
Fatal
(
err
)
}
<-
ctx
.
Done
()
time
.
Sleep
(
time
.
Millisecond
*
50
)
if
len
(
bswap
.
GetWantlist
())
>
0
{
t
.
Fatal
(
"should not have anyting in wantlist"
)
}
_
,
err
=
bswap
.
GetBlocks
(
context
.
Background
(),
keys
[
:
1
])
if
err
!=
nil
{
t
.
Fatal
(
err
)
}
ctx
,
cancel
=
context
.
WithCancel
(
context
.
Background
())
_
,
err
=
bswap
.
GetBlocks
(
ctx
,
keys
[
10
:
])
if
err
!=
nil
{
t
.
Fatal
(
err
)
}
time
.
Sleep
(
time
.
Millisecond
*
50
)
if
len
(
bswap
.
GetWantlist
())
!=
11
{
t
.
Fatal
(
"should have 11 keys in wantlist"
)
}
cancel
()
time
.
Sleep
(
time
.
Millisecond
*
50
)
if
!
(
len
(
bswap
.
GetWantlist
())
==
1
&&
bswap
.
GetWantlist
()[
0
]
==
keys
[
0
])
{
t
.
Fatal
(
"should only have keys[0] in wantlist"
)
}
}
decision/bench_test.go
View file @
2d3b3517
...
...
@@ -21,6 +21,6 @@ func BenchmarkTaskQueuePush(b *testing.B) {
}
b
.
ResetTimer
()
for
i
:=
0
;
i
<
b
.
N
;
i
++
{
q
.
Push
(
wantlist
.
Entry
{
Key
:
key
.
Key
(
i
),
Priority
:
math
.
MaxInt32
},
peers
[
i
%
len
(
peers
)])
q
.
Push
(
&
wantlist
.
Entry
{
Key
:
key
.
Key
(
i
),
Priority
:
math
.
MaxInt32
},
peers
[
i
%
len
(
peers
)])
}
}
decision/engine.go
View file @
2d3b3517
...
...
@@ -104,7 +104,7 @@ func NewEngine(ctx context.Context, bs bstore.Blockstore) *Engine {
return
e
}
func
(
e
*
Engine
)
WantlistForPeer
(
p
peer
.
ID
)
(
out
[]
wl
.
Entry
)
{
func
(
e
*
Engine
)
WantlistForPeer
(
p
peer
.
ID
)
(
out
[]
*
wl
.
Entry
)
{
e
.
lock
.
Lock
()
partner
,
ok
:=
e
.
ledgerMap
[
p
]
if
ok
{
...
...
@@ -233,7 +233,7 @@ func (e *Engine) MessageReceived(p peer.ID, m bsmsg.BitSwapMessage) error {
for
_
,
entry
:=
range
m
.
Wantlist
()
{
if
entry
.
Cancel
{
log
.
Debugf
(
"cancel %s"
,
entry
.
Key
)
log
.
Debugf
(
"
%s
cancel %s"
,
p
,
entry
.
Key
)
l
.
CancelWant
(
entry
.
Key
)
e
.
peerRequestQueue
.
Remove
(
entry
.
Key
,
p
)
}
else
{
...
...
decision/ledger.go
View file @
2d3b3517
...
...
@@ -87,7 +87,7 @@ func (l *ledger) CancelWant(k key.Key) {
l
.
wantList
.
Remove
(
k
)
}
func
(
l
*
ledger
)
WantListContains
(
k
key
.
Key
)
(
wl
.
Entry
,
bool
)
{
func
(
l
*
ledger
)
WantListContains
(
k
key
.
Key
)
(
*
wl
.
Entry
,
bool
)
{
return
l
.
wantList
.
Contains
(
k
)
}
...
...
decision/peer_request_queue.go
View file @
2d3b3517
...
...
@@ -13,7 +13,7 @@ import (
type
peerRequestQueue
interface
{
// Pop returns the next peerRequestTask. Returns nil if the peerRequestQueue is empty.
Pop
()
*
peerRequestTask
Push
(
entry
wantlist
.
Entry
,
to
peer
.
ID
)
Push
(
entry
*
wantlist
.
Entry
,
to
peer
.
ID
)
Remove
(
k
key
.
Key
,
p
peer
.
ID
)
// NB: cannot expose simply expose taskQueue.Len because trashed elements
...
...
@@ -45,7 +45,7 @@ type prq struct {
}
// Push currently adds a new peerRequestTask to the end of the list
func
(
tl
*
prq
)
Push
(
entry
wantlist
.
Entry
,
to
peer
.
ID
)
{
func
(
tl
*
prq
)
Push
(
entry
*
wantlist
.
Entry
,
to
peer
.
ID
)
{
tl
.
lock
.
Lock
()
defer
tl
.
lock
.
Unlock
()
partner
,
ok
:=
tl
.
partners
[
to
]
...
...
@@ -166,7 +166,7 @@ func (tl *prq) thawRound() {
}
type
peerRequestTask
struct
{
Entry
wantlist
.
Entry
Entry
*
wantlist
.
Entry
Target
peer
.
ID
// A callback to signal that this task has been completed
...
...
decision/peer_request_queue_test.go
View file @
2d3b3517
...
...
@@ -41,7 +41,7 @@ func TestPushPop(t *testing.T) {
for
_
,
index
:=
range
rand
.
Perm
(
len
(
alphabet
))
{
// add blocks for all letters
letter
:=
alphabet
[
index
]
t
.
Log
(
partner
.
String
())
prq
.
Push
(
wantlist
.
Entry
{
Key
:
key
.
Key
(
letter
),
Priority
:
math
.
MaxInt32
-
index
},
partner
)
prq
.
Push
(
&
wantlist
.
Entry
{
Key
:
key
.
Key
(
letter
),
Priority
:
math
.
MaxInt32
-
index
},
partner
)
}
for
_
,
consonant
:=
range
consonants
{
prq
.
Remove
(
key
.
Key
(
consonant
),
partner
)
...
...
@@ -78,10 +78,10 @@ func TestPeerRepeats(t *testing.T) {
// Have each push some blocks
for
i
:=
0
;
i
<
5
;
i
++
{
prq
.
Push
(
wantlist
.
Entry
{
Key
:
key
.
Key
(
i
)},
a
)
prq
.
Push
(
wantlist
.
Entry
{
Key
:
key
.
Key
(
i
)},
b
)
prq
.
Push
(
wantlist
.
Entry
{
Key
:
key
.
Key
(
i
)},
c
)
prq
.
Push
(
wantlist
.
Entry
{
Key
:
key
.
Key
(
i
)},
d
)
prq
.
Push
(
&
wantlist
.
Entry
{
Key
:
key
.
Key
(
i
)},
a
)
prq
.
Push
(
&
wantlist
.
Entry
{
Key
:
key
.
Key
(
i
)},
b
)
prq
.
Push
(
&
wantlist
.
Entry
{
Key
:
key
.
Key
(
i
)},
c
)
prq
.
Push
(
&
wantlist
.
Entry
{
Key
:
key
.
Key
(
i
)},
d
)
}
// now, pop off four entries, there should be one from each
...
...
message/message.go
View file @
2d3b3517
...
...
@@ -64,7 +64,7 @@ func newMsg(full bool) *impl {
}
type
Entry
struct
{
wantlist
.
Entry
*
wantlist
.
Entry
Cancel
bool
}
...
...
@@ -120,7 +120,7 @@ func (m *impl) addEntry(k key.Key, priority int, cancel bool) {
e
.
Cancel
=
cancel
}
else
{
m
.
wantlist
[
k
]
=
Entry
{
Entry
:
wantlist
.
Entry
{
Entry
:
&
wantlist
.
Entry
{
Key
:
k
,
Priority
:
priority
,
},
...
...
wantlist/wantlist.go
View file @
2d3b3517
...
...
@@ -7,8 +7,6 @@ import (
"sync"
key
"gx/ipfs/Qmce4Y4zg3sYr7xKM5UueS67vhNni6EeWgCRnb7MbLJMew/go-key"
"gx/ipfs/QmZy2y8t9zQH2a1b8q2ZSLKp17ATuJoCNxxyMFG5qFExpt/go-net/context"
)
type
ThreadSafe
struct
{
...
...
@@ -18,19 +16,17 @@ type ThreadSafe struct {
// not threadsafe
type
Wantlist
struct
{
set
map
[
key
.
Key
]
Entry
set
map
[
key
.
Key
]
*
Entry
}
type
Entry
struct
{
Key
key
.
Key
Priority
int
Ctx
context
.
Context
cancel
func
()
RefCnt
int
}
type
entrySlice
[]
Entry
type
entrySlice
[]
*
Entry
func
(
es
entrySlice
)
Len
()
int
{
return
len
(
es
)
}
func
(
es
entrySlice
)
Swap
(
i
,
j
int
)
{
es
[
i
],
es
[
j
]
=
es
[
j
],
es
[
i
]
}
...
...
@@ -44,41 +40,41 @@ func NewThreadSafe() *ThreadSafe {
func
New
()
*
Wantlist
{
return
&
Wantlist
{
set
:
make
(
map
[
key
.
Key
]
Entry
),
set
:
make
(
map
[
key
.
Key
]
*
Entry
),
}
}
func
(
w
*
ThreadSafe
)
Add
(
k
key
.
Key
,
priority
int
)
{
func
(
w
*
ThreadSafe
)
Add
(
k
key
.
Key
,
priority
int
)
bool
{
w
.
lk
.
Lock
()
defer
w
.
lk
.
Unlock
()
w
.
Wantlist
.
Add
(
k
,
priority
)
return
w
.
Wantlist
.
Add
(
k
,
priority
)
}
func
(
w
*
ThreadSafe
)
AddEntry
(
e
Entry
)
{
func
(
w
*
ThreadSafe
)
AddEntry
(
e
*
Entry
)
bool
{
w
.
lk
.
Lock
()
defer
w
.
lk
.
Unlock
()
w
.
Wantlist
.
AddEntry
(
e
)
return
w
.
Wantlist
.
AddEntry
(
e
)
}
func
(
w
*
ThreadSafe
)
Remove
(
k
key
.
Key
)
{
func
(
w
*
ThreadSafe
)
Remove
(
k
key
.
Key
)
bool
{
w
.
lk
.
Lock
()
defer
w
.
lk
.
Unlock
()
w
.
Wantlist
.
Remove
(
k
)
return
w
.
Wantlist
.
Remove
(
k
)
}
func
(
w
*
ThreadSafe
)
Contains
(
k
key
.
Key
)
(
Entry
,
bool
)
{
func
(
w
*
ThreadSafe
)
Contains
(
k
key
.
Key
)
(
*
Entry
,
bool
)
{
w
.
lk
.
RLock
()
defer
w
.
lk
.
RUnlock
()
return
w
.
Wantlist
.
Contains
(
k
)
}
func
(
w
*
ThreadSafe
)
Entries
()
[]
Entry
{
func
(
w
*
ThreadSafe
)
Entries
()
[]
*
Entry
{
w
.
lk
.
RLock
()
defer
w
.
lk
.
RUnlock
()
return
w
.
Wantlist
.
Entries
()
}
func
(
w
*
ThreadSafe
)
SortedEntries
()
[]
Entry
{
func
(
w
*
ThreadSafe
)
SortedEntries
()
[]
*
Entry
{
w
.
lk
.
RLock
()
defer
w
.
lk
.
RUnlock
()
return
w
.
Wantlist
.
SortedEntries
()
...
...
@@ -94,50 +90,50 @@ func (w *Wantlist) Len() int {
return
len
(
w
.
set
)
}
func
(
w
*
Wantlist
)
Add
(
k
key
.
Key
,
priority
int
)
{
func
(
w
*
Wantlist
)
Add
(
k
key
.
Key
,
priority
int
)
bool
{
if
e
,
ok
:=
w
.
set
[
k
];
ok
{
e
.
RefCnt
++
return
return
false
}
ctx
,
cancel
:=
context
.
WithCancel
(
context
.
Background
())
w
.
set
[
k
]
=
Entry
{
w
.
set
[
k
]
=
&
Entry
{
Key
:
k
,
Priority
:
priority
,
Ctx
:
ctx
,
cancel
:
cancel
,
RefCnt
:
1
,
}
return
true
}
func
(
w
*
Wantlist
)
AddEntry
(
e
Entry
)
{
if
_
,
ok
:=
w
.
set
[
e
.
Key
];
ok
{
return
func
(
w
*
Wantlist
)
AddEntry
(
e
*
Entry
)
bool
{
if
ex
,
ok
:=
w
.
set
[
e
.
Key
];
ok
{
ex
.
RefCnt
++
return
false
}
w
.
set
[
e
.
Key
]
=
e
return
true
}
func
(
w
*
Wantlist
)
Remove
(
k
key
.
Key
)
{
func
(
w
*
Wantlist
)
Remove
(
k
key
.
Key
)
bool
{
e
,
ok
:=
w
.
set
[
k
]
if
!
ok
{
return
return
false
}
e
.
RefCnt
--
if
e
.
RefCnt
<=
0
{
delete
(
w
.
set
,
k
)
if
e
.
cancel
!=
nil
{
e
.
cancel
()
}
return
true
}
return
false
}
func
(
w
*
Wantlist
)
Contains
(
k
key
.
Key
)
(
Entry
,
bool
)
{
func
(
w
*
Wantlist
)
Contains
(
k
key
.
Key
)
(
*
Entry
,
bool
)
{
e
,
ok
:=
w
.
set
[
k
]
return
e
,
ok
}
func
(
w
*
Wantlist
)
Entries
()
[]
Entry
{
func
(
w
*
Wantlist
)
Entries
()
[]
*
Entry
{
var
es
entrySlice
for
_
,
e
:=
range
w
.
set
{
es
=
append
(
es
,
e
)
...
...
@@ -145,7 +141,7 @@ func (w *Wantlist) Entries() []Entry {
return
es
}
func
(
w
*
Wantlist
)
SortedEntries
()
[]
Entry
{
func
(
w
*
Wantlist
)
SortedEntries
()
[]
*
Entry
{
var
es
entrySlice
for
_
,
e
:=
range
w
.
set
{
es
=
append
(
es
,
e
)
...
...
wantmanager.go
View file @
2d3b3517
...
...
@@ -75,6 +75,7 @@ func (pm *WantManager) WantBlocks(ctx context.Context, ks []key.Key) {
}
func
(
pm
*
WantManager
)
CancelWants
(
ks
[]
key
.
Key
)
{
log
.
Infof
(
"cancel wants: %s"
,
ks
)
pm
.
addEntries
(
context
.
TODO
(),
ks
,
true
)
}
...
...
@@ -83,16 +84,17 @@ func (pm *WantManager) addEntries(ctx context.Context, ks []key.Key, cancel bool
for
i
,
k
:=
range
ks
{
entries
=
append
(
entries
,
&
bsmsg
.
Entry
{
Cancel
:
cancel
,
Entry
:
wantlist
.
Entry
{
Entry
:
&
wantlist
.
Entry
{
Key
:
k
,
Priority
:
kMaxPriority
-
i
,
Ctx
:
ctx
,
RefCnt
:
1
,
},
})
}
select
{
case
pm
.
incoming
<-
entries
:
case
<-
pm
.
ctx
.
Done
()
:
case
<-
ctx
.
Done
()
:
}
}
...
...
@@ -241,33 +243,31 @@ func (pm *WantManager) Run() {
case
entries
:=
<-
pm
.
incoming
:
// add changes to our wantlist
var
filtered
[]
*
bsmsg
.
Entry
for
_
,
e
:=
range
entries
{
if
e
.
Cancel
{
pm
.
wl
.
Remove
(
e
.
Key
)
if
pm
.
wl
.
Remove
(
e
.
Key
)
{
filtered
=
append
(
filtered
,
e
)
}
}
else
{
pm
.
wl
.
AddEntry
(
e
.
Entry
)
if
pm
.
wl
.
AddEntry
(
e
.
Entry
)
{
filtered
=
append
(
filtered
,
e
)
}
}
}
// broadcast those wantlist changes
for
_
,
p
:=
range
pm
.
peers
{
p
.
addMessage
(
entries
)
p
.
addMessage
(
filtered
)
}
case
<-
tock
.
C
:
// resend entire wantlist every so often (REALLY SHOULDNT BE NECESSARY)
var
es
[]
*
bsmsg
.
Entry
for
_
,
e
:=
range
pm
.
wl
.
Entries
()
{
select
{
case
<-
e
.
Ctx
.
Done
()
:
// entry has been cancelled
// simply continue, the entry will be removed from the
// wantlist soon enough
continue
default
:
}
es
=
append
(
es
,
&
bsmsg
.
Entry
{
Entry
:
e
})
}
for
_
,
p
:=
range
pm
.
peers
{
p
.
outlk
.
Lock
()
p
.
out
=
bsmsg
.
New
(
true
)
...
...
workers.go
View file @
2d3b3517
package
bitswap
import
(
"math/rand"
"sync"
"time"
process
"gx/ipfs/QmSF8fPo3jgVBAy8fpdjjYqgG87dkJgUprRBHRd2tmfgpP/goprocess"
procctx
"gx/ipfs/QmSF8fPo3jgVBAy8fpdjjYqgG87dkJgUprRBHRd2tmfgpP/goprocess/context"
context
"gx/ipfs/QmZy2y8t9zQH2a1b8q2ZSLKp17ATuJoCNxxyMFG5qFExpt/go-net/context"
wantlist
"github.com/ipfs/go-ipfs/exchange/bitswap/wantlist"
logging
"gx/ipfs/QmSpJByNKFX1sCsHBEp3R73FL4NF6FnQTEGyNAXHm2GS52/go-log"
peer
"gx/ipfs/QmWXjJo15p4pzT7cayEwZi2sWgJqLnGDof6ZGMh9xBgU1p/go-libp2p-peer"
context
"gx/ipfs/QmZy2y8t9zQH2a1b8q2ZSLKp17ATuJoCNxxyMFG5qFExpt/go-net/context"
key
"gx/ipfs/Qmce4Y4zg3sYr7xKM5UueS67vhNni6EeWgCRnb7MbLJMew/go-key"
)
...
...
@@ -173,9 +172,17 @@ func (bs *Bitswap) rebroadcastWorker(parent context.Context) {
}
case
<-
broadcastSignal
.
C
:
// resend unfulfilled wantlist keys
log
.
Event
(
ctx
,
"Bitswap.Rebroadcast.active"
)
for
_
,
e
:=
range
bs
.
wm
.
wl
.
Entries
()
{
e
:=
e
bs
.
findKeys
<-
&
e
entries
:=
bs
.
wm
.
wl
.
Entries
()
if
len
(
entries
)
==
0
{
continue
}
// TODO: come up with a better strategy for determining when to search
// for new providers for blocks.
i
:=
rand
.
Intn
(
len
(
entries
))
bs
.
findKeys
<-
&
blockRequest
{
Key
:
entries
[
i
]
.
Key
,
Ctx
:
ctx
,
}
case
<-
parent
.
Done
()
:
return
...
...
@@ -185,33 +192,37 @@ func (bs *Bitswap) rebroadcastWorker(parent context.Context) {
func
(
bs
*
Bitswap
)
providerQueryManager
(
ctx
context
.
Context
)
{
var
activeLk
sync
.
Mutex
active
:=
make
(
map
[
key
.
Key
]
*
wantlist
.
Entry
)
kset
:=
key
.
NewKeySet
(
)
for
{
select
{
case
e
:=
<-
bs
.
findKeys
:
activeLk
.
Lock
()
if
_
,
ok
:=
active
[
e
.
Key
];
ok
{
if
kset
.
Has
(
e
.
Key
)
{
activeLk
.
Unlock
()
continue
}
active
[
e
.
Key
]
=
e
kset
.
Add
(
e
.
Key
)
activeLk
.
Unlock
()
go
func
(
e
*
wantlist
.
Entry
)
{
go
func
(
e
*
blockRequest
)
{
child
,
cancel
:=
context
.
WithTimeout
(
e
.
Ctx
,
providerRequestTimeout
)
defer
cancel
()
providers
:=
bs
.
network
.
FindProvidersAsync
(
child
,
e
.
Key
,
maxProvidersPerRequest
)
wg
:=
&
sync
.
WaitGroup
{}
for
p
:=
range
providers
{
wg
.
Add
(
1
)
go
func
(
p
peer
.
ID
)
{
defer
wg
.
Done
()
err
:=
bs
.
network
.
ConnectTo
(
child
,
p
)
if
err
!=
nil
{
log
.
Debug
(
"failed to connect to provider %s: %s"
,
p
,
err
)
}
}(
p
)
}
wg
.
Wait
()
activeLk
.
Lock
()
delete
(
active
,
e
.
Key
)
kset
.
Remove
(
e
.
Key
)
activeLk
.
Unlock
()
}(
e
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment