@@ -27,49 +27,54 @@ var (
27
27
_ p2p.Wrapper = (* tmcons .Message )(nil )
28
28
)
29
29
30
- // GetChannelDescriptor produces an instance of a descriptor for this
31
- // package's required channels.
32
- func getChannelDescriptors () map [p2p.ChannelID ]* p2p.ChannelDescriptor {
33
- return map [p2p.ChannelID ]* p2p.ChannelDescriptor {
34
- StateChannel : {
35
- ID : StateChannel ,
36
- MessageType : new (tmcons.Message ),
37
- Priority : 8 ,
38
- SendQueueCapacity : 64 ,
39
- RecvMessageCapacity : maxMsgSize ,
40
- RecvBufferCapacity : 128 ,
41
- Name : "state" ,
42
- },
43
- DataChannel : {
44
- // TODO: Consider a split between gossiping current block and catchup
45
- // stuff. Once we gossip the whole block there is nothing left to send
46
- // until next height or round.
47
- ID : DataChannel ,
48
- MessageType : new (tmcons.Message ),
49
- Priority : 12 ,
50
- SendQueueCapacity : 64 ,
51
- RecvBufferCapacity : 512 ,
52
- RecvMessageCapacity : maxMsgSize ,
53
- Name : "data" ,
54
- },
55
- VoteChannel : {
56
- ID : VoteChannel ,
57
- MessageType : new (tmcons.Message ),
58
- Priority : 10 ,
59
- SendQueueCapacity : 64 ,
60
- RecvBufferCapacity : 128 ,
61
- RecvMessageCapacity : maxMsgSize ,
62
- Name : "vote" ,
63
- },
64
- VoteSetBitsChannel : {
65
- ID : VoteSetBitsChannel ,
66
- MessageType : new (tmcons.Message ),
67
- Priority : 5 ,
68
- SendQueueCapacity : 8 ,
69
- RecvBufferCapacity : 128 ,
70
- RecvMessageCapacity : maxMsgSize ,
71
- Name : "voteSet" ,
72
- },
30
+ func GetStateChannelDescriptor () * p2p.ChannelDescriptor {
31
+ return & p2p.ChannelDescriptor {
32
+ ID : StateChannel ,
33
+ MessageType : new (tmcons.Message ),
34
+ Priority : 8 ,
35
+ SendQueueCapacity : 64 ,
36
+ RecvMessageCapacity : maxMsgSize ,
37
+ RecvBufferCapacity : 128 ,
38
+ Name : "state" ,
39
+ }
40
+ }
41
+
42
+ func GetDataChannelDescriptor () * p2p.ChannelDescriptor {
43
+ return & p2p.ChannelDescriptor {
44
+ // TODO: Consider a split between gossiping current block and catchup
45
+ // stuff. Once we gossip the whole block there is nothing left to send
46
+ // until next height or round.
47
+ ID : DataChannel ,
48
+ MessageType : new (tmcons.Message ),
49
+ Priority : 12 ,
50
+ SendQueueCapacity : 64 ,
51
+ RecvBufferCapacity : 512 ,
52
+ RecvMessageCapacity : maxMsgSize ,
53
+ Name : "data" ,
54
+ }
55
+ }
56
+
57
+ func GetVoteChannelDescriptor () * p2p.ChannelDescriptor {
58
+ return & p2p.ChannelDescriptor {
59
+ ID : VoteChannel ,
60
+ MessageType : new (tmcons.Message ),
61
+ Priority : 10 ,
62
+ SendQueueCapacity : 64 ,
63
+ RecvBufferCapacity : 128 ,
64
+ RecvMessageCapacity : maxMsgSize ,
65
+ Name : "vote" ,
66
+ }
67
+ }
68
+
69
+ func GetVoteSetChannelDescriptor () * p2p.ChannelDescriptor {
70
+ return & p2p.ChannelDescriptor {
71
+ ID : VoteSetBitsChannel ,
72
+ MessageType : new (tmcons.Message ),
73
+ Priority : 5 ,
74
+ SendQueueCapacity : 8 ,
75
+ RecvBufferCapacity : 128 ,
76
+ RecvMessageCapacity : maxMsgSize ,
77
+ Name : "voteSet" ,
73
78
}
74
79
}
75
80
@@ -103,8 +108,9 @@ type BlockSyncReactor interface {
103
108
GetRemainingSyncTime () time.Duration
104
109
}
105
110
106
- //go:generate ../../scripts/mockery_generate.sh ConsSyncReactor
107
111
// ConsSyncReactor defines an interface used for testing abilities of node.startStateSync.
112
+ //
113
+ //go:generate ../../scripts/mockery_generate.sh ConsSyncReactor
108
114
type ConsSyncReactor interface {
109
115
SwitchToConsensus (sm.State , bool )
110
116
SetStateSyncingMetrics (float64 )
@@ -127,7 +133,8 @@ type Reactor struct {
127
133
readySignal chan struct {} // closed when the node is ready to start consensus
128
134
129
135
peerEvents p2p.PeerEventSubscriber
130
- chCreator p2p.ChannelCreator
136
+
137
+ channels * channelBundle
131
138
}
132
139
133
140
// NewReactor returns a reference to a new consensus reactor, which implements
@@ -137,7 +144,6 @@ type Reactor struct {
137
144
func NewReactor (
138
145
logger log.Logger ,
139
146
cs * State ,
140
- channelCreator p2p.ChannelCreator ,
141
147
peerEvents p2p.PeerEventSubscriber ,
142
148
eventBus * eventbus.EventBus ,
143
149
waitSync bool ,
@@ -152,8 +158,8 @@ func NewReactor(
152
158
eventBus : eventBus ,
153
159
Metrics : metrics ,
154
160
peerEvents : peerEvents ,
155
- chCreator : channelCreator ,
156
161
readySignal : make (chan struct {}),
162
+ channels : & channelBundle {},
157
163
}
158
164
r .BaseService = * service .NewBaseService (logger , "Consensus" , r )
159
165
@@ -171,6 +177,22 @@ type channelBundle struct {
171
177
votSet * p2p.Channel
172
178
}
173
179
180
+ func (r * Reactor ) SetStateChannel (ch * p2p.Channel ) {
181
+ r .channels .state = ch
182
+ }
183
+
184
+ func (r * Reactor ) SetDataChannel (ch * p2p.Channel ) {
185
+ r .channels .data = ch
186
+ }
187
+
188
+ func (r * Reactor ) SetVoteChannel (ch * p2p.Channel ) {
189
+ r .channels .vote = ch
190
+ }
191
+
192
+ func (r * Reactor ) SetVoteSetChannel (ch * p2p.Channel ) {
193
+ r .channels .votSet = ch
194
+ }
195
+
174
196
// OnStart starts separate go routines for each p2p Channel and listens for
175
197
// envelopes on each. In addition, it also listens for peer updates and handles
176
198
// messages on that p2p channel accordingly. The caller must be sure to execute
@@ -180,37 +202,13 @@ func (r *Reactor) OnStart(ctx context.Context) error {
180
202
181
203
peerUpdates := r .peerEvents (ctx )
182
204
183
- var chBundle channelBundle
184
- var err error
185
-
186
- chans := getChannelDescriptors ()
187
- chBundle .state , err = r .chCreator (ctx , chans [StateChannel ])
188
- if err != nil {
189
- return err
190
- }
191
-
192
- chBundle .data , err = r .chCreator (ctx , chans [DataChannel ])
193
- if err != nil {
194
- return err
195
- }
196
-
197
- chBundle .vote , err = r .chCreator (ctx , chans [VoteChannel ])
198
- if err != nil {
199
- return err
200
- }
201
-
202
- chBundle .votSet , err = r .chCreator (ctx , chans [VoteSetBitsChannel ])
203
- if err != nil {
204
- return err
205
- }
206
-
207
205
// start routine that computes peer statistics for evaluating peer quality
208
206
//
209
207
// TODO: Evaluate if we need this to be synchronized via WaitGroup as to not
210
208
// leak the goroutine when stopping the reactor.
211
209
go r .peerStatsRoutine (ctx , peerUpdates )
212
210
213
- r .subscribeToBroadcastEvents (ctx , chBundle .state )
211
+ r .subscribeToBroadcastEvents (ctx , r . channels .state )
214
212
215
213
if ! r .WaitSync () {
216
214
if err := r .state .Start (ctx ); err != nil {
@@ -222,11 +220,11 @@ func (r *Reactor) OnStart(ctx context.Context) error {
222
220
223
221
go r .updateRoundStateRoutine (ctx )
224
222
225
- go r .processStateCh (ctx , chBundle )
226
- go r .processDataCh (ctx , chBundle )
227
- go r .processVoteCh (ctx , chBundle )
228
- go r .processVoteSetBitsCh (ctx , chBundle )
229
- go r .processPeerUpdates (ctx , peerUpdates , chBundle )
223
+ go r .processStateCh (ctx , * r . channels )
224
+ go r .processDataCh (ctx , * r . channels )
225
+ go r .processVoteCh (ctx , * r . channels )
226
+ go r .processVoteSetBitsCh (ctx , * r . channels )
227
+ go r .processPeerUpdates (ctx , peerUpdates , * r . channels )
230
228
231
229
return nil
232
230
}
0 commit comments