protected override async Task Dispatch(PeerInfo peerInfo, BodiesSyncBatch batch, CancellationToken cancellationToken) { ISyncPeer peer = peerInfo.SyncPeer; batch.ResponseSourcePeer = peerInfo; batch.MarkSent(); var hashes = batch.Infos.Where(i => i != null).Select(i => i !.BlockHash).ToArray(); Task <BlockBody[]> getBodiesTask = peer.GetBlockBodies(hashes, cancellationToken); await getBodiesTask.ContinueWith( (t, state) => { BodiesSyncBatch batchLocal = (BodiesSyncBatch)state !; if (t.IsCompletedSuccessfully) { if (batchLocal.RequestTime > 1000) { if (Logger.IsDebug) { Logger.Debug($"{batchLocal} - peer is slow {batchLocal.RequestTime:F2}"); } } batchLocal.Response = t.Result; } }, batch); }
protected override async Task Dispatch(PeerInfo peerInfo, BodiesSyncBatch batch, CancellationToken cancellationToken) { ISyncPeer peer = peerInfo.SyncPeer; batch.ResponseSourcePeer = peerInfo; batch.MarkSent(); Keccak[]? hashes = batch.Infos.Where(i => i != null).Select(i => i !.BlockHash).ToArray(); if (hashes.Length == 0) { if (Logger.IsDebug) { Logger.Debug($"{batch} - attempted send a request with no hash."); } return; } try { batch.Response = await peer.GetBlockBodies(hashes, cancellationToken); } catch (TimeoutException) { if (Logger.IsDebug) { Logger.Debug($"{batch} - Get block bodies timeout {batch.RequestTime:F2}"); } return; } if (batch.RequestTime > 1000) { if (Logger.IsDebug) { Logger.Debug($"{batch} - peer is slow {batch.RequestTime:F2}"); } } }
private async Task ExecuteRequest(CancellationToken token, FastBlocksBatch batch) { SyncPeerAllocation syncPeerAllocation = batch.Allocation; try { foreach (PeerInfo usefulPeer in _syncPeerPool.UsefulPeers) { if (usefulPeer.HeadNumber < Math.Max(0, (batch.MinNumber ?? 0) - 1024)) { if (_logger.IsDebug) { _logger.Debug($"Made {usefulPeer} sleep for a while - no min number satisfied"); } _syncPeerPool.ReportNoSyncProgress(usefulPeer); } } PeerInfo peerInfo = syncPeerAllocation?.Current; ISyncPeer peer = peerInfo?.SyncPeer; if (peer != null) { batch.MarkSent(); switch (batch.BatchType) { case FastBlocksBatchType.Headers: { Task <BlockHeader[]> getHeadersTask = peer.GetBlockHeaders(batch.Headers.StartNumber, batch.Headers.RequestSize, 0, token); await getHeadersTask.ContinueWith( t => { if (t.IsCompletedSuccessfully) { if (batch.RequestTime > 1000) { if (_logger.IsDebug) { _logger.Debug($"{batch} - peer is slow {batch.RequestTime:F2}"); } } batch.Headers.Response = getHeadersTask.Result; } else { if (t.Exception.InnerExceptions.Any(e => e is TimeoutException)) { _syncPeerPool.ReportInvalid(batch.Allocation, $"headers -> timeout"); } else { _syncPeerPool.ReportInvalid(batch.Allocation, $"headers -> {t.Exception}"); } } } ); break; } case FastBlocksBatchType.Bodies: { Task <BlockBody[]> getBodiesTask = peer.GetBlockBodies(batch.Bodies.Request, token); await getBodiesTask.ContinueWith( t => { if (t.IsCompletedSuccessfully) { if (batch.RequestTime > 1000) { if (_logger.IsDebug) { _logger.Debug($"{batch} - peer is slow {batch.RequestTime:F2}"); } } batch.Bodies.Response = getBodiesTask.Result; } else { if (t.Exception.InnerExceptions.Any(e => e is TimeoutException)) { _syncPeerPool.ReportInvalid(batch.Allocation, $"bodies -> timeout"); } else { _syncPeerPool.ReportInvalid(batch.Allocation, $"bodies -> {t.Exception}"); } } } ); break; } case FastBlocksBatchType.Receipts: { Task <TxReceipt[][]> getReceiptsTask = peer.GetReceipts(batch.Receipts.Request, token); await getReceiptsTask.ContinueWith( t => { if (t.IsCompletedSuccessfully) { if (batch.RequestTime > 1000) { if (_logger.IsDebug) { _logger.Debug($"{batch} - peer is slow {batch.RequestTime:F2}"); } } batch.Receipts.Response = getReceiptsTask.Result; } else { if (t.Exception.InnerExceptions.Any(e => e is TimeoutException)) { _syncPeerPool.ReportInvalid(batch.Allocation, $"receipts -> timeout"); } else { _syncPeerPool.ReportInvalid(batch.Allocation, $"receipts -> {t.Exception}"); } } } ); break; } default: { throw new InvalidOperationException($"{nameof(FastBlocksBatchType)} is {batch.BatchType}"); } } } (BlocksDataHandlerResult Result, int ItemsSynced)result = (BlocksDataHandlerResult.InvalidFormat, 0); try { result = _fastBlocksFeed.HandleResponse(batch); } catch (Exception e) { // possibly clear the response and handle empty response batch here (to avoid missing parts) // this practically corrupts sync if (_logger.IsError) { _logger.Error($"Error when handling response", e); } } Interlocked.Add(ref _downloadedHeaders, result.ItemsSynced); if (result.ItemsSynced == 0 && peer != null) { _syncPeerPool.ReportNoSyncProgress(peerInfo); } } finally { if (syncPeerAllocation != null) { _syncPeerPool.Free(syncPeerAllocation); } } }
private async Task ExecuteRequest(CancellationToken token, FastBlocksBatch batch) { SyncPeerAllocation nodeSyncAllocation = _syncPeerPool.Borrow(BorrowOptions.DoNotReplace | (batch.Prioritized ? BorrowOptions.None : BorrowOptions.LowPriority), "fast blocks", batch.MinNumber); foreach (PeerInfo peerInfo in _syncPeerPool.UsefulPeers) { if (peerInfo.HeadNumber < Math.Max(0, (batch.MinNumber ?? 0) - 1024)) { if (_logger.IsDebug) { _logger.Debug($"Made {peerInfo} sleep for a while - no min number satisfied"); } _syncPeerPool.ReportNoSyncProgress(peerInfo); } } try { ISyncPeer peer = nodeSyncAllocation?.Current?.SyncPeer; batch.Allocation = nodeSyncAllocation; if (peer != null) { batch.MarkSent(); switch (batch.BatchType) { case FastBlocksBatchType.Headers: { Task <BlockHeader[]> getHeadersTask = peer.GetBlockHeaders(batch.Headers.StartNumber, batch.Headers.RequestSize, 0, token); await getHeadersTask.ContinueWith( t => { if (t.IsCompletedSuccessfully) { if (batch.RequestTime > 1000) { if (_logger.IsDebug) { _logger.Debug($"{batch} - peer is slow {batch.RequestTime:F2}"); } } batch.Headers.Response = getHeadersTask.Result; ValidateHeaders(token, batch); } else { _syncPeerPool.ReportInvalid(batch.Allocation); } } ); break; } case FastBlocksBatchType.Bodies: { Task <BlockBody[]> getBodiesTask = peer.GetBlockBodies(batch.Bodies.Request, token); await getBodiesTask.ContinueWith( t => { if (t.IsCompletedSuccessfully) { if (batch.RequestTime > 1000) { if (_logger.IsDebug) { _logger.Debug($"{batch} - peer is slow {batch.RequestTime:F2}"); } } batch.Bodies.Response = getBodiesTask.Result; } else { _syncPeerPool.ReportInvalid(batch.Allocation); } } ); break; } case FastBlocksBatchType.Receipts: { Task <TxReceipt[][]> getReceiptsTask = peer.GetReceipts(batch.Receipts.Request, token); await getReceiptsTask.ContinueWith( t => { if (t.IsCompletedSuccessfully) { if (batch.RequestTime > 1000) { if (_logger.IsDebug) { _logger.Debug($"{batch} - peer is slow {batch.RequestTime:F2}"); } } batch.Receipts.Response = getReceiptsTask.Result; } else { _syncPeerPool.ReportInvalid(batch.Allocation); } } ); break; } default: { throw new InvalidOperationException($"{nameof(FastBlocksBatchType)} is {batch.BatchType}"); } } } (BlocksDataHandlerResult Result, int ItemsSynced)result = (BlocksDataHandlerResult.InvalidFormat, 0); try { if (batch.Bodies?.Response == null && batch.Headers?.Response == null && batch.Receipts?.Response == null) { // to avoid uncontrolled loop in case of a code error await Task.Delay(10); } result = _fastBlocksFeed.HandleResponse(batch); } catch (Exception e) { // possibly clear the response and handle empty response batch here (to avoid missing parts) if (_logger.IsError) { _logger.Error($"Error when handling response", e); } } Interlocked.Add(ref _downloadedHeaders, result.ItemsSynced); if (result.ItemsSynced == 0 && peer != null) { _syncPeerPool.ReportNoSyncProgress(nodeSyncAllocation); } } finally { if (nodeSyncAllocation != null) { _syncPeerPool.Free(nodeSyncAllocation); } } }