18 references to _messages
System.Threading.Tasks.Dataflow (18)
Blocks\BatchBlock.cs (18)
386
_messages
.Enqueue(messageValue!);
392
(_batchesCompleted + (
_messages
.Count / _batchSize)) >= _dataflowBlockOptions.ActualMaxNumberOfGroups)
441
if (dropPendingMessages)
_messages
.Clear();
506
bool noMoreMessages = _decliningPermanently &&
_messages
.Count < _batchSize;
517
if (
_messages
.Count > 0) MakeBatchIfPossible(evenIfFewerThanBatchSize: true);
571
int neededMessageCountToCompleteBatch = _batchSize -
_messages
.Count;
580
if (_nonGreedyState.AcceptFewerThanBatchSize &&
_messages
.Count > 0)
646
_messages
.Count + _nonGreedyState.PostponedMessages.Count);
732
bool fullBatch =
_messages
.Count >= _batchSize;
735
if (fullBatch || (evenIfFewerThanBatchSize &&
_messages
.Count > 0))
737
var newBatch = new T[fullBatch ? _batchSize :
_messages
.Count];
738
for (int i = 0; i < newBatch.Length; i++) newBatch[i] =
_messages
.Dequeue();
770
Debug.Assert(
_messages
.Count == 0, "The queue must be empty between batches in non-greedy mode");
890
itemCountNeededToCompleteBatch = _batchSize -
_messages
.Count;
1018
_messages
.Enqueue(sourceAndMessage.Value.Value);
1068
if (sourceAndMessage.Key != null)
_messages
.Enqueue(sourceAndMessage.Value.Value);
1150
private int InputCountForDebugger { get { return
_messages
.Count; } }
1177
public IEnumerable<T> InputQueue { get { return _target.
_messages
.ToList(); } }