18 references to _messages
System.Threading.Tasks.Dataflow (18)
Blocks\BatchBlock.cs (18)
389
_messages
.Enqueue(messageValue!);
395
(_batchesCompleted + (
_messages
.Count / _batchSize)) >= _dataflowBlockOptions.ActualMaxNumberOfGroups)
444
if (dropPendingMessages)
_messages
.Clear();
509
bool noMoreMessages = _decliningPermanently &&
_messages
.Count < _batchSize;
520
if (
_messages
.Count > 0) MakeBatchIfPossible(evenIfFewerThanBatchSize: true);
574
int neededMessageCountToCompleteBatch = _batchSize -
_messages
.Count;
583
if (_nonGreedyState.AcceptFewerThanBatchSize &&
_messages
.Count > 0)
649
_messages
.Count + _nonGreedyState.PostponedMessages.Count);
735
bool fullBatch =
_messages
.Count >= _batchSize;
738
if (fullBatch || (evenIfFewerThanBatchSize &&
_messages
.Count > 0))
740
var newBatch = new T[fullBatch ? _batchSize :
_messages
.Count];
741
for (int i = 0; i < newBatch.Length; i++) newBatch[i] =
_messages
.Dequeue();
773
Debug.Assert(
_messages
.Count == 0, "The queue must be empty between batches in non-greedy mode");
893
itemCountNeededToCompleteBatch = _batchSize -
_messages
.Count;
1021
_messages
.Enqueue(sourceAndMessage.Value.Value);
1071
if (sourceAndMessage.Key != null)
_messages
.Enqueue(sourceAndMessage.Value.Value);
1153
private int InputCountForDebugger { get { return
_messages
.Count; } }
1180
public IEnumerable<T> InputQueue { get { return _target.
_messages
.ToList(); } }