18 references to _messages
System.Threading.Tasks.Dataflow (18)
Blocks\BatchBlock.cs (18)
386_messages.Enqueue(messageValue!); 392(_batchesCompleted + (_messages.Count / _batchSize)) >= _dataflowBlockOptions.ActualMaxNumberOfGroups) 441if (dropPendingMessages) _messages.Clear(); 506bool noMoreMessages = _decliningPermanently && _messages.Count < _batchSize; 517if (_messages.Count > 0) MakeBatchIfPossible(evenIfFewerThanBatchSize: true); 571int neededMessageCountToCompleteBatch = _batchSize - _messages.Count; 580if (_nonGreedyState.AcceptFewerThanBatchSize && _messages.Count > 0) 646_messages.Count + _nonGreedyState.PostponedMessages.Count); 732bool fullBatch = _messages.Count >= _batchSize; 735if (fullBatch || (evenIfFewerThanBatchSize && _messages.Count > 0)) 737var newBatch = new T[fullBatch ? _batchSize : _messages.Count]; 738for (int i = 0; i < newBatch.Length; i++) newBatch[i] = _messages.Dequeue(); 770Debug.Assert(_messages.Count == 0, "The queue must be empty between batches in non-greedy mode"); 890itemCountNeededToCompleteBatch = _batchSize - _messages.Count; 1018_messages.Enqueue(sourceAndMessage.Value.Value); 1068if (sourceAndMessage.Key != null) _messages.Enqueue(sourceAndMessage.Value.Value); 1150private int InputCountForDebugger { get { return _messages.Count; } } 1177public IEnumerable<T> InputQueue { get { return _target._messages.ToList(); } }