1 write to _model
Microsoft.ML.Tokenizers (1)
Model\SentencePieceTokenizer.cs (1)
25_model = modelProto.TrainerSpec.ModelType switch
39 references to _model
Microsoft.ML.Tokenizers (39)
Model\SentencePieceTokenizer.cs (39)
36public IReadOnlyDictionary<string, int>? SpecialTokens => _model.SpecialTokens; 41public bool ByteFallback => _model.ByteFallback; 46public bool AddDummyPrefix => _model.AddDummyPrefix; 51public bool EscapeWhiteSpaces => _model.EscapeWhiteSpaces; 56public bool TreatWhitespaceAsSuffix { get => _model.TreatWhitespaceAsSuffix; private set => _model.TreatWhitespaceAsSuffix = value; } 61public bool AddBeginningOfSentence => _model.AddBeginningOfSentence; 66public bool AddEndOfSentence => _model.AddEndOfSentence; 71public string BeginningOfSentenceToken => _model.BeginningOfSentenceToken; 76public string EndOfSentenceToken => _model.EndOfSentenceToken; 81public string UnknownToken => _model.UnknownToken; 86public int BeginningOfSentenceId => _model.BeginningOfSentenceId; 91public int EndOfSentenceId => _model.EndOfSentenceId; 96public int UnknownId => _model.UnknownId; 106public override Normalizer? Normalizer => _model.Normalizer; 111public IReadOnlyDictionary<string, int> Vocabulary => _model.Vocabulary; 123Tokens = _model.EncodeToTokens(text, textSpan, out string? normalizedText, AddBeginningOfSentence, AddEndOfSentence, settings.ConsiderNormalization), 140=> _model.EncodeToTokens(text, Span<char>.Empty, out normalizedText, addBeginningOfSentence, addEndOfSentence, considerNormalization); 153=> _model.EncodeToTokens(null, text, out normalizedText, addBeginningOfSentence, addEndOfSentence, considerNormalization); 167Tokens = _model.EncodeToIds(text, textSpan, AddBeginningOfSentence, AddEndOfSentence, settings.ConsiderNormalization, out string? normalizedText, out int charsConsumed, settings.MaxTokenCount), 183=> _model.EncodeToIds(text, Span<char>.Empty, addBeginningOfSentence, addEndOfSentence, considerNormalization, out _, out _); 195=> _model.EncodeToIds(null, text, addBeginningOfSentence, addEndOfSentence, considerNormalization, out _, out _); 210=> _model.EncodeToIds(text, Span<char>.Empty, addBeginningOfSentence, addEndOfSentence, considerNormalization, out normalizedText, out charsConsumed, maxTokenCount); 225=> _model.EncodeToIds(null, text, addBeginningOfSentence, addEndOfSentence, considerNormalization, out normalizedText, out charsConsumed, maxTokenCount); 235=> _model.CountTokens(text, textSpan, AddBeginningOfSentence, AddEndOfSentence, settings.ConsiderNormalization, out _, out _, settings.MaxTokenCount); 247=> _model.CountTokens(text, ReadOnlySpan<char>.Empty, addBeginningOfSentence, addEndOfSentence, considerNormalization, out _, out _, int.MaxValue); 259=> _model.CountTokens(null, text, addBeginningOfSentence, addEndOfSentence, considerNormalization, out _, out _, int.MaxValue); 274=> _model.CountTokens(text, ReadOnlySpan<char>.Empty, addBeginningOfSentence, addEndOfSentence, considerNormalization, out normalizedText, out charsConsumed, maxTokenCount); 289=> _model.CountTokens(null, text, addBeginningOfSentence, addEndOfSentence, considerNormalization, out normalizedText, out charsConsumed, maxTokenCount); 311return _model.GetIndexByTokenCountFromEnd(text, textSpan, AddBeginningOfSentence, AddEndOfSentence, settings.MaxTokenCount, settings.ConsiderNormalization, out normalizedText, out tokenCount); 314tokenCount = _model.CountTokens(text, textSpan, AddBeginningOfSentence, AddEndOfSentence, settings.ConsiderNormalization, out normalizedText, out int charsConsumed, settings.MaxTokenCount); 336tokenCount = _model.CountTokens(text, ReadOnlySpan<char>.Empty, addBeginningOfSentence, addEndOfSentence, considerNormalization, out normalizedText, out int charsConsumed, maxTokenCount); 358tokenCount = _model.CountTokens(null, text, addBeginningOfSentence, addEndOfSentence, considerNormalization, out normalizedText, out int charsConsumed, maxTokenCount); 377=> _model.GetIndexByTokenCountFromEnd(text, ReadOnlySpan<char>.Empty, addBeginningOfSentence, addEndOfSentence, maxTokenCount, considerNormalization, out normalizedText, out tokenCount); 394=> _model.GetIndexByTokenCountFromEnd(null, text, addBeginningOfSentence, addEndOfSentence, maxTokenCount, considerNormalization, out normalizedText, out tokenCount); 401public override string Decode(IEnumerable<int> ids) => _model.Decode(ids, considerSpecialTokens: false); 409public string Decode(IEnumerable<int> ids, bool considerSpecialTokens) => _model.Decode(ids, considerSpecialTokens); 420=> _model.Decode(ids, destination, considerSpecialTokens: false, out idsConsumed, out charsWritten); 432=> _model.Decode(ids, destination, considerSpecialTokens, out idsConsumed, out charsWritten);