2 writes to _model
Microsoft.ML.Tokenizers (2)
Model\SentencePieceTokenizer.cs (2)
25_model = modelProto.TrainerSpec.ModelType switch 35_model = options.ModelType switch
39 references to _model
Microsoft.ML.Tokenizers (39)
Model\SentencePieceTokenizer.cs (39)
46public IReadOnlyDictionary<string, int>? SpecialTokens => _model.SpecialTokens; 51public bool ByteFallback => _model.ByteFallback; 56public bool AddDummyPrefix => _model.AddDummyPrefix; 61public bool EscapeWhiteSpaces => _model.EscapeWhiteSpaces; 66public bool TreatWhitespaceAsSuffix { get => _model.TreatWhitespaceAsSuffix; private set => _model.TreatWhitespaceAsSuffix = value; } 71public bool AddBeginningOfSentence => _model.AddBeginningOfSentence; 76public bool AddEndOfSentence => _model.AddEndOfSentence; 81public string BeginningOfSentenceToken => _model.BeginningOfSentenceToken; 86public string EndOfSentenceToken => _model.EndOfSentenceToken; 91public string UnknownToken => _model.UnknownToken; 96public int BeginningOfSentenceId => _model.BeginningOfSentenceId; 101public int EndOfSentenceId => _model.EndOfSentenceId; 106public int UnknownId => _model.UnknownId; 116public override Normalizer? Normalizer => _model.Normalizer; 121public IReadOnlyDictionary<string, int> Vocabulary => _model.Vocabulary; 133Tokens = _model.EncodeToTokens(text, textSpan, out string? normalizedText, AddBeginningOfSentence, AddEndOfSentence, settings.ConsiderNormalization), 150=> _model.EncodeToTokens(text, Span<char>.Empty, out normalizedText, addBeginningOfSentence, addEndOfSentence, considerNormalization); 163=> _model.EncodeToTokens(null, text, out normalizedText, addBeginningOfSentence, addEndOfSentence, considerNormalization); 177Tokens = _model.EncodeToIds(text, textSpan, AddBeginningOfSentence, AddEndOfSentence, settings.ConsiderNormalization, out string? normalizedText, out int charsConsumed, settings.MaxTokenCount), 193=> _model.EncodeToIds(text, Span<char>.Empty, addBeginningOfSentence, addEndOfSentence, considerNormalization, out _, out _); 205=> _model.EncodeToIds(null, text, addBeginningOfSentence, addEndOfSentence, considerNormalization, out _, out _); 220=> _model.EncodeToIds(text, Span<char>.Empty, addBeginningOfSentence, addEndOfSentence, considerNormalization, out normalizedText, out charsConsumed, maxTokenCount); 235=> _model.EncodeToIds(null, text, addBeginningOfSentence, addEndOfSentence, considerNormalization, out normalizedText, out charsConsumed, maxTokenCount); 245=> _model.CountTokens(text, textSpan, AddBeginningOfSentence, AddEndOfSentence, settings.ConsiderNormalization, out _, out _, settings.MaxTokenCount); 257=> _model.CountTokens(text, ReadOnlySpan<char>.Empty, addBeginningOfSentence, addEndOfSentence, considerNormalization, out _, out _, int.MaxValue); 269=> _model.CountTokens(null, text, addBeginningOfSentence, addEndOfSentence, considerNormalization, out _, out _, int.MaxValue); 284=> _model.CountTokens(text, ReadOnlySpan<char>.Empty, addBeginningOfSentence, addEndOfSentence, considerNormalization, out normalizedText, out charsConsumed, maxTokenCount); 299=> _model.CountTokens(null, text, addBeginningOfSentence, addEndOfSentence, considerNormalization, out normalizedText, out charsConsumed, maxTokenCount); 321return _model.GetIndexByTokenCountFromEnd(text, textSpan, AddBeginningOfSentence, AddEndOfSentence, settings.MaxTokenCount, settings.ConsiderNormalization, out normalizedText, out tokenCount); 324tokenCount = _model.CountTokens(text, textSpan, AddBeginningOfSentence, AddEndOfSentence, settings.ConsiderNormalization, out normalizedText, out int charsConsumed, settings.MaxTokenCount); 346tokenCount = _model.CountTokens(text, ReadOnlySpan<char>.Empty, addBeginningOfSentence, addEndOfSentence, considerNormalization, out normalizedText, out int charsConsumed, maxTokenCount); 368tokenCount = _model.CountTokens(null, text, addBeginningOfSentence, addEndOfSentence, considerNormalization, out normalizedText, out int charsConsumed, maxTokenCount); 387=> _model.GetIndexByTokenCountFromEnd(text, ReadOnlySpan<char>.Empty, addBeginningOfSentence, addEndOfSentence, maxTokenCount, considerNormalization, out normalizedText, out tokenCount); 404=> _model.GetIndexByTokenCountFromEnd(null, text, addBeginningOfSentence, addEndOfSentence, maxTokenCount, considerNormalization, out normalizedText, out tokenCount); 411public override string Decode(IEnumerable<int> ids) => _model.Decode(ids, considerSpecialTokens: false); 419public string Decode(IEnumerable<int> ids, bool considerSpecialTokens) => _model.Decode(ids, considerSpecialTokens); 430=> _model.Decode(ids, destination, considerSpecialTokens: false, out idsConsumed, out charsWritten); 442=> _model.Decode(ids, destination, considerSpecialTokens, out idsConsumed, out charsWritten);