SentencePieceTests.cs (39)
230byte[] model = MakeModelProto(ts.ToArray());
252byte[] model = MakeModelProto(MakeBpeTrainerSpec(), ns.ToArray());
267byte[] model = MakeModelProto(ts.ToArray());
304model.WriteMessageField(2, ts.ToArray()); // TrainerSpec
307SentencePieceTokenizer tokenizer = CreateFromSyntheticModel(model.ToArray());
325byte[] model = MakeModelProto(ts.ToArray());
357model.WriteMessageField(1, piece.ToArray());
358model.WriteMessageField(2, ts.ToArray());
359model.WriteMessageField(3, ns.ToArray());
362SentencePieceTokenizer tokenizer = CreateFromSyntheticModel(model.ToArray());
378byte[] model = MakeModelProto(ts.ToArray());
393SentencePieceTokenizer tokenizer = CreateFromSyntheticModel(model.ToArray());
440byte[] model = MakeModelProto(ts.ToArray());
452byte[] model = MakeModelProto(ts.ToArray());
470ts.ToArray(),
517using MemoryStream ms = new MemoryStream(w.ToArray());
531model.WriteMessageField(1, piece.ToArray());
534using MemoryStream ms = new MemoryStream(model.ToArray());
548model.WriteMessageField(2, ts.ToArray());
550using MemoryStream ms = new MemoryStream(model.ToArray());
562using MemoryStream ms = new MemoryStream(w.ToArray());
587SentencePieceTokenizer tokenizer = CreateFromSyntheticModel(model.ToArray());
601byte[] model = MakeModelProto(ts.ToArray());
616byte[] model = MakeModelProto(ts.ToArray());
638byte[] model = MakeModelProto(ts.ToArray(), ns.ToArray());
670byte[] model = MakeModelProto(ts.ToArray());
684SentencePieceTokenizer tokenizer = CreateFromSyntheticModel(model.ToArray());
704model.WriteMessageField(2, ts1.ToArray());
705model.WriteMessageField(2, ts2.ToArray());
707SentencePieceTokenizer tokenizer = CreateFromSyntheticModel(model.ToArray());
727model.WriteMessageField(2, ts1.ToArray());
728model.WriteMessageField(2, ts2.ToArray());
730SentencePieceTokenizer tokenizer = CreateFromSyntheticModel(model.ToArray());
747piece.ToArray());
764byte[] model = MakeModelProto(ts.ToArray());
813return w.ToArray();
820return w.ToArray();
836return w.ToArray();