99import Foundation
1010import Hub
1111
12- public class BertTokenizer : @ unchecked Sendable {
12+ public final class BertTokenizer : Sendable {
1313 private let basicTokenizer : BasicTokenizer
1414 private let wordpieceTokenizer : WordpieceTokenizer
1515 private let maxLen = 512
@@ -18,10 +18,10 @@ public class BertTokenizer: @unchecked Sendable {
1818 private let vocab : [ String : Int ]
1919 private let ids_to_tokens : [ Int : String ]
2020
21- public var bosToken : String ?
22- public var bosTokenId : Int ?
23- public var eosToken : String ?
24- public var eosTokenId : Int ?
21+ public let bosToken : String ?
22+ public let bosTokenId : Int ?
23+ public let eosToken : String ?
24+ public let eosTokenId : Int ?
2525
2626 public let fuseUnknownTokens : Bool
2727
@@ -177,7 +177,7 @@ extension BertTokenizer: PreTrainedTokenizerModel {
177177 }
178178}
179179
180- class BasicTokenizer {
180+ final class BasicTokenizer : Sendable {
181181 let doLowerCase : Bool
182182
183183 init ( doLowerCase: Bool = true ) {
@@ -243,7 +243,7 @@ private extension Character {
243243 }
244244}
245245
246- class WordpieceTokenizer {
246+ final class WordpieceTokenizer : Sendable {
247247 let unkToken = " [UNK] "
248248 private let maxInputCharsPerWord = 100
249249 private let vocab : [ String : Int ]
0 commit comments