@@ -134,7 +134,6 @@ public enum ModelState: CustomStringConvertible {
134134 }
135135}
136136
137- @available ( macOS 13 , iOS 16 , watchOS 10 , visionOS 1 , * )
138137public struct ModelComputeOptions {
139138 public var melCompute : MLComputeUnits
140139 public var audioEncoderCompute : MLComputeUnits
@@ -252,7 +251,6 @@ public struct ModelSupportConfig: Codable {
252251 computeDisabledModels ( )
253252 }
254253
255- @available ( macOS 13 , iOS 16 , watchOS 10 , visionOS 1 , * )
256254 public func modelSupport( for deviceIdentifier: String = WhisperKit . deviceName ( ) ) -> ModelSupport {
257255 // Find the support with the longest matching identifier prefix
258256 // i.e. `iPad13,16` should match exact `iPad13,16` instead of first prefix like `iPad13,1`
@@ -409,7 +407,6 @@ public enum ChunkingStrategy: String, Codable, CaseIterable {
409407 case vad
410408}
411409
412- @available ( macOS 13 , iOS 16 , watchOS 10 , visionOS 1 , * )
413410public struct DecodingFallback {
414411 public var needsFallback : Bool
415412 public var fallbackReason : String
@@ -420,7 +417,6 @@ public struct DecodingFallback {
420417 }
421418}
422419
423- @available ( macOS 13 , iOS 16 , watchOS 10 , visionOS 1 , * )
424420public extension DecodingFallback {
425421 init ? (
426422 options: DecodingOptions ,
@@ -447,7 +443,6 @@ public extension DecodingFallback {
447443 }
448444}
449445
450- @available ( macOS 13 , iOS 16 , watchOS 10 , visionOS 1 , * )
451446public struct DecodingResult {
452447 public var language : String
453448 public var languageProbs : [ String : Float ]
@@ -900,7 +895,6 @@ public class MelSpectrogramInput: MLFeatureProvider {
900895}
901896
902897/// Model Prediction Output Type
903- @available ( macOS 13 , iOS 16 , watchOS 10 , visionOS 1 , * )
904898public class MelSpectrogramOutput : MLFeatureProvider {
905899 /// Source provided by CoreML
906900 private let provider : MLFeatureProvider
@@ -937,7 +931,6 @@ public class MelSpectrogramOutput: MLFeatureProvider {
937931// MARK: AudioEncoder
938932
939933/// Model Prediction Input Type
940- @available ( macOS 13 , iOS 16 , watchOS 10 , visionOS 1 , * )
941934public class AudioEncoderInput : MLFeatureProvider {
942935 /// melspectrogram_features as 1 × {80,128} × 1 × 3000 4-dimensional array of floats
943936 public var melspectrogram_features : MLMultiArray
@@ -963,7 +956,6 @@ public class AudioEncoderInput: MLFeatureProvider {
963956}
964957
965958/// Model Prediction Output Type
966- @available ( macOS 13 , iOS 16 , watchOS 10 , visionOS 1 , * )
967959public class AudioEncoderOutput : MLFeatureProvider {
968960 /// Source provided by CoreML
969961 private let provider : MLFeatureProvider
@@ -1000,7 +992,6 @@ public class AudioEncoderOutput: MLFeatureProvider {
1000992// MARK: TextDecoder
1001993
1002994/// Model Prediction Input Type
1003- @available ( macOS 13 , iOS 16 , watchOS 10 , visionOS 1 , * )
1004995public class TextDecoderInput : MLFeatureProvider {
1005996 /// input_ids as 1 element vector of 32-bit integers
1006997 public var input_ids : MLMultiArray
@@ -1068,7 +1059,6 @@ public class TextDecoderInput: MLFeatureProvider {
10681059}
10691060
10701061/// Model Prediction Output Type
1071- @available ( macOS 13 , iOS 16 , watchOS 10 , visionOS 1 , * )
10721062public class TextDecoderOutput : MLFeatureProvider {
10731063 /// Source provided by CoreML
10741064 private let provider : MLFeatureProvider
@@ -1175,7 +1165,6 @@ public class TextDecoderCachePrefillInput: MLFeatureProvider {
11751165}
11761166
11771167/// Model Prediction Output Type
1178- @available ( macOS 13 , iOS 16 , watchOS 10 , visionOS 1 , * )
11791168public class TextDecoderCachePrefillOutput : MLFeatureProvider {
11801169 /// Source provided by CoreML
11811170 private let provider : MLFeatureProvider
0 commit comments