Description
"dependencies": {
"@tensorflow/tfjs-backend-cpu": "^3.5.0",
"@tensorflow/tfjs-backend-webgl": "^3.5.0",
"@tensorflow/tfjs-converter": "^3.5.0",
"@tensorflow/tfjs-core": "^3.5.0",
"fetch-wechat": "^0.0.3",
"regenerator-runtime": "^0.14.1"
},
"plugins": {
"tfjsPlugin": {
"version": "0.2.0",
"provider": "wx6afed118d9e81df9"
}
},
WeChat version: 8.0.50
WeChat base API version: 3.5.0
WeChat IDE version: 1.06.2405020 win32-x64
I am trying to use a model with dynamic axis to adapt to different size of input, like [1, 3, 255,255] or [1,3,127,127]. So I have to use executeAsync to get model inference, since there is error using 'predict' method:
nano_tracker.js:36 Preheat failed: Error: This execution contains the node 'StatefulPartitionedCall/assert_equal_5/Assert/AssertGuard/branch_executed/_114', which has the dynamic op 'Merge'. Please use model.executeAsync() instead. Alternatively, to avoid the dynamic ops, specify the inputs [Identity] at e.compile (index.js:17) at e.execute (index.js:17) at e.execute (index.js:17) at e.predict (index.js:17) at ModelBuilder._callee3$ (model_builder.js:54) at s (regeneratorRuntime.js:1) at Generator.<anonymous> (regeneratorRuntime.js:1) at Generator.next (regeneratorRuntime.js:1) at asyncGeneratorStep (asyncToGenerator.js:1) at c (asyncToGenerator.js:1)(env: Windows,mp,1.06.2405020; lib: 3.5.3)
Using executeAsync seems to get correct model inference, while memory increasing with every invoke until it fails.
Is there anyone knows how to fix this? Thank you so much