TDA4VE-Q1: tda4-sdk-J721s2: 8bit model accuracy decline

Part Number: TDA4VE-Q1


when the model is converted in the sdk, if both numParamBits and numFeatureBits in the import config are set to 8 bits, the model's inference results will differ significantly from those of the pc. howwever, if they are set to 16 bits, the model's inference results will be basically conssistent with those of pc.
for this purpose, we located the network layers that led to a significant drop in accuracy through a mixed-precision approach. these network layers are named in the.layer_info.txt file generated by th import tool: 37,38,39,40,41,195,196,197,198,199,204,266,267,268,269,270. these layers were transformed using 16-bit data precision to obtain a mixed-precision model. the inference results of the mixed_precision model have a significant improvement in precision compared to the 8-bit model. however, the inference speed of the mixed-precision model is much slower than that of the 8-bit model. our requirement is for fast inference speed and high precision

  • Hello!

    We have received your case and will take some time. Thank you for your patience.

  • Thank you for the question. 

    Generally speaking, you will expect some performance hit when you increase the setting from 8bit to 16bit.

    My understanding is that you are using mixed-precision model, and the inference performance (speed) is not good enough for you, correct? 

    Could you show us the exact settings in your import config file? (Prefer your original file or copy-pasted from your original file, not just describing them in English.)

    Also, are you running the inference on the EVM? If so, could you send us the log-text/screenshot about your inference running results. Because, sometimes the inference performance with PC simulation is not accurate.