Skip to content

Conversation

morgolock
Copy link

  • Updated convolution reference to branch epilogue:
    • TO=float: int32 to float dequant (acc * sA * sB + bias_f32)
    • TO!=float: usual quantize_down_scale_by_fixedpoint with int32 bias
  • Changed fixture to use F32 bias tensor for Q->F32 runs (instead of S32), matching arm_gemm dequant epilogue which only supports float bias.
  • Added explicit template instantiations for convolution_layer with TBias=float, TO=float to fix linker errors in validation.
  • Disabled activation in arm_gemm dequant path: offsets are applied afterwards by CpuGemmLowpOffsetContributionKernel, so activation must run there to see the correct final accumulator.

This aligns target and reference for quantized to F32 convolution tests and prevents premature clamping before offset contributions.

Change-Id: I6fffc98dc0798542a2702e6a593b850c16561e3b

@morgolock morgolock requested a review from gunes-arm September 19, 2025 10:56
@morgolock morgolock force-pushed the pr/conv_f32_dequant branch 3 times, most recently from ef3f28a to f3085cd Compare September 19, 2025 13:03
- Updated convolution reference to branch epilogue:
  * TO=float: int32 to float dequant (acc * sA * sB + bias_f32)
  * TO!=float: usual quantize_down_scale_by_fixedpoint with int32 bias
- Changed fixture to use F32 bias tensor for Q->F32 runs (instead of S32),
  matching arm_gemm dequant epilogue which only supports float bias.
- Added explicit template instantiations for convolution_layer with
  TBias=float, TO=float to fix linker errors in validation.
- Disabled activation in arm_gemm dequant path:
  offsets are applied afterwards by CpuGemmLowpOffsetContributionKernel,
  so activation must run there to see the correct final accumulator.

This aligns target and reference for quantized to F32 convolution tests
and prevents premature clamping before offset contributions.

Change-Id: I6fffc98dc0798542a2702e6a593b850c16561e3b
Signed-off-by: Pablo Marquez Tello <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants