Hi everyone,
I’m working with the micro speech example from the tflite-micro GitHub repository and have encountered an issue with the model architecture when training.
Observation:
-
The pre-trained
micro_speech_quantized.tflitemodel (included in the micro_speech example) usesDepthwiseConv2Dlayers, as documented intrain/README.md -
However, when I train a new model using
train_micro_speech_model.ipynbin Google Colab, the resulting model containsConv2Dlayers instead ofDepthwiseConv2D
Context: I believe this might be related to the TensorFlow version being used during training. Since DepthwiseConv2D is more efficient for microcontroller deployment, I’d prefer to train with that architecture.
Questions:
-
What’s the recommended approach to ensure the trained model uses DepthwiseConv2D layers?
-
Is there a specific TensorFlow version I should use with this notebook?
-
Are there any configuration parameters or model architecture flags I should set?
Environment:
-
Notebook:
train_micro_speech_model.ipynbfrom tflite-micro repo -
Platform: Google Colab
-
TensorFlow version: [I can provide this if it helps]
I’d really appreciate any guidance on this. Thank you in advance for your help!