OCR Implementation - Guidance #44
-
I was wondering if anyone could give me insight and tips on how I could implement OCR using the SDK. For more context, I am trying to read a word that is located above the user’s finger wherever he points to the document. I already made an algorithm that crops out the word using OpenCV with colour thresholding and some histogram analysis. I thought of maybe building a pipeline which takes the cropped image and feeds it to the NN but again I’m not sure it’s possible. If anyone has ideas I would love to hear them, thanks! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 3 replies
-
Hi @Excustic , And I think you should train your own NN model for the OCR. If you want to train your own yolov8n and finally put it on WE2, here is the tutorial. And about the crop function, you can use hx_lib_image_copy_helium for helium version. |
Beta Was this translation helpful? Give feedback.
Hi @Excustic ,
You can reference https://github.com/tensorflow/tflite-micro/blob/main/tensorflow/lite/micro/docs/memory_management.md to see the memory management of tensor arena because the
tail
part of tensor arena can not be share and other part of tensor arena can be reuse at difference model.OV5647_SUPPORT_BINNING
at here and disable the sub-sample and binning and set the crop position where you want. But be careful that each cropped areas coul…