Replies: 1 comment 3 replies
-
I do agree with everything you wrote and see your problems Regarding low resolution and blurred faces - we have a story in the backlog to add a blur plugin that will determine the level of blur. |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I am currently using CompreFace via Double Take and Frigate.. not that my question is specific to that but I am running into issues that seem to be related to both the source image being tested and the source images for training.
I looked at the few discussions on training and I am starting to think the documentation needs to explicitly state what resolutions make a good training images, example poses etc as well as don'ts.. Below are some of the challenges I have been finding and a few would things probably could have been avoided with some form of guide on training.
Garbage in equals garbage out:
While this may be intuitive and Double Take has some tools for this I have found that adding even a single lower res image or at least the face area is very small in total pixels or one with some artificing like pulling from surveillance footage seems to cause a drastic skew in results.. IE low res training images cause high confidence matches on almost all faces. I wonder if ComreFace itself should just cut off any face below a specific pixel count from training to being with as protection against this.
At this point I am basically avoiding training from any frigate source data unless the subject is filling the frame, free of artefacts and well lit.
Guidance on what makes a good training image:
Surveillance images as a source will catch people at all angles from a high angle above, in profile etc. Initially I was training with mostly dead on images using my higher quality phone, then I started adding up, down, left, right looking photos.. But as I try to improve detection I am now sometimes getting ears detected (face box is just the ear) as faces with high confidence matches ETC..
Is there guidance on how training images should be taken?
Family resemblance is a very small % certainty difference some times:
From a distance CompreFace seems to be all over the map in terms of separating our family of 5 from each other.. Where I am about the only one nearly always correct. I frequently get up to 98% confident wrong matches this is with high quality training images but low quality surveillance images at a distance sometimes, I have bumped up the min area in Double Take which cuts these down a lot, but even then I still get a lot of bad ones. Most seem to happen when the still uses has smear or artifacting from capturing a subject moving fast or in poor lite. I wonder if there is any way CompreFace it self could filter for "poor detail" or noisy images? Most of the time an unknown result would be better than a match.
Over all it is interesting that particularly with my kids that I am getting matching between 90-98% on the low smear or blurry images instead of a low confidence match.
Radom NOT A FACE things seem to turn up more since I have been adding training images:
I have det_prob_threshold set to 0.85 the default was 0.8 however I am thinking of setting it higher as it likes to find faces in video artifacts, most of the time those come back as unknown thankfully and since I stopped using the surveillance video as a training source.
edit: at .85 I still had a cup of coffee detected, but looks like pushing to .9 got rid of that.. much like some of the other values I am surprised I have to push so close to 100 essentially to get rid of some of these false positives.
I have looked at the documentation, and if I missed anything on guidelines for good training please let me know.
Also for reference most of my security cameras are currently lower end 1080p units and most problems happen "at a distance" but not always.. sometime is is just a matter of someone looking down when or in profile that causes a high confidence false positive.
I was playing around with deep stack as well and noticed confidence levels had much wider ranges even after training, however over all I have found CompreFace more accurate.
Beta Was this translation helpful? Give feedback.
All reactions