Skip to content

Conversation

@st-tuanmai
Copy link
Contributor

No description provided.

@PaulTR
Copy link
Collaborator

PaulTR commented Oct 9, 2024

Can you verify the classification happening in the app? It's picking up a lot of random things that aren't happening, and not picking up things like "speech" or "whistling" when that is happening. Thanks.

@st-tuanmai
Copy link
Contributor Author

Can you verify the classification happening in the app? It's picking up a lot of random things that aren't happening, and not picking up things like "speech" or "whistling" when that is happening. Thanks.

I will check it now

@st-tuanmai
Copy link
Contributor Author

I updated the branch to change TensorFlowLiteTaskAudio to TensorFlowLiteSwift.
Please help me check again.
Thanks you.

@PaulTR
Copy link
Collaborator

PaulTR commented Oct 28, 2024

I am still seeing wrong results consistently with this. Just whistling into the phone and not getting 'whistling'.

@st-tuanmai
Copy link
Contributor Author

Hi @PaulTR
I checked the model label file and the speech_commands label doesn't have whistling and speech. Please help me check.

labels:
background
down
go
left
off
on
right
stop
up

@PaulTR
Copy link
Collaborator

PaulTR commented Oct 29, 2024 via email

@PaulTR
Copy link
Collaborator

PaulTR commented Oct 29, 2024 via email

@st-tuanmai
Copy link
Contributor Author

I am using label.txt file extracted from tflite file using python code, I think label is added wrongly to metadata.
vocal = zipfile.ZipFile('./demo/speech_commands.tflite').extractall('label')

@PaulTR
Copy link
Collaborator

PaulTR commented Oct 30, 2024

Not comparing the speech commands model - can you verify this all with the standard audio classification model (the one with 'music', 'whispering', 'whistling', etc.)? Thanks.

@st-tuanmai
Copy link
Contributor Author

st-tuanmai commented Nov 1, 2024

@PaulTR
Copy link
Collaborator

PaulTR commented Nov 1, 2024

OK again we're not looking at the speech models for this issue. Only looking at the sound classification model that does whistling/whispering/music/etc. the label that shows up is not matching the sound.

Delete the speech model if that's adding a complication, it doesn't matter for the sample. We need the first general sound classification model to work correctly.

@PaulTR
Copy link
Collaborator

PaulTR commented Nov 5, 2024

You can compare the iOS sample to the Android sample. They're using this model: https://storage.googleapis.com/ai-edge/interpreter-samples/audio_classification/android/yamnet.tflite (I copied the same one into the ios folder) and returning the correct results.

@st-tuanmai
Copy link
Contributor Author

@PaulTR
Copy link
Collaborator

PaulTR commented Nov 6, 2024 via email

@st-tuanmai
Copy link
Contributor Author

@PaulTR I removed the speech commands model, I think we can use only yamnet model now.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants