Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The recognition ability of the Audio model #69

Open
b3326023 opened this issue Dec 10, 2019 · 0 comments
Open

The recognition ability of the Audio model #69

b3326023 opened this issue Dec 10, 2019 · 0 comments
Labels
question Further information is requested

Comments

@b3326023
Copy link

Excuse me, I have search that the audio pretrained model used in this project is Speech-Command, and it use over 105,000 WAVE audio files of people saying thirty different words.

So this base model have the ability to well recognize many different words, and its learned low level feature should be only associated with the speech only.

But my question is that why the transfer learning model trained on some very different audio samples like clap table, water sounds, whistle, etc, such non-speech sounds, are also magically perform very well?

@irealva irealva added the question Further information is requested label Dec 13, 2019
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants