-
Notifications
You must be signed in to change notification settings - Fork 1.2k
Description
Is segmentation of sounds works well ?
I`ve listened to TalkingMachinesPodcast.wav, watched waves spectre in Camtasia Studio and compared with output of
segmentation.py:
('0.0m 0.0s', 0)
('0.0m 2.Ss', 1)
('0.0m 5.0s', 0)
( '0.0m 7 .Ss ', 1)
('0.0m 10.0s', 1)
('0.0m 12.Ss', 1)
('0.0m 15.0s', 1)
('0.0m 17.Ss', 0)
('0.0m 20.0s', 1)
('0.0m 22.Ss', 1)
('0.0m 25.0s', 0)
('0.0m 27.Ss', 0)
And i dont think that it works correct as it makes segment no exectly close where they should be.
Moreover if i put k=3 i see:
('0.0m 0.0s', 0)
('0.0m 5.0s', 1)
('0.0m 10.0s', 2)
('0.0m 15.0s', 2)
('0.0m 20.0s', 1)
('0.0m 25.0s', 2)
('0.0m 30.0s', 1)
('0.0m 35.0s', 2)
('0.0m 40.0s', 1)
('0.0m 45.0s', 2)
('0.0m 50.0s', 2)
('0.0m 55.0s', 2)
Why timeline in code run away ? And we have much more clusters ?