I have a bunch of brief (~1 second each) wave files, some of which contain a frequency-modulated sound of interest between about 3900-4500 Hz. I thought I might run MFCC calculations on these sound files and use the results to pursue clustering/separation between files that contain the sound of interest and files that do not.
Since I'm not interested in the other frequency bands, I limited the calculation bounds to the aforementioned range. However, my MFCC matrix outputs as a bunch of NaNs, presumably because the range is so narrow. This either means that MFCCs are an inappropriate measure in general for such a narrow frequency range, or that I have yet to take advantage of the numerous other parameters involved in this calculation (e.g. window size, DCT type, etc.). What might be going on here? It could be that I should be investigating other acoustic features for this purpose instead of MFCCs, but I am a total audio rookie and don't know where to start.
(I'm only looking for a general/big picture answer to this question, but in case it helps provide context, I'm using R's melfcc() function to calculate these. I believe it is based on the same code as that for Matlab.)
EDIT: For more background.
I have a bunch of "detections" of an animal sound event from real field recordings, detected via a spectrogram cross correlation template. Some of these detections are true positives, some are false positives. I want to find acoustic features that might help me predict which are true and which are false. (In addition to the acoustic features, I can use other things such as time of day detected, etc.) Ultimately pursuing a Bayesian system in which the model can learn over time as new detection data are incorporated. Not attached to any acoustic features in particular, but not sure what to try and MFCCs seemed like somewhere to start.