-
Notifications
You must be signed in to change notification settings - Fork 313
/
MF_hmm_fit.m
135 lines (115 loc) · 4.84 KB
/
MF_hmm_fit.m
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
function out = MF_hmm_fit(y,trainp,numStates,randomSeed)
% MF_hmm_fit Fits a Hidden Markov Model to sequential data.
%
% Actually highly stochastic, so for reproducible results helps to reset the
% random seed...
%
%---INPUTS:
% y, the input time series
% trainp, the proportion of data to train on, 0 < trainp < 1
% numStates, the number of states in the HMM
% Uses Zoubin Gharamani's implementation of HMMs for real-valued Gaussian
% observations:
% http://www.gatsby.ucl.ac.uk/~zoubin/software.html
% or, specifically:
% http://www.gatsby.ucl.ac.uk/~zoubin/software/hmm.tar.gz
%
% Uses ZG_hmm (renamed from hmm) and ZG_hmm_cl (renamed from hmm_cl)
% ------------------------------------------------------------------------------
% Copyright (C) 2020, Ben D. Fulcher <ben.d.fulcher@gmail.com>,
% <http://www.benfulcher.com>
%
% If you use this code for your research, please cite the following two papers:
%
% (1) B.D. Fulcher and N.S. Jones, "hctsa: A Computational Framework for Automated
% Time-Series Phenotyping Using Massive Feature Extraction, Cell Systems 5: 527 (2017).
% DOI: 10.1016/j.cels.2017.10.001
%
% (2) B.D. Fulcher, M.A. Little, N.S. Jones, "Highly comparative time-series
% analysis: the empirical structure of time series and their methods",
% J. Roy. Soc. Interface 10(83) 20130048 (2013).
% DOI: 10.1098/rsif.2013.0048
%
% This function is free software: you can redistribute it and/or modify it under
% the terms of the GNU General Public License as published by the Free Software
% Foundation, either version 3 of the License, or (at your option) any later
% version.
%
% This program is distributed in the hope that it will be useful, but WITHOUT
% ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
% FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
% details.
%
% You should have received a copy of the GNU General Public License along with
% this program. If not, see <http://www.gnu.org/licenses/>.
% ------------------------------------------------------------------------------
% Check required function files exist:
if ~exist(fullfile('ZG_hmm','ZG_hmm_cl'),'file') || ~exist(fullfile('ZG_hmm','ZG_hmm_cl'),'file')
error('Could not find the required HMM fitting functions (Zoubin Gharamani''s code)');
end
% ------------------------------------------------------------------------------
%% Check Inputs
% ------------------------------------------------------------------------------
N = length(y); % number of samples in time series
if nargin < 2 || isempty(trainp)
fprintf(1,'Training on 80%% of the data by default\n');
trainp = 0.8; % train on 80% of the data
end
if nargin < 3 || isempty(numStates)
fprintf(1,'Using 3 states by default\n');
numStates = 3; % use 3 states
end
if nargin < 4
randomSeed = [];
end
%-------------------------------------------------------------------------------
% Deal with random seeds
%-------------------------------------------------------------------------------
BF_ResetSeed(randomSeed); % reset the random seed if specified
% ------------------------------------------------------------------------------
%% Train the HMM
% ------------------------------------------------------------------------------
% Divide up dataset into training (ytrain) and test (ytest) portions
Ntrain = floor(trainp*N);
if Ntrain == N
error('No data for test set for HMM fitting?!')
end
ytrain = y(1:Ntrain);
if Ntrain < N
ytest = y(Ntrain+1:end);
Ntest = length(ytest);
end
% Train HMM with <numStates> states for 30 cycles of EM (or until
% convergence); default termination tolerance
[Mu, Cov, P, Pi, LL] = ZG_hmm(ytrain,Ntrain,numStates,30);
% ------------------------------------------------------------------------------
%% Output statistics on the training
% ------------------------------------------------------------------------------
% Mean vector, Mu
Musort = sort(Mu,'ascend');
for i = 1:length(Mu)
out.(sprintf('Mu_%u',i)) = Musort(i); % use dynamic field referencing
% eval(sprintf('out.Mu_%u = Musort(%u);',i,i));
end
out.meanMu = mean(Mu);
out.rangeMu = max(Mu) - min(Mu);
out.maxMu = max(Mu);
out.minMu = min(Mu);
% Covariance Cov
out.Cov = Cov;
% Transition matrix
out.Pmeandiag = mean(diag(P));
out.stdmeanP = std(mean(P));
out.maxP = max(P(:));
out.meanP = mean(P(:)); % I guess this is just 1/numStates? A constant so not a useful output.
out.stdP = std(P(:));
% Within-sample log-likelihood
out.LLtrainpersample = max(LL)/Ntrain; % loglikelihood per sample
out.nit = length(LL); % number of iterations
% ------------------------------------------------------------------------------
%% Calculate log likelihood for the test data
% ------------------------------------------------------------------------------
lik = ZG_hmm_cl(ytest,Ntest,numStates,Mu,Cov,P,Pi);
out.LLtestpersample = lik/Ntest;
out.LLdifference = out.LLtestpersample - out.LLtrainpersample;
end