From 54814cd9632799af5f01c7166f83a3a6b8a4f20a Mon Sep 17 00:00:00 2001 From: Vadim Mazalov Date: Sat, 18 Mar 2017 11:09:49 -0700 Subject: [PATCH] Updated baselines --- Tests/EndToEndTests/Speech/LSTM_CTC/baseline.linux.cpu.txt | 2 +- Tests/EndToEndTests/Speech/LSTM_CTC/baseline.linux.gpu.txt | 2 +- Tests/EndToEndTests/Speech/LSTM_CTC/baseline.windows.cpu.txt | 2 +- Tests/EndToEndTests/Speech/LSTM_CTC/baseline.windows.gpu.txt | 2 +- Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.linux.cpu.txt | 2 +- Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.linux.gpu.txt | 2 +- .../EndToEndTests/Speech/LSTM_CTC_MLF/baseline.windows.cpu.txt | 2 +- .../EndToEndTests/Speech/LSTM_CTC_MLF/baseline.windows.gpu.txt | 2 +- 8 files changed, 8 insertions(+), 8 deletions(-) diff --git a/Tests/EndToEndTests/Speech/LSTM_CTC/baseline.linux.cpu.txt b/Tests/EndToEndTests/Speech/LSTM_CTC/baseline.linux.cpu.txt index 8b20c655629b..f04fcdf7901c 100644 --- a/Tests/EndToEndTests/Speech/LSTM_CTC/baseline.linux.cpu.txt +++ b/Tests/EndToEndTests/Speech/LSTM_CTC/baseline.linux.cpu.txt @@ -474,7 +474,7 @@ Here are the ones that don't share memory: 02/18/2017 07:06:38: Starting Epoch 1: learning rate per sample = 0.005000 effective momentum = 0.900000 momentum as time constant = 189.8 samples 02/18/2017 07:06:38: Starting minibatch loop. -02/18/2017 07:06:39: Finished Epoch[ 1 of 10]: [Training] cr = 4.16293501 * 368; Err = 2.52941181 * 368; totalSamplesSeen = 368; learningRatePerSample = 0.0049999999; epochTime=1.00518s +02/18/2017 07:06:39: Finished Epoch[ 1 of 10]: [Training] cr = 4.16293501 * 368; Err = 2.36764709 * 368; totalSamplesSeen = 368; learningRatePerSample = 0.0049999999; epochTime=0.475233s 02/18/2017 07:06:39: SGD: Saving checkpoint model '/tmp/cntk-test-20170218070416.834755/Speech_LSTM_CTC@debug_gpu/models/simple.dnn.1' 02/18/2017 07:06:39: Starting Epoch 2: learning rate per sample = 0.005000 effective momentum = 0.900000 momentum as time constant = 189.8 samples diff --git a/Tests/EndToEndTests/Speech/LSTM_CTC/baseline.linux.gpu.txt b/Tests/EndToEndTests/Speech/LSTM_CTC/baseline.linux.gpu.txt index 7593a080593e..828845704076 100644 --- a/Tests/EndToEndTests/Speech/LSTM_CTC/baseline.linux.gpu.txt +++ b/Tests/EndToEndTests/Speech/LSTM_CTC/baseline.linux.gpu.txt @@ -342,7 +342,7 @@ Memory Sharing: Out of 200 matrices, 71 are shared as 27, and 129 are not shared 02/18/2017 07:06:38: Starting Epoch 1: learning rate per sample = 0.005000 effective momentum = 0.900000 momentum as time constant = 189.8 samples 02/18/2017 07:06:38: Starting minibatch loop. -02/18/2017 07:06:39: Finished Epoch[ 1 of 10]: [Training] cr = 4.16293501 * 368; Err = 2.52941181 * 368; totalSamplesSeen = 368; learningRatePerSample = 0.0049999999; epochTime=1.00518s +02/18/2017 07:06:39: Finished Epoch[ 1 of 10]: [Training] cr = 4.16293501 * 368; Err = 2.36764709 * 368; totalSamplesSeen = 368; learningRatePerSample = 0.0049999999; epochTime=0.984606s 02/18/2017 07:06:39: SGD: Saving checkpoint model '/tmp/cntk-test-20170218070416.834755/Speech_LSTM_CTC@debug_gpu/models/simple.dnn.1' 02/18/2017 07:06:39: Starting Epoch 2: learning rate per sample = 0.005000 effective momentum = 0.900000 momentum as time constant = 189.8 samples diff --git a/Tests/EndToEndTests/Speech/LSTM_CTC/baseline.windows.cpu.txt b/Tests/EndToEndTests/Speech/LSTM_CTC/baseline.windows.cpu.txt index de57957e504d..00ed93fadff0 100644 --- a/Tests/EndToEndTests/Speech/LSTM_CTC/baseline.windows.cpu.txt +++ b/Tests/EndToEndTests/Speech/LSTM_CTC/baseline.windows.cpu.txt @@ -467,7 +467,7 @@ Here are the ones that don't share memory: 02/22/2017 21:20:37: Starting Epoch 1: learning rate per sample = 0.005000 effective momentum = 0.900000 momentum as time constant = 189.8 samples 02/22/2017 21:20:37: Starting minibatch loop. -02/22/2017 21:20:38: Finished Epoch[ 1 of 10]: [Training] cr = 4.16293534 * 368; Err = 2.52941181 * 368; totalSamplesSeen = 368; learningRatePerSample = 0.0049999999; epochTime=0.926568s +02/22/2017 21:20:38: Finished Epoch[ 1 of 10]: [Training] cr = 4.16293534 * 368; Err = 2.36764709 * 368; totalSamplesSeen = 368; learningRatePerSample = 0.0049999999; epochTime=0.793934s 02/22/2017 21:20:38: SGD: Saving checkpoint model 'C:\Users\svcphil\AppData\Local\Temp\cntk-test-20170222211959.232036\Speech_LSTM_CTC@release_cpu/models/simple.dnn.1' 02/22/2017 21:20:38: Starting Epoch 2: learning rate per sample = 0.005000 effective momentum = 0.900000 momentum as time constant = 189.8 samples diff --git a/Tests/EndToEndTests/Speech/LSTM_CTC/baseline.windows.gpu.txt b/Tests/EndToEndTests/Speech/LSTM_CTC/baseline.windows.gpu.txt index 591628df917e..03ab3e514fe1 100644 --- a/Tests/EndToEndTests/Speech/LSTM_CTC/baseline.windows.gpu.txt +++ b/Tests/EndToEndTests/Speech/LSTM_CTC/baseline.windows.gpu.txt @@ -340,7 +340,7 @@ Memory Sharing: Out of 200 matrices, 71 are shared as 27, and 129 are not shared 02/18/2017 09:28:11: Starting Epoch 1: learning rate per sample = 0.005000 effective momentum = 0.900000 momentum as time constant = 189.8 samples 02/18/2017 09:28:11: Starting minibatch loop. -02/18/2017 09:28:13: Finished Epoch[ 1 of 10]: [Training] cr = 4.16293501 * 368; Err = 2.52941181 * 368; totalSamplesSeen = 368; learningRatePerSample = 0.0049999999; epochTime=1.03913s +02/18/2017 09:28:13: Finished Epoch[ 1 of 10]: [Training] cr = 4.16293501 * 368; Err = 2.36764709 * 368; totalSamplesSeen = 368; learningRatePerSample = 0.0049999999; epochTime=1.01936s 02/18/2017 09:28:13: SGD: Saving checkpoint model 'C:\Users\svcphil\AppData\Local\Temp\cntk-test-20170218092651.300722\Speech_LSTM_CTC@release_gpu/models/simple.dnn.1' 02/18/2017 09:28:13: Starting Epoch 2: learning rate per sample = 0.005000 effective momentum = 0.900000 momentum as time constant = 189.8 samples diff --git a/Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.linux.cpu.txt b/Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.linux.cpu.txt index 9fa3b21c13f9..f8bba40dfea4 100644 --- a/Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.linux.cpu.txt +++ b/Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.linux.cpu.txt @@ -474,7 +474,7 @@ Here are the ones that don't share memory: 02/18/2017 07:06:38: Starting Epoch 1: learning rate per sample = 0.005000 effective momentum = 0.900000 momentum as time constant = 189.8 samples 02/18/2017 07:06:38: Starting minibatch loop. -02/18/2017 07:06:39: Finished Epoch[ 1 of 10]: [Training] cr = 4.16290018 * 368; Err = 2.51470582 * 368; totalSamplesSeen = 368; learningRatePerSample = 0.0049999999; epochTime=0.486473s +02/18/2017 07:06:39: Finished Epoch[ 1 of 10]: [Training] cr = 4.16290018 * 368; Err = 2.36764709 * 368; totalSamplesSeen = 368; learningRatePerSample = 0.0049999999; epochTime=0.515702s 02/18/2017 07:06:39: SGD: Saving checkpoint model '/tmp/cntk-test-20170218070416.834755/Speech_LSTM_CTC@debug_gpu/models/simple.dnn.1' 02/18/2017 07:06:39: Starting Epoch 2: learning rate per sample = 0.005000 effective momentum = 0.900000 momentum as time constant = 189.8 samples diff --git a/Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.linux.gpu.txt b/Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.linux.gpu.txt index 4fb6045860e8..006f7ace2716 100644 --- a/Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.linux.gpu.txt +++ b/Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.linux.gpu.txt @@ -342,7 +342,7 @@ Memory Sharing: Out of 200 matrices, 71 are shared as 27, and 129 are not shared 02/18/2017 07:06:38: Starting Epoch 1: learning rate per sample = 0.005000 effective momentum = 0.900000 momentum as time constant = 189.8 samples 02/18/2017 07:06:38: Starting minibatch loop. -02/18/2017 07:06:39: Finished Epoch[ 1 of 10]: [Training] cr = 4.16290018 * 368; Err = 2.51470582 * 368; totalSamplesSeen = 368; learningRatePerSample = 0.0049999999; epochTime=1.20341s +02/18/2017 07:06:39: Finished Epoch[ 1 of 10]: [Training] cr = 4.16290018 * 368; Err = 2.36764709 * 368; totalSamplesSeen = 368; learningRatePerSample = 0.0049999999; epochTime=1.21637s 02/18/2017 07:06:39: SGD: Saving checkpoint model '/tmp/cntk-test-20170218070416.834755/Speech_LSTM_CTC@debug_gpu/models/simple.dnn.1' 02/18/2017 07:06:39: Starting Epoch 2: learning rate per sample = 0.005000 effective momentum = 0.900000 momentum as time constant = 189.8 samples diff --git a/Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.windows.cpu.txt b/Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.windows.cpu.txt index b58d797519e1..4281afb8808b 100644 --- a/Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.windows.cpu.txt +++ b/Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.windows.cpu.txt @@ -467,7 +467,7 @@ Here are the ones that don't share memory: 02/22/2017 21:20:37: Starting Epoch 1: learning rate per sample = 0.005000 effective momentum = 0.900000 momentum as time constant = 189.8 samples 02/22/2017 21:20:37: Starting minibatch loop. -02/22/2017 21:20:38: Finished Epoch[ 1 of 10]: [Training] cr = 4.16290018 * 368; Err = 2.51470582 * 368; totalSamplesSeen = 368; learningRatePerSample = 0.0049999999; epochTime=0.926568s +02/22/2017 21:20:38: Finished Epoch[ 1 of 10]: [Training] cr = 4.16290018 * 368; Err = 2.36764709 * 368; totalSamplesSeen = 368; learningRatePerSample = 0.0049999999; epochTime=0.963661s 02/22/2017 21:20:38: SGD: Saving checkpoint model 'C:\Users\svcphil\AppData\Local\Temp\cntk-test-20170222211959.232036\Speech_LSTM_CTC@release_cpu/models/simple.dnn.1' 02/22/2017 21:20:38: Starting Epoch 2: learning rate per sample = 0.005000 effective momentum = 0.900000 momentum as time constant = 189.8 samples diff --git a/Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.windows.gpu.txt b/Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.windows.gpu.txt index 6188aca7f5db..1017dbcc00ae 100644 --- a/Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.windows.gpu.txt +++ b/Tests/EndToEndTests/Speech/LSTM_CTC_MLF/baseline.windows.gpu.txt @@ -340,7 +340,7 @@ Memory Sharing: Out of 200 matrices, 71 are shared as 27, and 129 are not shared 02/18/2017 09:28:11: Starting Epoch 1: learning rate per sample = 0.005000 effective momentum = 0.900000 momentum as time constant = 189.8 samples 02/18/2017 09:28:11: Starting minibatch loop. -02/18/2017 09:28:13: Finished Epoch[ 1 of 10]: [Training] cr = 4.16290018 * 368; Err = 2.51470582 * 368; totalSamplesSeen = 368; learningRatePerSample = 0.0049999999; epochTime=0.74581s +02/18/2017 09:28:13: Finished Epoch[ 1 of 10]: [Training] cr = 4.16290018 * 368; Err = 2.36764709 * 368; totalSamplesSeen = 368; learningRatePerSample = 0.0049999999; epochTime=1.00822s 02/18/2017 09:28:13: SGD: Saving checkpoint model 'C:\Users\svcphil\AppData\Local\Temp\cntk-test-20170218092651.300722\Speech_LSTM_CTC@release_gpu/models/simple.dnn.1' 02/18/2017 09:28:13: Starting Epoch 2: learning rate per sample = 0.005000 effective momentum = 0.900000 momentum as time constant = 189.8 samples