Skip to content
Merged
Show file tree
Hide file tree
Changes from 1 commit
Commits
Show all changes
53 commits
Select commit Hold shift + click to select a range
214e1cc
ResourceConfigField -> ResourceField
1vn May 5, 2019
719ed13
go transformer changes
1vn May 6, 2019
b26eccc
Merge branch 'master' into spec-optional
1vn May 7, 2019
5c835c2
comment out error stack printing
1vn May 7, 2019
b54ffd3
format
1vn May 7, 2019
162bc86
simplify fetch_metadata
1vn May 7, 2019
b8db138
clean up
1vn May 8, 2019
a35f089
fix transform spark bug and training_columns
1vn May 8, 2019
fbf9f56
format
1vn May 8, 2019
78e0095
fix url
1vn May 8, 2019
09992cc
unfix url
1vn May 8, 2019
aa9e900
Merge branch 'spec-optional' of github.com:cortexlabs/cortex into spe…
1vn May 8, 2019
0f6954d
merge master
1vn May 9, 2019
36c5fcf
address some comments
1vn May 10, 2019
6a3a9bd
check 5 samples
1vn May 10, 2019
c5f760d
remove skip validation
1vn May 13, 2019
0bd6b73
remove outdated check
1vn May 13, 2019
081b9b3
add AlphaNumericDashDotUnderscoreEmpty
1vn May 13, 2019
7fedb6f
address some comments - lowercase resourceFields and replace error
1vn May 13, 2019
9445688
cache anon specs
1vn May 13, 2019
14e45ff
fix autogen
1vn May 13, 2019
03a22bc
progress
1vn May 14, 2019
c4a04bb
lazy load metadata
1vn May 15, 2019
cd38e52
clean up spark
1vn May 15, 2019
909a001
format and lint
1vn May 15, 2019
8c93f4c
don't skip cast
1vn May 16, 2019
106fd58
default to None
1vn May 16, 2019
28fcdc8
fix context
1vn May 16, 2019
c5504ac
format
1vn May 16, 2019
5271c3f
fix tests
1vn May 16, 2019
19a69bb
fix test
1vn May 16, 2019
ad9d8b3
address some python comments
1vn May 16, 2019
b6d4337
format
1vn May 16, 2019
181da9a
add test, remove camel case
1vn May 17, 2019
487cf40
Merge branch 'master' into spec-optional
1vn May 17, 2019
0f7c363
clean up type checking logic, formatting
1vn May 17, 2019
076c020
remove more camel case
1vn May 17, 2019
486c026
fix more camel case
1vn May 17, 2019
79657c4
address some comments
1vn May 17, 2019
471e98f
refactor metadata
1vn May 18, 2019
2d5602c
use raw_dataset key
1vn May 18, 2019
8c7fa9b
address some comments
1vn May 21, 2019
e493684
move type inference to validate_trannsformer
1vn May 21, 2019
c1551c0
validate inferred types from transform spark and python
1vn May 21, 2019
8abc9dc
pass type downstream
1vn May 21, 2019
b36218e
add comment about transform_python
1vn May 21, 2019
5a7bdd3
remove unused
1vn May 21, 2019
7583e39
Merge branch 'master' into spec-optional
1vn May 21, 2019
2aa5a99
address comments
1vn May 21, 2019
95d98e1
format
1vn May 21, 2019
694b08d
wrap more code in try
1vn May 21, 2019
5b5f0b6
move type check before value check
1vn May 21, 2019
b2fd9d7
address comments
1vn May 21, 2019
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
progress
  • Loading branch information
1vn committed May 14, 2019
commit 03a22bc2a82a626993b4433d231612dca16bdcf0
4 changes: 2 additions & 2 deletions pkg/operator/context/models.go
Original file line number Diff line number Diff line change
Expand Up @@ -85,7 +85,6 @@ func getModels(
datasetIDWithTags := hash.Bytes(buf.Bytes())

datasetRoot := filepath.Join(root, consts.TrainingDataDir, datasetID)

trainingDatasetName := strings.Join([]string{
modelConfig.Name,
resource.TrainingDatasetType.String(),
Expand All @@ -97,7 +96,7 @@ func getModels(
ID: modelID,
IDWithTags: modelID,
ResourceType: resource.ModelType,
MetadataKey: filepath.Join(datasetRoot, "metadata.json"),
MetadataKey: filepath.Join(root, consts.ModelsDir, modelID+"_metadata.json"),
},
},
Model: modelConfig,
Expand All @@ -115,6 +114,7 @@ func getModels(
ID: datasetID,
IDWithTags: datasetIDWithTags,
ResourceType: resource.TrainingDatasetType,
MetadataKey: filepath.Join(datasetRoot, "metadata.json"),
},
},
ModelName: modelConfig.Name,
Expand Down
18 changes: 18 additions & 0 deletions pkg/workloads/lib/context.py
Original file line number Diff line number Diff line change
Expand Up @@ -479,6 +479,17 @@ def update_metadata(self, metadata, context_key, context_item=""):
self.ctx[context_key][context_item]["metadata"] = metadata
self.storage.put_json(metadata, self.ctx[context_key][context_item]["metadata_key"])

def get_metadata(self, context_key, context_item, use_cache=True):
if use_cache and self.ctx[context_key][context_item]["metadata"]:
return self.ctx[context_key][context_item]["metadata"]

metadata_uri = self.ctx[context_key][context_item]["metadata_key"]
metadata = self.storage.get_json(metadata_uri, allow_missing=True)
self.ctx[context_key][context_item]["metadata"] = metadata
return metadata



def fetch_metadata(self):
resources = [
"python_packages",
Expand All @@ -499,6 +510,13 @@ def fetch_metadata(self):
metadata = {}
self.ctx[resource][k]["metadata"] = metadata

# fetch dataset metadata for models
for k, v in self.ctx["models"].items():
metadata = self.storage.get_json(v["dataset"]["metadata_key"], allow_missing=True)
if not metadata:
metadata = {}
self.ctx["models"][k]["dataset"]["metadata"] = metadata

metadata = self.storage.get_json(self.raw_dataset["metadata_key"], allow_missing=True)
if not metadata:
metadata = {}
Expand Down
5 changes: 3 additions & 2 deletions pkg/workloads/tf_train/train_util.py
Original file line number Diff line number Diff line change
Expand Up @@ -149,9 +149,10 @@ def train(model_name, model_impl, ctx, model_dir):
exporter = tf.estimator.FinalExporter("estimator", serving_input_fn, as_text=False)

train_num_steps = model["training"]["num_steps"]
dataset_metadata = model["dataset"]["metadata"]
if model["training"]["num_epochs"]:
train_num_steps = (
math.ceil(model["metadata"]["training_size"] / float(model["training"]["batch_size"]))
math.ceil(dataset_metadata["training_size"] / float(model["training"]["batch_size"]))
* model["training"]["num_epochs"]
)

Expand All @@ -160,7 +161,7 @@ def train(model_name, model_impl, ctx, model_dir):
eval_num_steps = model["evaluation"]["num_steps"]
if model["evaluation"]["num_epochs"]:
eval_num_steps = (
math.ceil(model["metadata"]["training_size"] / float(model["evaluation"]["batch_size"]))
math.ceil(model["metadata"]["eval_size"] / float(model["evaluation"]["batch_size"]))
* model["evaluation"]["num_epochs"]
)

Expand Down