-
Notifications
You must be signed in to change notification settings - Fork 1
Tutorials
If you want to run a quick test of the pipeline, simply execute:
python main.py +experiment=default_runThis command pulls all submodule parameters from the submodule folder and gets an experiment up and running. It will load a dataset and a model, train it, run some analysis and generate some visualizations. It’s a smooth way to check if everything in the pipeline is functioning as expected!
If training multiple models using worms with a certain number of named neurons or evaluating models with worms from a specific experimental dataset is your aim, let's dive into how that’s done!
Firstly, a validation dataset according to your preferences needs to be generated:
- Navigate to
configs/submodule/dataset.yaml - Tailor the parameters as you want and set
save_datasetstotrue. The pipeline will then store the validation and training datasets in the log folder.
The next step is to use your custom dataset:
- Open the
datasetfolder located within thelogfolder. - Move
validation_dataset.ptandval_dataset_info.csvtodata/datasets/your_custom_dataset, whereyour_custom_datasetis a custom dataset name of your choice. - When running your experiment, indicate the path to your custom dataset directory by specifying it in the
use_these_datasetsparameter inside thedataset.yamlconfig file. This tells the pipeline to load this validation dataset during the model training process, while using the worms specified in theexperimental_datasetsparameter for training.
Imagine training models using worms with, let’s say, 90 named neurons from our experimental datasets.
Here’s a step-by-step guide on how you’d do that, very similar to what we did before:
-
Run the Dataset Submodule: Execute it individually with the specifications you desire, e.g.,
num_named_neurons: 90. -
Utilize Your Train Dataset: Move
train_dataset.ptandtrain_dataset_info.csvinto a custom directory. -
Configuration: Specify the path to your custom train dataset directory using the
use_these_datasetsparameter inside thedataset.yamlconfig file. -
Run the Pipeline: Launch the pipeline. Now, models will be trained using your standard dataset, but validated using parameters specified at
experimental_datasets.
-
Quilee Simeon 📧 qsimeon@mit.edu
-
Leandro Risso Venâncio 📧 rissov.leandro@gmail.com
We appreciate all contributions and look forward to continuing this journey together.