Data Migration is like Migrate, but it will changes the data of your tables, not your schema. This gems creates a file where you can write code to make change along you schema changes.
Extracts the necessary files including a migrate that create a table used to keep you data migration version.
rails g data_migrater:installThen execute the migrations to create the table.
rake db:migrateNow we can create a data migration with name we want.
rails g data_migrater:create nameCheck your db/data_migrate folder, the data migration will be there.
Next time your application run, all pending data migration will be executed.
You can send your log to a file including the module DataMigrater::Logger.
class MyDataMigration
  include DataMigrater::Logger
  def execute
    logger.info "going to log/my_data_migration.log"
  end
endBy default, the class name is used and the file goes to log folder. You can change it:
class MyDataMigration
  include DataMigrater::Logger
  data_logger path: "db/data_migrate/log/global.log"
  def execute
    logger.info "going to db/data_migrate/log/global.log"
  end
end- dir: Directory where log will be;
- file: File name;
- path: Composition of- dir/filewhen you want give a fully qualified path. Default:- log/class_name.log.
You can parse CSV on the fly using the module DataMigrater::CSV.
class MyDataMigration
  include DataMigrater::CSV
  def execute
    csv.each do |line|
      Object.create! line
    end
  end
endBy default, the class name is used and the file is parsed from db/data_migrate/support/csv folder. You can change it:
class MyDataMigration
  include DataMigrater::CSV
  data_csv path: '/tmp/objects.csv'
  def execute
    csv.each do |line|
      Object.create! line
    end
  end
endYou can process a batch of items using the chunk_size:
class MyDataMigration
  include DataMigrater::CSV
  data_csv chunk_size: 2
  def execute
    # [
    #   { first_name: 'Washington', last_name: 'Botelho' },
    #   { first_name: 'Vanessa'   , last_name: 'Queiroz' }
    # ]
    csv.each { |line| Object.create line }
  end
endYou can rename the keys inside the iterated data using the option key_mapping:
class MyDataMigration
  include DataMigrater::CSV
  data_csv key_mapping: { first_name: :first }
  def execute
    csv.each do |line|
      Object.create! line # { first: 'Washington', last_name: 'Botelho' }
    end
  end
end- dir: Directory where CSV is located, by default- db/data_migrate/support/csv;
- file: File name, by default is the class name underscored:- my_data_migration.csv;
- path: Composition of- dir+- /+- filewhen you want give a fully qualified path.
- chunk_size: Batch parse size;
- key_mapping: Key name alias.
For more CSV options, check the project Smarter CSV:
You can download your CSV directly from Amazon S3 using the module DataMigrater::CSV with some configs. You must set provider as :s3 to activate S3 feature.
class MyDataMigration
  include DataMigrater::CSV
  data_csv bucket: 'my-bucket', provider: :s3
  def execute
    csv.each { |line| Object.create line }
  end
endBy default, when you use the S3 feature, the envs ACCESS_KEY_ID, REGION (default us-east-1) and SECRET_ACCESS_KEY will be used.
If you do not want export it globally and need to pass it inside you class, just declare de credentials options:
class MyDataMigration
  include DataMigrater::CSV
  data_csv provider: :s3, credentials: {
    access_key_id:     'foo',
    region:            'us-east-1',
    secret_access_key: 'bar'
  }
  def execute
    csv.each { |line| Object.create line }
  end
endYou can delete the S3 and local file from your migration after process the CSV.
class MyDataMigration
  include DataMigrater::CSV
  data_csv provider: :s3
  def execute
    csv.each { |line| Object.create line }
    csv_delete
  end
end- bucket: The bucket name. By default- data-migrater.
- credentials: AWS credentials:- access_key_id,- regionand- secret_access_key;
- provider:- :s3to indicate the S3 provider;
You can use ENV to prevent Data Migrater to run. By default, RAILS_ENV as test already prevents it.
DATA_MIGRATER=false rails s