Ansible role for deploying spark-operator.
It is available also on Ansible Galaxy (link).
ansible-galaxy install jiri_kremser.spark_operatorAnd then use it in your playbook (example-playbook.yaml):
- hosts: localhost
  roles:
    - role: jiri_kremser.spark_operatoror with your own settings (these are the default settings):
- hosts: localhost
  roles:
    - role: jiri_kremser.spark_operator
      namespace: myproject
      image: quay.io/radanalyticsio/spark-operator:latest-released
      crd: false
      prometheus: false
      runExampleCluster: falseansible-playbook example-playbook.yamlor with this one liner:
ansible localhost -m include_role -a name=jiri_kremser.spark_operatorAfter running the playbook, the operator should be up and running and you can continue with creating either config maps or custom resources for your Spark clusters or Spark applications. For more details please consult the readme of spark-operator (you can skip the operator deployment step) or check the examples.
ansible-playbook -i tests/inventory tests/test-playbook.yml