@@ -93,7 +93,7 @@ def on_cluster_created(self, ray_cluster_handler):
93
93
_logger .warning (
94
94
"Registering Ray cluster spark job as background job failed. "
95
95
"You need to manually call `ray.util.spark.shutdown_ray_cluster()` "
96
- "before detaching your databricks notebook."
96
+ "before detaching your Databricks notebook."
97
97
)
98
98
99
99
auto_shutdown_minutes = float (
@@ -102,7 +102,7 @@ def on_cluster_created(self, ray_cluster_handler):
102
102
if auto_shutdown_minutes == 0 :
103
103
_logger .info (
104
104
"The Ray cluster will keep running until you manually detach the "
105
- "databricks notebook or call "
105
+ "Databricks notebook or call "
106
106
"`ray.util.spark.shutdown_ray_cluster()`."
107
107
)
108
108
return
@@ -117,26 +117,24 @@ def on_cluster_created(self, ray_cluster_handler):
117
117
db_api_entry .getIdleTimeMillisSinceLastNotebookExecution ()
118
118
except Exception :
119
119
_logger .warning (
120
- "Databricks `getIdleTimeMillisSinceLastNotebookExecution` API "
121
- "is unavailable, it is probably because that "
122
- "your current Databricks Runtime version does not support API "
123
- "`getIdleTimeMillisSinceLastNotebookExecution`, we cannot "
124
- "automatically shut down Ray cluster when databricks notebook "
125
- "is inactive, you need to manually detach databricks notebook "
120
+ "Failed to retrieve idle time since last notebook execution, "
121
+ "so that we cannot automatically shut down Ray cluster when "
122
+ "Databricks notebook is inactive for the specified minutes. "
123
+ "You need to manually detach Databricks notebook "
126
124
"or call `ray.util.spark.shutdown_ray_cluster()` to shut down "
127
125
"Ray cluster on spark."
128
126
)
129
127
return
130
128
131
129
_logger .info (
132
130
"The Ray cluster will be shut down automatically if you don't run "
133
- "commands on the databricks notebook for "
131
+ "commands on the Databricks notebook for "
134
132
f"{ auto_shutdown_minutes } minutes. You can change the "
135
- "automatically shutdown minutes by setting "
133
+ "auto- shutdown minutes by setting "
136
134
f"'{ DATABRICKS_RAY_ON_SPARK_AUTOSHUTDOWN_MINUTES } ' environment "
137
135
"variable, setting it to 0 means that the Ray cluster keeps running "
138
136
"until you manually call `ray.util.spark.shutdown_ray_cluster()` or "
139
- "detach databricks notebook."
137
+ "detach Databricks notebook."
140
138
)
141
139
142
140
def auto_shutdown_watcher ():
0 commit comments