@@ -34,24 +34,24 @@ class PySparkException(Exception):
3434 def __init__ (
3535 self ,
3636 message : Optional [str ] = None ,
37- error_class : Optional [str ] = None ,
38- message_parameters : Optional [Dict [str , str ]] = None ,
39- query_contexts : Optional [List ["QueryContext" ]] = None ,
37+ errorClass : Optional [str ] = None ,
38+ messageParameters : Optional [Dict [str , str ]] = None ,
39+ contexts : Optional [List ["QueryContext" ]] = None ,
4040 ):
41- if query_contexts is None :
42- query_contexts = []
41+ if contexts is None :
42+ contexts = []
4343 self ._error_reader = ErrorClassesReader ()
4444
4545 if message is None :
4646 self ._message = self ._error_reader .get_error_message (
47- cast (str , error_class ), cast (Dict [str , str ], message_parameters )
47+ cast (str , errorClass ), cast (Dict [str , str ], messageParameters )
4848 )
4949 else :
5050 self ._message = message
5151
52- self ._error_class = error_class
53- self ._message_parameters = message_parameters
54- self ._query_contexts = query_contexts
52+ self ._errorClass = errorClass
53+ self ._messageParameters = messageParameters
54+ self ._contexts = contexts
5555
5656 def getErrorClass (self ) -> Optional [str ]:
5757 """
@@ -66,7 +66,7 @@ def getErrorClass(self) -> Optional[str]:
6666 :meth:`PySparkException.getQueryContext`
6767 :meth:`PySparkException.getSqlState`
6868 """
69- return self ._error_class
69+ return self ._errorClass
7070
7171 def getMessageParameters (self ) -> Optional [Dict [str , str ]]:
7272 """
@@ -81,7 +81,7 @@ def getMessageParameters(self) -> Optional[Dict[str, str]]:
8181 :meth:`PySparkException.getQueryContext`
8282 :meth:`PySparkException.getSqlState`
8383 """
84- return self ._message_parameters
84+ return self ._messageParameters
8585
8686 def getSqlState (self ) -> Optional [str ]:
8787 """
@@ -128,28 +128,28 @@ def getQueryContext(self) -> List["QueryContext"]:
128128 :meth:`PySparkException.getMessage`
129129 :meth:`PySparkException.getSqlState`
130130 """
131- return self ._query_contexts
131+ return self ._contexts
132132
133133 def _log_exception (self ) -> None :
134- query_contexts = self .getQueryContext ()
135- query_context = query_contexts [0 ] if len (query_contexts ) != 0 else None
136- if query_context :
137- if query_context .contextType ().name == "DataFrame" :
134+ contexts = self .getQueryContext ()
135+ context = contexts [0 ] if len (contexts ) != 0 else None
136+ if context :
137+ if context .contextType ().name == "DataFrame" :
138138 logger = PySparkLogger .getLogger ("DataFrameQueryContextLogger" )
139- call_site = query_context .callSite ().split (":" )
139+ call_site = context .callSite ().split (":" )
140140 line = call_site [1 ] if len (call_site ) == 2 else ""
141141 logger .exception (
142142 self .getMessage (),
143143 file = call_site [0 ],
144144 line = line ,
145- fragment = query_context .fragment (),
146- error_class = self .getErrorClass (),
145+ fragment = context .fragment (),
146+ errorClass = self .getErrorClass (),
147147 )
148148 else :
149149 logger = PySparkLogger .getLogger ("SQLQueryContextLogger" )
150150 logger .exception (
151151 self .getMessage (),
152- error_class = self .getErrorClass (),
152+ errorClass = self .getErrorClass (),
153153 )
154154
155155 def __str__ (self ) -> str :
@@ -299,11 +299,11 @@ class PySparkAssertionError(PySparkException, AssertionError):
299299 def __init__ (
300300 self ,
301301 message : Optional [str ] = None ,
302- error_class : Optional [str ] = None ,
303- message_parameters : Optional [Dict [str , str ]] = None ,
302+ errorClass : Optional [str ] = None ,
303+ messageParameters : Optional [Dict [str , str ]] = None ,
304304 data : Optional [Iterable ["Row" ]] = None ,
305305 ):
306- super ().__init__ (message , error_class , message_parameters )
306+ super ().__init__ (message , errorClass , messageParameters )
307307 self .data = data
308308
309309
0 commit comments