2525else :
2626 from itertools import imap as map
2727
28- from py4j .java_collections import ListConverter , MapConverter
29-
3028from pyspark .context import SparkContext
3129from pyspark .rdd import RDD , _load_from_socket , ignore_unicode_prefix
3230from pyspark .serializers import BatchedSerializer , PickleSerializer , UTF8Deserializer
@@ -186,9 +184,7 @@ def saveAsTable(self, tableName, source=None, mode="error", **options):
186184 source = self .sql_ctx .getConf ("spark.sql.sources.default" ,
187185 "org.apache.spark.sql.parquet" )
188186 jmode = self ._java_save_mode (mode )
189- joptions = MapConverter ().convert (options ,
190- self .sql_ctx ._sc ._gateway ._gateway_client )
191- self ._jdf .saveAsTable (tableName , source , jmode , joptions )
187+ self ._jdf .saveAsTable (tableName , source , jmode , options )
192188
193189 def save (self , path = None , source = None , mode = "error" , ** options ):
194190 """Saves the contents of the :class:`DataFrame` to a data source.
@@ -211,9 +207,7 @@ def save(self, path=None, source=None, mode="error", **options):
211207 source = self .sql_ctx .getConf ("spark.sql.sources.default" ,
212208 "org.apache.spark.sql.parquet" )
213209 jmode = self ._java_save_mode (mode )
214- joptions = MapConverter ().convert (options ,
215- self ._sc ._gateway ._gateway_client )
216- self ._jdf .save (source , jmode , joptions )
210+ self ._jdf .save (source , jmode , options )
217211
218212 @property
219213 def schema (self ):
@@ -819,7 +813,6 @@ def fillna(self, value, subset=None):
819813 value = float (value )
820814
821815 if isinstance (value , dict ):
822- value = MapConverter ().convert (value , self .sql_ctx ._sc ._gateway ._gateway_client )
823816 return DataFrame (self ._jdf .na ().fill (value ), self .sql_ctx )
824817 elif subset is None :
825818 return DataFrame (self ._jdf .na ().fill (value ), self .sql_ctx )
@@ -932,9 +925,7 @@ def agg(self, *exprs):
932925 """
933926 assert exprs , "exprs should not be empty"
934927 if len (exprs ) == 1 and isinstance (exprs [0 ], dict ):
935- jmap = MapConverter ().convert (exprs [0 ],
936- self .sql_ctx ._sc ._gateway ._gateway_client )
937- jdf = self ._jdf .agg (jmap )
928+ jdf = self ._jdf .agg (exprs [0 ])
938929 else :
939930 # Columns
940931 assert all (isinstance (c , Column ) for c in exprs ), "all exprs should be Column"
@@ -1040,8 +1031,7 @@ def _to_seq(sc, cols, converter=None):
10401031 """
10411032 if converter :
10421033 cols = [converter (c ) for c in cols ]
1043- jcols = ListConverter ().convert (cols , sc ._gateway ._gateway_client )
1044- return sc ._jvm .PythonUtils .toSeq (jcols )
1034+ return sc ._jvm .PythonUtils .toSeq (cols )
10451035
10461036
10471037def _unary_op (name , doc = "unary operator" ):
0 commit comments