Number of Partitions of Spark Dataframe -


can explain number of partitions created spark dataframe.

i know rdd, while creating can mention number of partitions below.

val rdd1 = sc.textfile("path" , 6)  

but spark dataframe while creating looks not have option specify number of partitions rdd.

only possibility think is, after creating dataframe can use repartition api.

df.repartition(4) 

so can please let me know if can specify number of partitions while creating dataframe.

you cannot, or @ least not in general case not different compared rdd. example textfile example code you've provides sets limit on minimum number of partitions.

in general:

  • datasets generated locally using methods range or todf on local collection use spark.default.parallelism.
  • datasets created rdd inherit number of partitions parent.
  • datsets created using data source api:

  • some data sources may provide additional options give more control on partitioning. example jdbc source allows set partitioning column, values range , desired number of partitions.

Comments

Popular posts from this blog

java - Jasper subreport showing only one entry from the JSON data source when embedded in the Title band -

mapreduce - Resource manager does not transit to active state from standby -

serialization - Convert Any type in scala to Array[Byte] and back -