But why ? Any specific reason behind it ? I am aware of that we can persist the dataframes but before proceeding would like to know the memory level of my DFs. I am working on performance tuning of my Spark jobs , looking for Storage Level APIs like RDDs.
Thanks, Divya On 6 May 2016 at 11:16, Ted Yu <yuzhih...@gmail.com> wrote: > I am afraid there is no such API. > > When persisting, you can specify StorageLevel : > > def persist(newLevel: StorageLevel): this.type = { > > Can you tell us your use case ? > > Thanks > > On Thu, May 5, 2016 at 8:06 PM, Divya Gehlot <divya.htco...@gmail.com> > wrote: > >> Hi, >> How can I get and set storage level for Dataframes like RDDs , >> as mentioned in following book links >> >> https://jaceklaskowski.gitbooks.io/mastering-apache-spark/content/spark-rdd-caching.html >> >> >> >> Thanks, >> Divya >> > >