Skip to content

Releases: dataflint/spark

Version 0.2.6

06 Nov 15:05
Compare
Choose a tag to compare
  • Nvidia RAPIDS support
  • Supporting proxy servers before spark/history server better
  • bug fixes

Version 0.2.5

28 Oct 09:16
Compare
Choose a tag to compare

Fix "Return to History Server" button when history server is behind a proxy

Full Changelog: v0.2.4...v0.2.5

Version 0.2.4

15 Oct 15:48
Compare
Choose a tag to compare
  • Added driver memory usage visibility & alerts
  • Update README
  • bug fixes

Version 0.2.3

13 Aug 10:48
Compare
Choose a tag to compare
  • New alert - Large data Broadcast, for requesting to broadcast large data sets with the broadcast() function
  • New alert - Large filter conditions, for wiring long filter conditions instead of using join logic
  • UI Improvements

Version 0.2.2

04 Jun 12:43
Compare
Choose a tag to compare

Support spark versions 2.4 logs in history server with version later than 3.2
Limited feature-set is available due to events having less data than spark 3.0 and up

Version 0.2.1

03 Jun 16:56
Compare
Choose a tag to compare
  1. Better Databricks stage to node support
  2. Support spark.dataflint.runId in custom history server providers when appId is not the spark appId

Version 0.2.0

20 May 13:09
Compare
Choose a tag to compare
  • Better support for Databricks Photon plans
  • Input nodes shows partitions filters and push down filters
  • Stage Breakdown - press the blue down arrow on sql node to see stage information
  • New alert - large number of small tasks

Version 0.1.7

17 Apr 17:03
Compare
Choose a tag to compare

Apache Iceberg alerts improvements

Add avg file size in read/write

More information when hovering on stage

Version 0.1.6

26 Mar 11:06
Compare
Choose a tag to compare

Apache Iceberg support:

  1. Better node naming
  2. Read metrics and reading small files alerts
  3. Write metrics and overwriting most of table alerts
    Write metrics require enabling an iceberg metric reporter, can be done for you by enabling spark.dataflint.iceberg.autoCatalogDiscovery to true, or setting the iceberg metric reporter manually for each catalog, for example:
    spark.sql.catalog.[catalog name].metrics-reporter-impl org.apache.spark.dataflint.iceberg.DataflintIcebergMetricsReporter

Version 0.1.5

24 Feb 19:48
Compare
Choose a tag to compare
  • Add support for history server with cluster-mode jobs (i.e. with attempt numbet)
  • Fix "wasted cores" calculation
  • Fix status tab SQL is flickering when there is SQL with sub queriers