Releases: dataflint/spark
Releases · dataflint/spark
Version 0.2.6
- Nvidia RAPIDS support
- Supporting proxy servers before spark/history server better
- bug fixes
Version 0.2.5
Fix "Return to History Server" button when history server is behind a proxy
Full Changelog: v0.2.4...v0.2.5
Version 0.2.4
- Added driver memory usage visibility & alerts
- Update README
- bug fixes
Version 0.2.3
- New alert - Large data Broadcast, for requesting to broadcast large data sets with the broadcast() function
- New alert - Large filter conditions, for wiring long filter conditions instead of using join logic
- UI Improvements
Version 0.2.2
Support spark versions 2.4 logs in history server with version later than 3.2
Limited feature-set is available due to events having less data than spark 3.0 and up
Version 0.2.1
- Better Databricks stage to node support
- Support spark.dataflint.runId in custom history server providers when appId is not the spark appId
Version 0.2.0
- Better support for Databricks Photon plans
- Input nodes shows partitions filters and push down filters
- Stage Breakdown - press the blue down arrow on sql node to see stage information
- New alert - large number of small tasks
Version 0.1.7
Apache Iceberg alerts improvements
Add avg file size in read/write
More information when hovering on stage
Version 0.1.6
Apache Iceberg support:
- Better node naming
- Read metrics and reading small files alerts
- Write metrics and overwriting most of table alerts
Write metrics require enabling an iceberg metric reporter, can be done for you by enabling spark.dataflint.iceberg.autoCatalogDiscovery to true, or setting the iceberg metric reporter manually for each catalog, for example:
spark.sql.catalog.[catalog name].metrics-reporter-impl org.apache.spark.dataflint.iceberg.DataflintIcebergMetricsReporter
Version 0.1.5
- Add support for history server with cluster-mode jobs (i.e. with attempt numbet)
- Fix "wasted cores" calculation
- Fix status tab SQL is flickering when there is SQL with sub queriers