Skip to content

zoom 20201201

stvoutsin edited this page Dec 1, 2020 · 16 revisions
  1. Zoom meeting 1st December

20201201 16:00 UTC

In progress

  1. Wiki page to plan notebooks
  2. Deploy a larger cluster to work with the full size data set
  3. User space ssh rsync access

New issues

  1. Configuration for Ansible deployment
  2. Experiment with scaling the Ansible deployment
  3. User accounts in Drupal
    • in progress 242
  4. Integration with IRIS IAM
    • in progress 243
  5. Resource booking in Drupal
    • in progress 244
  6. Automated testing for Kubernetes deployment
    • in progress 245
  7. Investigate IRIS echo S3 service for user data

New questions

User data space

  1. Simple implementation reserves 10G per user.
  2. Simple implementation for now - works for small number of users.
  3. Longer term - How do we recover unused space?
  4. Longer term - How do we handle dormant accounts?
  5. Longer term - Staging mechanism to push older data to an archive and recover unused space?

Spark version

  • Current live system is spark-2.7.
  • Zeppelin Hadoop-Yarn deploy is spark-2.7.
  • Kubernetes deploy is spark-3.x.
  • Nigel's Random Forrest example uses spark-2.7?
    • Does it need spark-2.7?
  • AXS distribution is based on spark-2.7.
    • Does it need spark-2.7 or can we create a spark-3.x version?

Do we stick with spark-2.7 or try to upgrade to spark-3.x.

  1. Are there issues with Zeppelin Hadoop Yarn deployment?
  2. Are there issues with getting AXS to work with spark-3.x?
  3. The Kubernetes deployment probably won't work with spark-2.x.

Questions about AXS

  1. Can we figure out how to apply AXS changes to a standard Spark distribution?
  2. What benefits does AXS give us?
  3. Can we create an example that demonstrates this?

AXS issues

  1. Differences between a standard distribution and the AXS distribution.
  2. Apply differences to add AXS to our deployment.
  3. Tests that demonstrate that AXS in installed and working
  4. Benchmark to compare performance of AXS augmented deployment

Actions

Create script that shows conversion from csv to parquet for Gaia, writing the results to the Ceph shares (stv) Test out multiple concurrent users running jobs via Zeppelin