79631307

Date: 2025-05-20 23:39:31
Score: 0.5
Natty:
Report link

The known issue is described here:

https://learn.microsoft.com/en-us/azure/hdinsight/hdinsight-known-issues

Spark History Service => Decommissioned node logs cannot be accessed directly from Spark / YARN UI (Expected behavior)

This issue can be very bothersome. A Spark job that has executed recently (in the past hour) may stop presenting its logs in the Spark U/I.  I think the bug needs to be fixed on priority.

In the meantime here are a few alternative approaches that the PG suggested for customers to use:

Alternative #1: Manually construct the URL to the Job History to access the decommissioned aggregated logs.

Example:

https://<CLUSTERDNSNAME>.azurehdinsight.net/yarnui/jobhistory/logs/<Decommissioned worker node FQDN>/port/30050/<CONTAINER-ID>/<CONTAINER-ID>/root/stderr?start=-4096

Alternative #2: Use the schedule-based autoscaling workflow. This allows developers time to debug job failures before the cluster scales down.

Alternative #3: Use the yarn logs command via the Azure CLI.

Alternative #4: Use an open-source converter to translate TFile-formatted logs in the Azure Storage account to plain text

Reasons:
  • Long answer (-1):
  • No code block (0.5):
  • Contains question mark (0.5):
  • Self-answer (0.5):
Posted by: David Beavon