Terrascope
  • Overview
  • Get started
  • Introduction
    • Terrascope Introduction
    • The Copernicus Programme
    • Registration and authentication
  • Data
    • Sentinel Missions
    • Sentinel-1
    • Sentinel-2
    • Sentinel-3
    • Sentinel-5P
    • PROBA-V mission
    • PROBA-V
    • SPOT-VGT mission
    • SPOT-VGT
    • Additional Products
  • APIs
    • catalogue APIs
    • OpenSearch
    • TerraCatalogueClient
    • STAC
    • Product download
    • Streamlined Data Access APIs
    • openEO
    • Additional Web-services
    • CropSAR Service
    • Web Map Service (WMS)
    • Web Map Tile Service (WMTS)
  • Tools
    • Terrascope GUI
    • Terrascope Viewer
    • openEO web editor
    • Virtual Environments
    • Virtual Machine
    • JupyterLab
    • Hadoop Cluster
    • EOplaza
    • Getting started
    • Manage your Organisation
    • Publish a Service
    • Execute a Service
    • Manage a Service
    • Reporting
  • Quotas and Limitations
  • Support
    • Contact
    • Terrascope Forum
    • Terrascope Sample Examples
  • FAQ
  1. Virtual Environments
  2. Hadoop Cluster
  3. Access Spark Logs
  • Terrascope GUI
    • Terrascope Viewer
    • openEO web editor
  • Virtual Environments
    • Virtual Machine
      • Terrascope Policies
    • JupyterLab
    • Hadoop Cluster
      • Manage Spark Resources
      • Advanced Kerberos
      • Access Spark Logs
      • Use Docker on Hadoop
      • Manage permissions and ownership
  • EOplaza
    • Getting started
    • Manage your Organisation
    • Publish a Service
    • Execute a Service
    • Manage a Service
      • Service Maturity
      • Credit Strength
    • Reporting

On this page

  • YARN ResourceManager UI
  • YARN logs command
  • YARN Spark Access Control Lists (ACLs)
  1. Virtual Environments
  2. Hadoop Cluster
  3. Access Spark Logs

Accessing Spark logs

Access to Spark logs is available for monitoring and troubleshooting Spark jobs. This page provides instructions for accessing Spark logs using the YARN ResourceManager UI and the yarn logs command. The YARN ResourceManager UI and the yarn logs command can only be used from the User Virtual Machine.

YARN ResourceManager UI

The YARN ResourceManager UI provides access to the Spark UI, which contains detailed information about the Spark job, including the DAG visualization, job statistics, and logs.

Access the Spark UI by clicking the ApplicationMaster link in the YARN ResourceManager UI. To access the YARN ResourceManager UI, visit https://epod-master1.vgt.vito.be:8090:

  • Use a Firefox web browser from the User Virtual Machine.
  • Have a valid Ticket Granting Ticket (TGT); instructions for creating a TGT can be found in Advanced Kerberos.
  • Once in the UI, select an application ID to view the details of an application and click on the ApplicationMaster link to get to the Spark UI.

Figure 1: YARN ResourceManager UI.

An improved UI is available at https://epod-master1.vgt.vito.be:8090/ui2 or https://epod-master2.vgt.vito.be:8090/ui2, depending on the current master node. This new UI offers an enhanced view of the resources used per queue, among other features.

Figure 2: Improved YARN ResourceManager UI.

YARN logs command

The YARN logs command can retrieve logs for a specific application ID. The application ID can be found in the YARN ResourceManager UI.

To get the logs, use the yarn logs command:


    yarn logs -applicationId <application_id>

You can also pipe the output to less to enable scrolling:


    yarn logs -applicationId <application_id> | less

The yarn logs command will output logs for different log types:

  • stdout: application logs printed to stdout
  • stderr: application logs printed to stderr
  • directory.info: prints the content of the container working directory
  • prelaunch.out, prelaunch.err, launch_container.sh: logs related to container start

Filter log types by adding the -log_files_pattern parameter, for example:


    yarn logs -applicationId <application_id> -log_files_pattern stderr | less

YARN Spark Access Control Lists (ACLs)

To allow other users or user groups to view Spark job logs they didn’t submit themselves (e.g. when running the Spark job as a service user), add the following configuration parameters to spark-submit:

These parameters set permissions for viewing logs of running and finished jobs, as well as modifying (including killing) the submitted job.


--conf spark.ui.view.acls=user1,user2 \
--conf spark.ui.view.acls.groups=group1 \
 
--conf spark.modify.acls=user1,user2
--conf spark.modify.acls.groups=group1
Back to top
Advanced Kerberos
Use Docker on Hadoop

Copyright 2018 - 2024 VITO NV All Rights reserved

 
  • Terms of Use

  • Privacy declaration

  • Cookie Policy

  • Contact