The Jobs page displays a list of all backup, restore, defreeze, and log request jobs. The Jobs page shows the progress of each File Server job. Job details are available for all jobs. On the job details page you can see the job details and the logs.
To access the File Servers Jobs page, perform the following tasks:
- Log in to the Hybrid Workloads Management Console.
- Select the workload from the Protect menu. Note that if the All Organizations menu is enabled, you have to first select an organization and then select the workload.
- In the navigation pane on the left, click Jobs. The right pane displays the following data:
|Job ID||A unique identification number associated with the job.|
|Job Type||The type of job operation, such as Backup, Restore, Defreeze, Log Request.|
|Backup Set||Name of the FS backup set for which the job was created.|
|Start time||The time when the job started.|
|End time||The time when the job ended.|
|Status||The status of the job. For more information, see Job Status.|
You can filter jobs by Job Type, Status, and Started in.
You can filter jobs by the following Job Types:
- Log Request
The Status filter allows you to filter jobs based on their statuses.
- Queued (1): The triggered jobs that are waiting to be processed.
Running (2): The jobs that are being processed.
During backup job for a File server, one of the following statuses is displayed:
The estimated data to be backed up and the total number of files scanned is displayed.
E.g: "472 MB estimated:6097 files scanned"
After the estimation is completed, the percentage of data uploaded to the cloud is displayed.
Successful (3): The jobs processed without any errors.
Successful with errors (4): The jobs that are processed but have some errors.
Failed (5): The jobs that were not processed successfully.
Note: A failed job is displayed with an error code. Hover on the error code to view the error message. The error code hyperlink redirects to the exact error message description on the documentation portal.
Canceled (6): The jobs that were canceled.
Waiting for Retry (7): The jobs that failed in the previous attempt and waiting to be processed again.
Note: By default, Druva retries a job twice with a ten-minute interval between the two attempts.
The Backup Retry feature ensures the following:
- Higher backup success rate, if a backup is failing due to transient issues in your environment.
- Optimum utilization of the scheduled window.
The backup retry for a job is performed only if a job fails. Druva does not perform backup retry for jobs that have successfully finished but with errors.
An automatic job retry is not triggered for a manually triggered job.
Skipped (8): The jobs that did not start within the scheduled window because another job is in progress.
Backup window expired (9): The list of jobs Druva could not complete because the entire data set was not backed up within the specified duration, and a recovery point was not created.
Scanning (13): The jobs that are still being scanned.
Scan failed (14): The jobs for which the scan has failed.
You can also filter jobs that Started in:
- Last 24 hours
- Last 7 days
- Last 1 month
- Custom date
Note: You can choose a specific date range.
The Job ID and the Backup Set values are clickable. Clicking the Job ID value for an FS job takes you to the job details page. Clicking the value in the Backup Set tab takes you to the backup set details page.
From the Jobs page within the Protect > File Servers page, click the Job ID of the job whose details you want to view. The jobs details page is divided into the Details tab, and the Logs tab.
The Details tab is divided into the following sections:
|Job Details||Data Estimation|
|Data Transfer||Environment Details|
The Job Details section has the following fields:
|Job Type||The type of job that was executed. This can be backup, restore, defreeze, or log request.|
|Backup Set Type||The type of server that was backed up. For File Servers, the backup set type is File.|
|Start Time||The time when the job started.|
|Pre Script Status||
The status of the execution of the pre-backup script. The status can be one of the following:
|Server||The file server on which the backup was initiated.|
|Backup Policy||The Backup Policy associated with the backup set.|
|End Time||The time when the job ended.|
|Post Script Status||
The status of the execution of the post-backup script.The status can be one of the following:
|Status||Completion status of the job.|
|Backup Content||The directories and folders that were backed up.|
|Scan Mode||The type of scan conducted which is either a Folder Walk, USN Journal Scan, or a Smart Scan.|
The data estimation section has the following fields:
|Source Data Scanned||The amount of data scanned at the source for the backup.|
|#Files Scanned||The total number of files scanned.|
|#Files Changed||The total number of files that changed as compared to the previous backup.|
|#Files Added||The total number of files that were added as compared to the previous backup.|
|#Files Removed||The total number of files that were removed as compared to the previous backup.|
|Time taken for estimation||The approximate time taken to scan the data. The time taken for estimation does not include network retry time.|
|File Size Distribution||
The percentage of the files backed up based on their sizes. The percentages are marked with the following file size ranges:
The Data Transfer section has the following details:
|Data Transferred to Cloud||The incremental data that is directly uploaded to Cloud after deduplication and compression.|
|Data Transferred to CloudCache||The incremental data that is directly uploaded to CloudCache after deduplication and compression.|
|Backup Duration||The total time taken to upload data to Cloud and CloudCache. Backup duration excludes the estimation time, network retry time, and the waiting for retry time.|
|Backup Speed||The rate at which the source data is scanned for backup.|
|Bandwidth Consumed||The bit rate to transfer data to cloud and CloudCache.|
The Environment Details section has the following details:
|Disk Read Rate||The bit rate for reading I/O by the agent. The Disk Read Rate is an average of the total data read and the total time taken to read data at various intervals for all drives where data is distributed. For UNC Shares, the Disk Read Rate is the average value of drive(s).|
|#Network Retries||The number of network retry attempts made within a job session.|
|Network Retry Duration||The total time spent in network retries. It is cumulative of all network session duration.|
Log files help analyze and troubleshoot issues encountered while performing a task. You can share the logs with Druva Technical Support to aid in issue resolution.
Displays the progress logs of the job.
The Detailed Logs section gives you the option to Request Detailed Logs. Detailed logs are available only after the upload job completes. Till then, the Download Detailed Logs button remains disabled. The logs are available for download till a certain date and time specified under the Available for download till field. The Requested on field shows you when the request for detailed logs was made.
Upload requested logs
The log files are used to analyze and troubleshoot the issues you might encounter while performing a task. For assistance in resolving the issues, you can share the log files with technical support. The job details are displayed in the Phoenix Management Console for the last 180 days.
Note: Process logs are only available for backup and restore jobs.
Detailed logs for jobs are available for Phoenix agent version 3.4 and later and GovCloud client version 4.0 and later. The detailed logs include the following:
|Common logs for Window, Linux, and backup proxy||Window logs||Linux logs||Backup proxy logs|
|Druva Phoenix Config||Window event/Application logs||Dmesg logs||VMware logs|
|Agent-specific Logs||VSS information||System information|
|Main Service logs|
The procedure varies depending on the Phoenix agent version you are using. If a job is executed with Phoenix agent version 3.4 to 4.5, then the Request Server Logs option will fetch the consolidated job logs available for that server. However, if a job is executed with Phoenix agent version 4.6 and later, the Request Job Logs option will fetch logs only for that particular job. You can request job logs within 30 days of triggering the job. You must download the requested logs within 7 days of triggering the request. The job details are available in the Phoenix Management Console for the last 180 days.
Note: If you execute a job on Phoenix agent 4.5 or earlier and then upgrade Phoenix agent to version 4.6 or later before downloading the logs, you will see the Request Job Logs option on the Detailed Logs tab of the Job Details page. However, clicking this button will still fetch the server logs for the job executed before the client upgrade.
You can request job logs within 30 days of triggering the job. You must download the requested logs within 7 days of triggering the request.If the log file is 4.5 MB or smaller, you can send it to technical support as an email attachment. If the log file exceeds 4.5 MB in size, perform the following tasks to send the logs to support:
- Go to https://upload.druva.com/
- Enter the case number in the Ticket Number field.
- Click Choose File, and add the compressed files to upload.
- Click Upload. Notify the support engineer that the logs have been uploaded on the portal by responding to the ongoing support ticket email.