The All Jobs page displays a list of all backup, restore, defreeze, and log request jobs. You can monitor the progress of each File Server job. Additionally, job details are accessible for all listed tasks. Upon navigating to the job details page, you can delve into specific job details and review associated logs.
To access the File Servers All Jobs page, perform the following tasks:
Log in to the Management Console.
Select the workload from the Protect menu. Note that if the All Organizations menu is enabled, you have to first select an organization and then select the workload.
In the navigation pane on the left, click Jobs. The right pane displays the following data:
Field | Description |
Job ID | A unique identification number associated with the job. |
Job Type | The type of job operation, such as Backup, Restore, Defreeze, Log Request. |
Backup Set | Name of the FS backup set for which the job was created. |
Start time | The time when the job started. |
End time | The time when the job ended. |
Status | The status of the job. For more information, see Job Status. |
Filters
You can filter jobs by Job Type, Status, and Started in.
Job Type
You can filter jobs by the following Job Types:
Backup
Restore
Defreeze
Log Request
You can filter jobs by Status:
Status
The Status filter allows you to filter jobs based on their statuses.
Queued (1): The triggered jobs that are waiting to be processed.
Running (2): The jobs that are being processed.
โ
During backup job for a File server, one of the following statuses is displayed:The estimated data to be backed up and the total number of files scanned is displayed.
E.g: "472 MB estimated:6097 files scanned"After the estimation is completed, the percentage of data uploaded to the cloud is displayed.
Successful (3): The jobs processed without any errors.
Successful with errors (4): The jobs that are processed but have some errors.
Failed (5): The jobs that were not processed successfully.
Note: A failed job is displayed with an error code. Hover on the error code to view the error message. The error code hyperlink redirects to the exact error message description on the documentation portal.
Canceled (6): The jobs that were canceled.
Waiting for Retry (7): The jobs that failed in the previous attempt and waiting to be processed again.
Note: By default, Druva retries a job twice with a ten-minute interval between the two attempts.
The Backup Retry feature ensures the following:
Higher backup success rate, if a backup is failing due to transient issues in your environment.
Optimum utilization of the scheduled window.
The backup retry for a job is performed only if a job fails. Druva does not perform backup retry for jobs that have successfully finished but with errors.
An automatic job retry is not triggered for a manually triggered job.
Skipped (8): The jobs that did not start within the scheduled window because another job is in progress.
Backup window expired (9): The list of jobs Druva could not complete because the entire data set was not backed up within the specified duration, and a recovery point was not created.
Scanning (13): The jobs that are still being scanned.
Scan failed (14): The jobs for which the scan has failed.
You can also filter jobs that Started in:
Last 24 hours
Last 7 days
Last 1 month
Custom dateNote: You can choose a specific date range.
The Job ID and the Backup Set values are clickable. Clicking the Job ID value for an FS job takes you to the job details page. Clicking the value in the Backup Set tab takes you to the backup set details page.
From the Jobs page within the Protect > File Servers page, click the Job ID of the job whose details you want to view. The jobs details page is divided into the Details tab, and the Logs tab.
Job Details
The Job Details section has the following fields:
Field | Description |
Job Type | The type of job that was executed. This can be backup, restore, defreeze, or log request. |
Backup Set Type | The type of server that was backed up. For File Servers, the backup set type is File. |
Start Time | The time when the job started. |
Pre Script Status | The status of the execution of the pre-backup script. The status can be one of the following:
|
Server | The file server on which the backup was initiated. |
Backup Policy | The Backup Policy associated with the backup set. |
End Time | The time when the job ended. |
Post Script Status | The status of the execution of the post-backup script.The status can be one of the following:
|
Status | Completion status of the job. |
Backup Content | The directories and folders that were backed up. |
Scan Mode | The type of scan conducted which is either a Folder Walk, or a USN Journal Scan. |
Data Estimation
The data estimation section has the following fields:
Field | Description |
Source Data Scanned | The amount of data scanned at the source for the backup. |
#Files Scanned | The total number of files scanned. |
#Files Changed | The total number of files that changed as compared to the previous backup. |
#Files Added | The total number of files that were added as compared to the previous backup. |
#Files Removed | The total number of files that were removed as compared to the previous backup. |
Time taken for estimation | The approximate time taken to scan the data. The time taken for estimation does not include network retry time. |
File Size Distribution | The percentage of the files backed up based on their sizes. The percentages are marked with the following file size ranges:
|
Data Transfer
The Data Transfer section has the following details:
Field | Description |
Data Transferred to Cloud | The incremental data that is directly uploaded to Cloud after deduplication and compression. |
Data Transferred to CloudCache | The incremental data that is directly uploaded to CloudCache after deduplication and compression. |
Backup Duration | The total time taken to upload data to Cloud and CloudCache. Backup duration excludes the estimation time, network retry time, and the waiting for retry time. |
Backup Speed | The rate at which the source data is scanned for backup. |
Bandwidth Consumed | The bit rate to transfer data to cloud and CloudCache. |
Environment Details
The Environment Details section has the following details:
Field | Description |
Disk Read Rate | The bit rate for reading I/O by the agent. The Disk Read Rate is an average of the total data read and the total time taken to read data at various intervals for all drives where data is distributed. For UNC Shares, the Disk Read Rate is the average value of drive(s). |
#Network Retries | The number of network retry attempts made within a job session. |
Network Retry Duration | The total time spent in network retries. It is cumulative of all network session duration. |
Monitor file server logs
Log files help analyze and troubleshoot issues encountered while performing a task. You can share the logs with Druva Technical Support to aid in issue resolution.
Progress Logs
Displays the progress logs of the job.
Detailed Logs
The Detailed Logs section gives you the option to Request Detailed Logs. Detailed logs are available only after the upload job completes. Till then, the Download Detailed Logs button remains disabled. The logs are available for download till a certain date and time specified under the Available for download till field. The Requested on field shows you when the request for detailed logs was made.
Upload requested logs
The log files are used to analyze and troubleshoot the issues you might encounter while performing a task. For assistance in resolving the issues, you can share the log files with technical support. The job details are displayed in the Management Console for the last 180 days.
Note: Process logs are only available for backup and restore jobs.
Detailed logs for jobs are available for Hybrid Workloads agent / Enterprise Workloads agent. The detailed logs include the following:
Common logs for Window, Linux, and backup proxy | Window logs | Linux logs | Backup proxy logs |
Druva Phoenix Config | Window event/Application logs | Dmesg logs | VMware logs |
Agent-specific Logs | VSS information | System information | |
Main Service logs |
You can request job logs within 30 days of triggering the job. You must download the requested logs within 7 days of triggering the request. The job details are available in the Management Console for the last 180 days.
Note: If you execute a job on Hybrid Workloads agent (version prior to 7.0.0) and then upgrade to Enterprise Workloads agent (version 7.0.0 and later) before downloading the logs, you will see the Request Job Logs option on the Detailed Logs tab of the Job Details page. However, clicking this button will still fetch the server logs for the job executed before the client upgrade.
You can request job logs within 30 days of triggering the job. You must download the requested logs within 7 days of triggering the request.If the log file is 4.5 MB or smaller, you can send it to technical support as an email attachment. If the log file exceeds 4.5 MB in size, perform the following tasks to send the logs to support:
Enter the case number in the Ticket Number field.
Click Choose File, and add the compressed files to upload.
Click Upload. Notify the support engineer that the logs have been uploaded on the portal by responding to the ongoing support ticket email.