Please enter search query.
Search <product_name> all support & community content...
Article: 100045147
Last Published: 2025-07-24
Ratings: 0 1
Product(s): Data Insight
Problem
Data Insight (DI) runs tasks in the concept of jobs on a scheduled basis to move and process data between worker nodes in the DI environment.
Cause
Jobs can be listed using the command: <INSTALLDIR>\Program Files\DataInsight\bin\configcli.exe list_jobs
A job can be run via the CLI using the following command: <INSTALLDIR>\Program Files\DataInsight\bin\configcli.exe execute_job <JobName>
Note: JobName is case-sensitive
Example:
C:\Program Files\DataInsight\bin\configcli execute_job IndexWriterJob
Result expected:
Job started
Solution
DataInsightComm Service Jobs
Job | Description | Default Schedule | Node | Files Processed/Updated | Comments |
ADScanJob | Scans the directory services (AD, LDAP, NIS). Updates users.db | 3:00:00 AM - Daily | MS | Processes users.db and _local_users.zip files | |
ADScanJob_watchlist | Updates information and details about watchlisted users | On Demand | MS | ||
ActivityIndexJob | 7:00:00 AM - Daily | Indexer | Updates dir-activity.idx.### and file-activity.idx.#### for each MSU | Allows faster computation of data ownership | |
CancelClassifyRequestJob | Fetches the list of classification requests that are cancelled and distributes this request between Information Governance nodes | Every 20 seconds | Classification | Before classifying files, all the classification jobs consult this list to identify the requests that are marked for cancelation. If they observe any canceled request in the new request that is submitted for classification, then that request is deleted | |
ChangeLogJob | Merges changelog files generated by collector.exe from collector\changelog\inbox into a main device-specific changelog file. This file is later used by incremental scanner | Every 1 hour | Collector | Processes <NODE ID>_<timestamp>_changelog.db files | |
ClassifyBatchJob | Splits classification batch input databases for consumption by the scanner on the collector | Every 1 minute | Indexer | ||
ClassifyIndexJob | Updates index databse for each MSU with classifcation tabs and updates the bookkeeping DB | Every 1 minute | Indexer | ||
ClassifyInputJob | Processes the classification requests form the Information Governance Mangement Console and from reports for the consumption by the bookkeeping DB | Every 10 seconds | MS | ||
ClassifyMergeStatusJob | Pulls classification update status files from indexer and updates the global bookkeeping DB which is used to display high level classification status in the console | Every 10 seconds | MS | ||
CloudDeviceAuditJob | Collects audit data from configured Box account | Every 15 minutes | Collector | ||
CloudDeviceAuditJob_amazon_s3 | Collects audit data from AmazonS3 account | Every 10 minutes | Collector | ||
CloudDeviceAuditJob_onedrive | Collects the audit records for OneDrive accounts | Every 30 minutes | Collector | ||
CloudDeviceAuditJob_sponline | Collects the SharePoint Online site collection audit data | Every 30 minutes | Collector | ||
CloudUtilJob | Creates user to email mapping for Box device | Once every day | All | ||
CollectorJob | Initiates collector.exe to process raw event files and generates audit files | Every 2 hours | Collector | Input: fpolicy, cee, winnas*.sqlite files File Format: fpolicy_<timestamp>_<gen>_<DEVICE ID>.sqlite Output: audit_sqlite files File Format: audit_cifs_<timestamp>_<shareId> |
|
CollectorJob_size | Initiates collector.exe runs for a device if the input size for the device exceeds the batch_size limit | Every 10 minutes | Collector | Input: fpolicy, cee, winnas*.sqlite files File Format: fpolicy_<timestamp>_<gen>_<DEVICE ID>.sqlite Output: audit_sqlite files File Format: audit_cifs_<timestamp>_<shareId> |
Runs more frequently than the regular CollectorJob, so that files for an active device are processed as soon as enough input is available |
CompactJob | Compress the attic folder and err folders in <DATADIR>\collector, <DATADIR>\scanner, and <DATADIR>\indexer | 5:00:00 AM - Daily | All | Uses the Windows compression feature to set the "compression" attribute for the folders | |
Compact_Job_Report | Compresses completed reports under <DATADIR>\console\reports\reportruns folder | Every 1 hour | MS | ||
ConsumeUnresolvedEntitiesJob | Resolves all unresolved Users and Groups that are fetched from the configured devices based on the latest user scan information | 9:00:00 AM & 9:00:00 PM - Daily | |||
ControlPointJob | Triggers controlpoint.exe for each share on the indexer and generates tag files | 12:00:00 PM - Saturday | Indexer | Generates tags file in inbox with the format <MSU ID>_timestamp_tags.sqlite Generated tags files are consumed up by the next IndexWriterJob to update MSU DB |
|
CreateWorkflowDBJob | Creates the database containing the data for DLP Incident Management, Entitlement Review , and Ownership Confirmation workflows based on the input provided by users | Every 1 minute | MS | ||
DataRetentionJob | Enforces data retention policies which include archiving old index segments, deleting old segments, indexes for deleted objects, old system events, and old alerts | 12:00:00 AM - 1st and 15th of the Month | MS, Indexer | ||
DeviceAuditJob | Fetches audit records from Hitachi HNAS EVS configured in Information Governance | Every 1 second | Collector | ||
DiscoverDFSJob | Executes mxdfsscan.exe. It accepts csv containing DFS namespace as an input. It discovers this namespaces and creates csv as an output | 2:00:00 AM & 2:00:00 PM - Daily | MS | Requires dfs_<timestamp>.csv file Generates dfs_<timestamp>_out.csv |
|
DiscoverDFSJob_import | Validates and imports DFS paths generated by DiscoverDFSJob | 3:00:00 AM & 3:00:00 PM - Daily | MS | Consumes dfs_<timestamp>_out.csv | |
DiscoverSharesJob | Discovers shares on filers and web applications on site collections | 2:00:00 AM & 2:00:00 PM - Daily | All | Creates <NODE ID>_<timestamp>_discovered_shares.zip for the node | The zip file is sent to the MS via the FileTransferJob |
DiscoverSharesJob_mrg | Merges all of the zip files produced by DiscoverSharesJob | 3:00:00 AM & 3:00:00 PM - Daily | MS | Consumes _discovered_shares.zip files for all nodes | Runs an hour later than the regular DiscoverSharesJob so as to let the share discovery complete |
DlpSensitiveFilesJob | Pulls classification information from the configured DLP server and generates tags.db file for each MSU | 12:00:00 AM - Daily | MS | Generates tags file in outbox for each msu with the format msu<MSU ID>_<timestamp>_dlp-tags.sqlite | Generated tags files will be transferred to respective indexers and will be consumed by next IndexWriterJob |
EVGetCategoriesJob | Pulls the retention categories of all the configured EV servers | Every 1 hour | MS | Updates <DATADIR>\conf\workflow\steps\ev\EV.db | |
EmailAlertsJob | Send email alerts which are created when any Information Governance policy is violated | Every 15 minutes | MS | Reads <DATADIR>\alerts\alerts.db | SMTP settings need to be configured successfully |
EmailEventsJob | Sends email notifications as configured in Information Governance. The email notifications pertain to events happening in the product | Every 15 minutes | MS | Reads <DATADIR>\events\events_email.db | SMTP settings need to be configured successfully |
FileTransferJob | Transfers the files from the <DATADIR>\outbox folder from a Collector node to the <DATADIR>\inbox folder of the appropriate Indexer or Management Server node | Every 1 minute | All | ||
FileTransferJob_Evt | Transfers event files (internal events of the product) to MS | Every 10 seconds | All | Proccesses <timestamp>_<NODE ID>_<version>_events.sqlite files | Ensures that events make it to the management server asap without waiting in queue for other files to get transferred |
FileTransferJob_WF | Transfers workflow DB from MS to appropriate Portal node | Every 1 minute | MS, Portal | Processes <DATADIR>\workflow\outbox\workflow.db | |
FileTransferJob_alerts | Transfers alerts files between Information Governance nodes | Every 10 seconds | All | ||
FileTranferJob_classify | Distributes the classification events between nodes | Every 10 seconds | All | ||
FileTransferJob_content | Routes content files and classification sqlite files to the assigned classification server | Every 10 seconds | WinNAS w/agent | ||
FileTransferJob_mipscanout | Routes sqlite databases containing the MIP labels from the scan to the associated indexer node | Every 1 minute | Collector | ||
HNasEnableAuditJob | Enables SACLs for the shares when a Hitachi HNAS filer is configured | Every 10 minutes | Collector | ||
IScannerJob | Initiates the incremental scan process for shares/site collections for paths that have changed on those devices since the last full scan | 7:00:00 PM - Daily | Collector, WinNAS w/agent | Consumes msu<MSU ID>_<timestamp>_path.db files Generates scan_cifs_<MSU ID>_<timestamp>.sqlite and scan_sharepoint_<MSU ID>_<timestamp>.sqlite files |
|
IScannerJob_sponline | Initiates the incremental scan process for paths that have changed on configured SharePoint online devices since the last scan | 7:00:00 PM - Daily | Collector | Consumes msu<MSU ID>_<timestamp>_path.db files Generates scan_sponline_<MSU ID>_<timestamp>.sqlite files |
|
IndexCheckJob | Verifies the integrity of the index databases | 6:00:00 PM - Monday | Indexer | ||
IndexVoldbJob | Index all volume usage information received from all collectors and Windows File Server agents | Every 10 minutes | MS | Consumes voldb files <DEVICE ID>_<timestamp>_voldb.sqlite | |
IndexWriterJob | Initiates the idxwriter process to update the Indexer database with scan (incremental and full) and audit data | Every 4 hours | Indexer | Consumes audit_sqlite, scan_sqlite and _tags.sqlite files | |
IndexWriterJob_autoResume | Executes IndexWriterJob if the disk safeguard threshold percentage is below CRITICAL Level which will only process paused/stopped MSU | Every 5 minutes | Indexer | ||
IndexWriterJob_cleanup | Performs clean-up activity on orphan index DBs and temp DBs associated with the MSUs in a paused/stopped state. For stopped/paused MSUs, this job will execute idxwriter.exe to checkpoint the database |
Every 15 minutes | Indexer | ||
IndexWriterJob_erroredFilesCleanup | Performs clean-up activity on the error files generated during IndexWriterJob | 12:00:00 AM - Daily | Indexer | ||
KeyRotationJob | Changes encryption keys used by the nodes | On Demand | MS | ||
LabelsConsumerJob | Processes AIP Labels fetched from the configured OneDrive and SPOnline devices | Every 10 minutes | Consumes msu<MSU ID>_<timestamp>_aip_tags.sqlite files | ||
LocalUserScanJob | Scans the local users and groups on the configured devices | 12:30:00 AM - Daily | Collector, WinNAS w/agent | Generates <NODE ID>_<timestamp>_ local_users.zip files | The local_users.zip files are transferred to MS via FileTransferJob to be consumed by ADScanJob |
MergeStatsJob | Aggregates the published statistics into a primary <DATADIR>\stats\statistics.db that are rolled up into hourly, daily, weekly buckets | Every 15 minutes | Collector, WinNAS w/agent | For WinNAS nodes, this job creates a statistics file to be sent over to its collector node. The collector node consolidates stats from all its WinNas nodes |
|
MsuAuditJob | This job computes the storage statistics for each msu folder on the indexer. Statistics which are consumed per msu are activity index files size, hash files size, segment files size, index db size and total size | 6:00:00 PM - Saturday | Indexer | Stores statistics in <DATADIR>\stats\msu_stats.db file | This information is used to show msu level statistics in GUI. Also useful for monitoring historical growth |
MsuMigrationJob | This job starts/resumes any filer migration initiated by user | 7:00:00 AM - Daily & immediately after migration is executed | Indexer | Copies all MSU folders from source Indexer to destination Indexer | |
NFSUserMappingJob | Maps every UID in raw audit files for NFS and VxFS to an ID generated for use in Information Governance (either a SID of a local user or NIS/LDAP user) | Every 1 hour | Collector | Processes fpolicynfs_<timestamp>_<gen>_<DEVICE ID>.sqlite and vxfs_<gen>_<timestamp>_<DEVICE ID>.sqlite | Enabled only on collector nodes that have VxFS or NFS filers configured |
PauseResumeClassifyRequestJob | Create and refreshes an In-memory list of Paused CRID's | Every 20 seconds | MS | Updates <DATADIR>\classification\conf\paused_resume_classification.db | |
PingHeartBeatJob | Sends the heartbeat from the worker node to the Management Server | Every 1 minute | All, except MS | ||
PingMonitorJob | Monitors heartbeats from the worker nodes. Sends notifications in case it does not get a heartbeat from the worker node | Every 1 minute | MS | ||
ProcessEventsJob | Processes all internal Information Governance events received from worker nodes and adds them to the events.db files on the Management Server | Every 1 minute | MS | Processes _events.sqlite files in inbox Adds events to the MS date wise file stored in <DATADIR>\events\yyyy-mm-dd_events.db |
|
ProcessEventsJob_SE | Consumes the scanerror files and inserts the information about failed paths into a per-share file on MS | Every 1 minute | MS | Input: <DATADIR>\inbox\scanerror_<msuId>_<timestamp>_snap.sqlite Output: <DATADIR>/console/stat/scan_errors/msu<id>.db |
|
RTWBJob | Consumes copies of the audit files generated from the CollectorJob to use for Real Time | Every 1 minute | Indexer | Consumes wb_audit_<namespace>_<MSU ID>_<timestamp>.sqlite files generated by the CollectorJob for Real Time Policies get_rt_alerts.exe creates alerts_<NODE ID>_<timestamp>_wbrt.sqlite files to alerts\outbox folder |
Generated alert files are transferred to the MS via FileTransferJob and processed on the MS by the PolicyJob_alerts job to create the alerts |
RiskDossierJob | Computes the number of files accessible and number of sensitive files accessible to each user on each MSU | 11:00:00 PM - Daily | Indexer | ||
SFAuditJob | Collects audit logs from the Veritas file system servers | Every 10 minutes | Collector | Generates <DATADIR>\unix\vxfs_<gen>_<timestamp>_<DEVICE ID>.sqlite files | Enabled only on collector nodes that have VxFS filers configured |
SFUpdateJob | Sends Information Governance configuration changes to the Veritas file system server | Every 10 minutes | Collector | Updates VxFS server with share configuration changes | |
SPAuditJob | Collects the audit logs from the SQL Server database for SharePoint web applications and generates SharePoint audit databases | Every 2 hours | Collector | Generates audit_sharepoint_<MSU ID>_<timestamp>.sqlite | Enabled only if a SharePoint web application is configured on the collector |
SPEnableAuditJob | Enables auditing for site collections which have been added to Information Governance for monitoring | Every 10 minutes | Collector | ||
SPScannerJob | Scans the site collections at the scheduled time | 11:00:00 PM - Daily | Collector | Generates scan_sharepoint_ files | Enabled only if a SharePoint web application is configured on the collector |
ScanMipLabelJob | Executes mipactionsutil.exe for each msu and fetches the labels and creates output databases | Every 2 minute | Collector | Consumes <DATADIR>\collector\scan_mip\scan_mip_inputs\mip_input_<MSU ID>_<timestamp>.sqlite Generates <DATADIR>\collector\scan_mip\scan_mip_outputs\scan_mip_labels_<MSU ID>_<timestamp>.sqlite |
|
ScanPauseResumeJob | Checks the changes to the pause and resume settings on the Information Governance servers, and accordingly, pauses or resumes scans | Every 1 minute | Collector, WinNAS w/agent | ||
ScannerJob | Initiates the scanner process to scan the shares/site collections added to Information Governance. Creates the scan database for each share that it scanned in the outbox folder | 7:00:00 PM - Last Friday of the Month | Collector, WinNAS w/agent | Output: scan_cifs_<MSU ID>_<timestamp>_snapshot.sqlite or scan_sharepoint_<MSU ID>_<timestamp>.sqlite | There may be additional ScannerJob_<num> instances if custom scan schedules are configured |
SendNodeInfoJob | Sends the node information, such as the operating system, Information Governance version running on the worker node to the MS | 11:00:00 PM - Saturday | All | Information is sent only if it differs from information in the current DB | |
SpoolEventsJob | Spools events on worker nodes to be sent to Management Server | Every 30 seconds | All | Generates _events.sqlite files | |
StatsJob_Index_Size | Publishes statistics related to index size | 6:00:00 AM - Daily | Indexer | Updates <DATADIR>\stats\lstats.db file | The information is used to display the filer statistics on the Information Governance Management Console |
StatsJob_Latency | Records filer latency statistics for NetApp filers | Every 15 minutes | Collector | ||
SyncScansJob | Collector: Gathers all scan progress information from WinNAS nodes MS: Gathers all scan progress information from Collectors |
Every 1 minute | MS, Collector | The scan status is displayed on the Settings>Scanning Dashboard > In-progress Scans tab of the Management Console | |
SystemMonitorJob | Monitors status of Watchdog service | Every 1 minute | All | ||
TagsConsumerJob | Consumes the CSV file containing tags for paths. Imports the attributes into Information Governance and generates a tags.db file for each file system object | 11:00:00 PM - Daily | MS | ||
UpdateConfigJob | Updates the configuration database on the worker nodes based on the configuration changes made on the MS | Every 1 minute | All | Updates <DATADIR>\conf\config.db.<v> to the latest version | |
UpdateConfigJob_lic | Updates portal license in the configuration database properly | 1:00:00 AM - Daily | MS | ||
UpdateCustodiansJob | Downloads custodian updates from each indexer | Every 1 minute | MS | ||
UpdateLabelCacheJob | This job creates a list of unique labels(AIP) for every OneDrive & SPOnline share and updates it in the MS cache | 12:00:00 AM - Daily | MS | ||
UpdateMIPLabelInfoJob | Generates the output sqlite database files containing the file MIP Label information to be consumed by the indexer | Every 1 minute | MS | ||
UpdateWFCentralAuditDBJob | Updates workflow audit information | Every 1 minute | MS | Updates <DATADIR>\workflow\workflow_audit.db | |
UserIndexJob_idx | Generates user indexes | On Demand | Indexer | ||
UserRiskJob | Updates hashes used to compute the user risk score | 2:00:00 AM - Daily | Indexer | ||
VolInfoJob | Collects volume utilization information from WinNAS filers | Every 24 hours | WinNAS w/agent | Generates <DEVICE ID>_<timestamp>_voldb.sqlite files | The voldb files are sent over to MS and consumed by IndexVoldbJob |
WFStatusMergeJob | Merges the workflow and action status updates for remediation workflows (DLP Incident Remediation, Entitlement Reviews, Ownership Confirmation), Enterprise Vault archiving, and custom actions. Updates the primary workflow database with the details so that users can monitor the progress of workflows and actions from the Management Console | Every 5 minutes | MS | ||
WinnasConfigHashJob | Updates the config hash for Windows File Servers associated with a collector | Every 1 minute | WinNAS w/agent | ||
WinnasConfigHashJob_F | Forces all Windows Filers to update the configuration database | 2:00:00 AM - Sunday | WinNAS w/agent | ||
WorkflowActionExecutionJob | Consumes the request file from the MS when a Records Classification workflow is submitted from the Portal. The request file contains the paths on which an Enterprise Vault action is submitted. When the action on the paths is complete, the job updates the request file with the status of the action | Every 1 hour | MS, Portal |
DataInsightWeb Service Jobs
Job | Description | Default Schedule | Node | Files Processed\Updated | Comments |
CustodianSummaryReportJob | Periodically runs the custodian summary report, which is used to determine the custodians assigned in Data Insight for various resources. The output produced by this report is used in DLP Incident Remediation, Entitlement Review, and Ownership Confirmation workflows | Daily | MS | ||
HealthAuditReportJob | Generates Health Audit Report | 5:00:00 AM - Daily | MS | Generates latest output at <INSTALLDIR>\log\health_audit | |
PolicyJob | Evaluates the policies configured and raises alerts | 12:00:00 AM - Daily | MS | ||
PolicyJob_alerts | Merges all of the Real Time alerts files generated on the Indexers by RTWBJob and updates the alerts DB | Every 10 seconds | MS | Consumes alerts_<NODE ID>_<timestamp>_wbrt.sqlite files Generates alerts_wb_<NODE ID>_<timestamp>_rt.sqlite files which are used to update the alerts_master.db |
The EmailAlertsJob then reads the data from the alerts_master.db and generates the email alerts |
PurgeReportsJob | Deletes report outputs older than threshold configured | 12:00:00 AM - Daily | MS | ||
ReportJob | Schedules report runs as per user configuration | 12:00:00 AM - Daily | MS | ||
UpdateConfigJob | Updates the configuration DB on the worker nodes based on changes made on the MS | Every 1 minute | MS | ||
UserIndexJob_merge | Consolidates user activity and permission map from all indexers | Every 2 hours | MS | ||
UserIndexJob_split | Requests each Indexer for user activity and permission map via UserIndexJob_idx | 12:00:00 AM - Daily | MS | ||
UserRiskMergeJob | Combines data from all MSUs into a single risk score value for each user and generates the User Risk report | 6:00:00 AM - Daily | MS | Generates the <DATADIR>\conf\userrisk_dashboard.db |
DataInsightWorkflow Service Jobs
Job | Description | Default Schedule | Node | Files Processed\Updated | Comments |
WFStepExecuterJob | Processes actions for Enterprise Vault archiving, requests for permission remediation, and custom actions | Every 5 minutes | MS, Portal | ||
WFStepExecuterJob_im | Processes Entitlement Review, DLP Incident Remediation, and Ownership Confirmation workflows. Sends email reminders containing links to the remediation portal to the custodians at a specified interval | Every 1 hour | MS, Portal | ||
UpdateConfigJob | Updates the workflow schedules based on configuration changes from the MS | Every 1 minute | All | ||
WFSpoolStatusJob | Monitors workflow data and generates status DB with any new updates | Every 5 minutes | Portal | ||
FileTransferJob_WF | Transfers workflow status DB's from Portal to MS | Every 1 minute | Portal |
DataInsightWatchdog Service Jobs
Job | Description | Default Schedule | Node | Files Processed\Updated | Comments |
RegenerateCommdCertsJob | Regenerates keys, copies it to all the nodes at the scheduled time. It replaces old keys with new ones and restarts all the services. | 2:00:00 AM - Daily | All | ||
SyncPerformanceStatsJob | Fetches performance related statistics from all nodes | Every 30 minutes | MS | ||
SystemMonitorJob | Collects resources (disk usage, CPU, memory usage) from the node | Every 1 minute | All | ||
SystemMonitorJob_backlog | Gathers statistics for unprocessed backlog files | Every 1 hour | All | ||
SystemMonitorJob_lic | Captures and prints total users and total storage information in WatchDog Logs | Every 12 hours | MS | ||
SystemMonitorJob_cleanup | Performs clean-up activity on classification database files | 6:00:00 AM - Daily | All | ||
SystemMonitorJob_monitor_config | Monitors the Information Governance configuration database updates | Every 6 hours | MS | ||
SystemMonitorJob_monitor_coredump | Monitors <installdir>\dumps folder for core dumps from process crashes | Every 1 hour | All | ||
SystemMonitorJob_monitor_service_status | Monitors the Information Governance service status on the nodes | Every 1 minute | All | ||
UpdateConfigJob | Reconfigures watchdog jobs based on configuration updates from MS | Every 1 minute | All |
DataInsight Classification Service Jobs
Job | Description | Default Schedule | Node | Files Processed\Updated | Comments |
CancelClassifyRequestJob | Fetches the list of classification requests that are canceled and distributes this request between Information Governance nodes | Every 20 seconds | Classification | Before classifying files, all the classification jobs consult this list to identify the requests that are marked for cancelation. If they observe any canceled request in the new request that is submitted for classification, then that request is deleted | |
ClassifyFetchJob | Processes classification input files and adds to the priority queues and tracks a list of files that could not be fetched | Every 1 minute | Classification | Input: <DATADIR>\classification\inbox\<priority>_<CR ID>_<BATCH ID>_<NODE ID>_<MSU ID>_<timestamp>_snap<N>.csqlite files | The input file contains the location where the actual file has been kept in the <datadir>\classification\content folder |
ClassifyFetchPauseJob | Refreshes the pause or resume status of fetch jobs as per the duration configured for content fetching | Every 1 minute | Classification | ||
ClassifyJob | Checks the <DATADIRr>\classification\inbox folder for input files submitted for classification folder and adds them to three separate priority queues. It picks a file from the highest queue in FIFO order, and starts classifying content using Arctera Information Classifier. All files in that input file are submitted for classification. Once all paths in the file have been classified, the results of the classification and any resulting errors are written to a database in the <DATADIR>\classification\outbox folder | Every 1 minute | Classification | ||
MIPGetLabelsJob | Fetches MIP Label information from VIC policies | 12:00:00 AM - Daily | MS | Updates mip_labels.db | |
MIPLabelsSyncJob | Creates policies and tags using VIC API's | On Demand | MS | Executed when Settings - Data Management - Set Microsoft Purview Information Protection (MIP) Label configuration is saved | |
UpdateVICPolicyMapJob | It ensures that Information Governance configuration database is in sync with the Classification Policy Manager | Every 10 seconds | MS | ||
UpdateConfigJob | Reconfigures classification jobs based on the configuration changes made on the Management Server | Every 1 minute | Classification | ||
CreateFeaturesJob | Checks if sufficient classified data is available for the supervised learning algorithm to create predictions (training sets) | 12:01:00 AM Sunday | Indexer | The job has a multi-threaded execution framework which executes actions in parallel. The default thread count is 2. You can set the value using the matrix.classification.sl.features.threads property at global or node level. NOTE: The node level property always takes precedence over the global level property |
|
PredictJob | Copies the prediction files from the temp output directory to a classification outbox | 5:00:00 AM Sunday | Indexer | ||
SLCreateBatchesJob | Creates batches of files for the consumption of Arctera Information Classifier. These files are classified with high priority | Every 2 hours | Indexer | ||
ClassifyManageWorkloadJob | Checks the classification or workload folder on the primary Classification Server and counts batches based on their priority. If the workload needs to be distributed, the job fetches a list of servers in its pool and fetches the number of batches based on their priority in the classification or inbox folder. If the number of batches on any secondary that have a priority less than 10, then the job distributes the batches across that secondary and copies them to the secondary's classification or inbox folder | Every 1 minute | Classification (Primary) |
Reference
How Data Insight works, and the ports that it uses for communication with devices