Documente Academic
Documente Profesional
Documente Cultură
Version 7.5.0.1
SC27-5692-00
IBM StoredIQ Platform
Version 7.5.0.1
SC27-5692-00
Note
Before using this information and the product it supports, read the information in Notices on page 161.
This edition applies to Version 7.5.0.1 of (product number 5725-M86) and to all subsequent releases and
modifications until otherwise indicated in new editions.
Copyright IBM Corporation 2001, 2013.
US Government Users Restricted Rights Use, duplication or disclosure restricted by GSA ADP Schedule Contract
with IBM Corp.
Contents
How to send your comments . . . . . v Configuring audit settings . . . . . . . . 30
Configuring hash settings . . . . . . . . 30
Contacting IBM . . . . . . . . . . . vii Configuring desktop settings . . . . . . . 31
Downloading the IBM Desktop Data Collector
installer from the application . . . . . . . 31
Contacting IBM StoredIQ customer Upgrading the IBM Desktop Data Collector agent 31
support. . . . . . . . . . . . . . . ix Using the Encrypted File System recovery agent 32
Notices . . . . . . . . . . . . . . 161
Using Folders. . . . . . . . . . . . 85
Understanding folder types . . . . . . . . . 85
Creating a folder . . . . . . . . . . . 85 Index . . . . . . . . . . . . . . . 165
Deleting a folder . . . . . . . . . . . 85
iv Administration Guide
How to send your comments
Your feedback helps IBM to provide quality information. Please share any
comments that you have about this information or other documentation that IBM
Software Development ships with its products.
To learn about available service options, call one of the following numbers:
v In the United States: 1-888-426-4343
v In Canada: 1-800-465-9600
For more information about how to contact IBM, see the Contact IBM web site at
http://www.ibm.com/contact/us/.
For more information about how to contact IBM, see the Contact IBM web site at
http://www.ibm.com/contact/us/.
The Administrator is the person responsible for managing the IBM StoredIQ
Platform. This individual has strong understanding of data sources, indexes, data
servers, jobs, infosets, and actions. This list provides an overview as to how IBM
StoredIQ Administrator works:
v Viewing data servers and volumesUsing IBM StoredIQ Administrator, the
Administrator can identify what data servers are deployed, their location, what
data is being managed, and the status of each data server in the system. Volume
management is a central component of IBM StoredIQ Platform. IBM StoredIQ
Administrator also allows the Administrator to see what volumes are currently
under management, which data server is responsible for that volume, the state
of the volume after indexing, and the amount and size of information that is
contained by each volume. Administrators can also add volumes to data servers
through this interface.
v Creating system infosetsSystem infosets that use only specific indexed
volumes can be created and managed within IBM StoredIQ Administrator.
Although infosets are a core component of IBM StoredIQ Data Workbench,
system infosets are created as a shortcut for users in IBM StoredIQ
Administrator.
v Managing usersThe user management area allows Administrators to create
users and manage users' access to the various IBM StoredIQ applications.
v Configuring and managing actionsAn action is any process that is taken
upon the data that is represented by the indexes. Actions are run by data servers
on indexed data objects. Any errors or warnings that are generated as a result of
an action are recorded as exceptions in IBM StoredIQ Data Workbench.
Note: Actions can be created within IBM StoredIQ Administrator and then made
available to otherIBM StoredIQ applications such as IBM StoredIQ Data
Workbench.
v Managing target setsProvides an interface that allows the user to set the
wanted targets for specific actions that require a destination volume for their
actions.
v Managing conceptsThis feature allows you to relate business concepts to
indexed data.
Related concepts:
Viewing data servers and adding volumes
Creating system infosets
Managing users
2 Administration Guide
Configuring and managing actions
Managing target sets
Managing concepts
You have a big data problem: SharePoint sites, wikis, email, files, blogs, discussion
threads, and attachments. Your company's information is its most valuableand
potentially most dangerousasset.
IBM StoredIQ Data Workbench is a data visualization and management tool that
helps you to actively manage your company's data. IBM StoredIQ Data Workbench
helps you to determine how much data you have, where it resides, who owns it,
when it was last utilized, and so on. Then, once you have a clear understanding of
your company's data landscape, IBM StoredIQ Data Workbench helps you take
control of and make informed decisions about your data and act on that
knowledge by copying, copying to retention, or performing a discovery export.
That once-risky data is now a legitimate company asset.
Here are just some of the ways you could use IBM StoredIQ Data Workbench.
v Let's say that you need to find all company email sent from or received by
Eileen Sideways (esideways@thecompany.com). You can use IBM StoredIQ Data
Workbench to find all email and then copy that data to a predefined repository.
You could also use IBM StoredIQ Data Workbench to find all of the
esideways@thecompany.com email that occurred between specific dates and then
make that email available for review.
v As an administrator, you'd like to rid your networks and storage of unused data.
You can use IBM StoredIQ Data Workbench to find all files that have not been
modified in more than five years.
v You would like to find all image files created in 2007. Not only can IBM
StoredIQ Data Workbench find all image files created in 2007, it can also tell you
how much space they occupy on your network.
v A user needs to understand how data regarding Windows is being retained.
Using IBM StoredIQ Data Workbench, you can provide that user with a visual
overview of the number of objects retained and a breakdown of files per data
source. Additionally, you can apply overlays to show the user if those files
contain forbidden information such as credit-card numbers or Social Security
numbers.
Related concepts:
Understanding IBM StoredIQ Platform distributed architecture on page 1
IBM StoredIQ Data Workbench is the tool that allows you to visualize this indexed
data, helping you to identify potential red-flag issues, to know how much data you
have on different servers, and to alert people about potentially interesting or useful
data. IBM StoredIQ Data Workbench helps you to ensure that your company's data
is an asset, not a liability.
IBM StoredIQ eDiscovery is an end-user application that helps legal users during
the initial phases of the eDiscovery process. By helping you to capture needed
electronic data, you can use IBM StoredIQ eDiscovery to communicate easily with
other users regarding the status and review process of work being done. IBM
StoredIQ eDiscovery does not drive the eDiscovery process, but instead helps legal
users to control and communicate those processes more effectively.
IBM StoredIQ eDiscovery helps to address the left side of the Electronic Discovery
Reference Model, as shown below.
Here are just some of the ways you could use IBM StoredIQ eDiscovery.
v Let's say that you need to find all electronic information regarding an upcoming
personal-injury lawsuit. You can use IBM StoredIQ eDiscovery to create a matter
for the suit, and then create boxes to contain email and reports regarding the
case.
v As a legal user, you'd like to see the status of all currently active matters. You
can use the Matter Dashboard to visualize the different matters' states.
4 Administration Guide
v As a data expert, you know that you need to respond to users in a timely
fashion. Using IBM StoredIQ eDiscovery, people can email you directly from the
application regarding a matter's questions.
Within IBM StoredIQ eDiscovery, the user is a non-technical end user who needs to
obtain certain pieces of data in order to complete a legal process. The IBM
StoredIQ eDiscovery user has the knowledge of the information or parameters for
what kinds of data they want, but does not always have knowledge of how to
obtain that data.
Before using IBM StoredIQ eDiscovery, ensure that prerequisites are met:
v The IBM StoredIQ Platform must be deployed, configured, and ready for use.
v IBM StoredIQ eDiscovery is dependent upon other IBM StoredIQ applications
such as IBM StoredIQ Data Workbench and IBM StoredIQ Administrator.
Related concepts:
Understanding IBM StoredIQ Platform distributed architecture on page 1
Through the IBM StoredIQ Data Script interface, you can monitor each of the steps
as they are executed and then view any defined outputs, such as reports, infosets,
or exports that are generated as a result. By running IBM StoredIQ Data
Workbench workflows, the user can reapply processes that have been reviewed
and approved to apply to a wide variety of different data problems.
Related concepts:
Understanding IBM StoredIQ Platform distributed architecture on page 1
IBM StoredIQ Policy Manager allows users to run mature policies and processes at
scale across a wider range of data. What makes IBM StoredIQ Policy Manager so
powerful is that it lets users define and execute systemwide policies, focusing on
the execution of the process rather than understanding or reviewing affected data
objects. Additionally, IBM StoredIQ Policy Manager's reports lets you record what
actions were performed, when they were performed, and what data was affected
by the policy's execution. IBM StoredIQ Policy Manager is an extremely powerful
tool in managing your data effectively.
Primary Tabs
IBM StoredIQ Platform users perform most tasks using the Web interface. The
menu bar at the top of the interface contains three primary tabs that are described
in this table.
Name Description
Related concepts:
Web interface icons and buttons on page 9
Administration
This topic provides descriptions of the Administration tab as well as its subtabs:
Dashboard, Data Sources, and Configuration.
Dashboard
The Dashboard subtab provides an overview of the systems current, ongoing, and
previous processes as well as its current status. This table describes
Administrator-level features and descriptions.
Page refresh Choose from 30-, 60-, or 90-second intervals to refresh the page.
Todays job schedule View a list of jobs scheduled for that day with links to the jobs
summary.
Jobs in progress View details of each job step as it is running, including estimated
time to completion, average speed, total system and contained
objects encountered, harvest exceptions, and binary processing
information.
Harvest statistics Review the performance over the last hour for all harvests.
Event log Review the last 500 events or download the entire event log for
the current date or previous dates.
Appliance status Provides a status view of the appliance. Reboot or restart the
appliance through the about appliance link. View cache details for
volumes and discovery exports.
Data Sources
The Data sources subtab is where Administrators define servers and volumes.
These can be places that are indexed or copied to. A variety of server types and
volumes can be configured for use in managing data. Administrators can add
Enterprise Vault sites, Centera pools, Dell DX Storage Clusters, Celerra, NetApp,
and FileNet servers through the Specify servers area. Volumes are configured and
imported in the Specify volumes section.
Configuration
The Administrator configures system and application settings for IBM StoredIQ
Platform through the Configuration subtab.
Settings Description
System
DA Gateway settings Configure the DA Gateway host or IP address.
Mail server settings Configure what mail server to use and how often to
send email.
System time and date Set the system time and date on the appliance.
Application
8 Administration Guide
Settings Description
Harvester settings Set basic parameters and limits, data object extensions
and directories to skip, and reasons to run binary
processing.
Full-text settings Set full-text search limits for length of word and
numbers and edit stop words. See Configuring full-text
index settings on page 28.
Data object types Set the object types that appear in the disk use by data
object type report.
Audit settings Configure how long and how many audits will be kept.
Related concepts:
Creating volumes and data sources on page 33
Navigating within IBM StoredIQ Platform on page 7
Folders
The Folders tab is where users create and manage application objects.
Related concepts:
Navigating within IBM StoredIQ Platform on page 7
Using Folders on page 85
Audit
The audit feature allows you to review all actions taken using the data server.
These icons and buttons are seen within the IBM StoredIQ Platform dashboard.
The table describes the different icons and buttons seen within the Folders tab.
10 Administration Guide
Icon or Button Name Description
Option Description
View appliance details Shows software version and details of
harvester processes running on the controller
for the appliance.
Application services Shows a list of all services and current
status, including:
v Servicethe name of each service on the
appliance component
v PIDthe process ID associated with each
service
v Current memory (MB)the memory
being used by each service
v Total memory (MB)total memory being
used by each service and all child services
v CPU percentagethe percentage of CPU
usage for each service. This value is zero
when a service is idle.
v Statusthe current status of each service.
Status messages include Running,
Stopped, Error, Initializing, Unknown.
14 Administration Guide
Option Description
Storage
Storage information for a controller includes:
v Volume
v Total Space
v Used space
v Percentage
Controller and compute node status
Indicator lights show component status:
v GreenRunning
v YellowThe node is functional but is in
the process of rebuilding; performance
may be degraded during this time. Note:
The rebuild will progress faster if the
system is not in use.
v Rednot running
Related concepts:
Performing IBM StoredIQ Platform administration on page 13
System Section
Application Section
Option Description
Primary Network Interface v IP typeSet to stic or dynamic. If set to
dynamic, the IP address, Netmask, and
Default Gateway fields are disabled.
v IP addressEnter the IP address if
specifying the address manually.
v NetmaskEnter the network mask of the
IP address.
v Default gatewayEnter the IP address of
the default gateway.
v HostnameEnter the fully qualified
domain name assigned to the appliance.
v Ethernet speedSelect the Ethernet speed.
v Available portsIndicate the available
ports.
v Separate network for file/email
serversSpecify the additional subnet for
accessing file/email servers. Select this
check box if you are using the Web
application from one subnet and
harvesting from another subnet.
18 Administration Guide
A system restart is required for any primary network interface changes to take
effect. See Restarting and Rebooting the Appliance.
Option Description
DNS Settings v DNS search orderEnter the DNS search
order for multiple DNS servers.
v Nameserver 1Set the IP address of the
primary DNS server for name resolution
(required).
v Nameserver 2 Set the IP address of the
secondary DNS server for name resolution
(optional).
v Nameserver 3Set the IP address of the
tertiary DNS server for name resolution
(optional).
DNS settings take effect after they have been saved. Changes to the server's IP
address take effect immediately. Because the server has a new IP address, you
must reflect this new address in the browser address line before continuing.
3. Click OK.
4. Click Server name resolution. Set the following for the data server:
v Windows Share (CIFS) file server name resolutionThese settings take
effect upon saving.
Option Description
Windows Share File Server Name v LMHOSTSEnter the IP hostname
Resolution format.
v WINS ServerEnter the name of the
WINS server.
Option Description
NIS (for NFS) v Use NISSelect this box to enable NIS to
perform UID/GID to friendly name
resolution in an NFS environment.
v NIS DomainSpecify the NIS domain.
v Broadcast for server on local
networkSelect this box if the NIS
domain server is located on the local
network and can be discovered by
broadcasting. This option does not work if
the NIS domain server is on another
subnet.
v Specify NIS serverIf not using
broadcast, specify the IP address of the
NIS domain server here.
Option Description
Active Directory Active Directory serverEnter the name of
the Active Directory server.
5. Click OK.
Related concepts:
Configuring system configuration options on page 18
Option Description
siqConsoleLogLineTrap This is a straight conversion of a console log
line into a trap. It uses these parameters:
messageSource, messageID, severity,
messageText.
siqRaidControllerTrap Sent when the RAID controller status is
anything but normal. Refer to the MIB for
status code information. It uses this
parameter: nodeNum.
siqRaidDiskTrap Sent when any attached raid disk's status is
anything but OK. It uses this parameter:
nodeNum.
siqBbuTrap Battery Backup Unit (BBU) error on the
RAID controller detected. It uses this
parameter: nodeNum.
siqCacheBitTrap Caching indicator for RAID array is off. It
uses this parameter: nodeNum.
siqNetworkTrap Network interface is not UP when it should
be. It uses this parameter: nodeNum.
siqDbConnTrap Delivered when the active Postgres
connection percentage exceeds an acceptable
threshold. It uses this parameter: nodeNum.
siqFreeMemTrap Delivered when available memory falls too
low. It uses this parameter: nodeNum.
siqSwapUseTrap Sent when swap use exceeds an acceptable
threshold. Often indicates memory leakage.
It uses this parameter: nodeNum.
siqCpuTrap Sent when CPU load averages are too high.
It uses this parameter: nodeNum.
siqTzMismatchTrap Sent when a nodes time zone offset does
not match that of the controller. It uses this
parameter: nodeNum.
Related concepts:
Configuring system configuration options on page 18
Note: A system restart is required for any changes made to the system time and
date. See Restarting and Rebooting the Appliance.
The gateway must be manually configured to support this backup. For additional
information, configure the gateway in the util menu. This can done by contacting
your service representative.
22 Administration Guide
Configuring system configuration options on page 18
Related reference:
Contacting IBM StoredIQ customer support on page ix
Managing users
When managing users, you can log in and out of the system and perform various
account administration tasks.
Related concepts:
Configuring system configuration options on page 18
4. To log out of the application, click in the upper right-hand corner, and
then click the Log out link.
Related concepts:
Managing users
Creating users
To create a user:
1. Go to Administration > Configuration > Manage users, and then click
Create new user.
2. In the First name field, enter the user's first name.
3. In the Last name field, enter the user's last name.
4. In the Email address field, enter the user's email address.
5. For Authentication, select Active Directory or Local.
Editing users
To edit a user:
1. Go to Administration > Configuration > Manage users, and then click the
name of the user that you would like to edit.
2. Click Edit User and edit the fields as needed.
3. Click OK to save your changes.
Related concepts:
Managing users on page 23
24 Administration Guide
Managing users on page 23
26 Administration Guide
during a harvest. A 0 in this field means unlimited. This setting applies to all
user-defined expressions (keyword, regular expression, scoped, and
proximity) and all standard attributes.
6. Configure Binary Processing.
a. Run binary processing when text processing failsSelect this option to
run binary processing. The system runs further processes against content
that failed in a harvesting. You can select options for when to perform this
extended processing and how to scan content. Binary processing does not
search image file types such as .GIF and .JPG for text extraction.
b. Failure reasons to begin binary processingSelect the check boxes of the
options that define when to perform extended processing. Binary processing
can enact when extracting text from a file fails, when the format of the file
is unknown to the system parameters, when the data object type is not
supported by the harvester scan, or when the data object format does not
contain actual text.
c. Data object extensionsSet binary processing to process all data files or
only files of entered extensions. To add extensions, enter one per line
without a period.
d. Text encodingSet options for what data to scan and extract when
performing binary processing. This extended processing can accept
extended characters and UTF-16 and UTF-32 encoded characters as text. The
system searches UTF-16 and UTF-32 by default.
e. MinimumsSet the minimum required number of located, consecutive
characters to begin processing for text extraction. For example, if you enter
4, the system begins text processing when four consecutive characters of a
particular select text encoding are found. This setting helps find and extract
helpful data from the binary processing, reducing the amount of false
positives.
7. Click OK. Changes to harvester settings do not take effect until the appliance
is rebooted or the application services are restarted.
Related concepts:
Configuring application configuration options on page 25
OCR processing enables text extraction from graphic image files stored directly on
or inside archives stored on volumes where the Include content tagging and
full-text index option is selected. After content typing inside theIBM StoredIQ
Platform processing pipeline, enabling OCR processing will route the following
files types through an optical character recognition engine OCR to extract
recognizable text.
v Windows or OS/2 bitmap (BMP)
v Tag image bitmap file (TIFF) (TIFF)
v Bitmap (CompuServe) (GIF)
v Portable Network Graphics (PNG)
The text extracted from image files is processed through theIBM StoredIQ Platform
pipeline in the same manner as text extracted from other supported file types.
The OCR processing rate of image files is approximately 710 KB/sec per IBM
StoredIQ Platform harvester process.
Related concepts:
Configuring application configuration options on page 25
Related tasks:
Configuring harvester settings on page 26
28 Administration Guide
Number lengthOnly include numbers that are longer than ____
characters. Enter the number of characters a number must contain in order
to be indexed. The Number length feature enables you to index longer
numbers and ignore shorter numbers. By not indexing shorter numbers,
such as one- and two-character numbers that rarely mean anything, you
can focus your filter on meaningful numbers, such as account numbers,
Social Security numbers, credit card numbers, license plate numbers,
telephone numbers, and more.
ExtensionsIndex numbers based on the file extensions of the data
objects in which they appear. Select Limit numbers for all extensions to
limit numbers in all file extensions to the character limits set above in
Number length. Alternatively, select Limit numbers for these extensions
to limit the numbers selected in Numbers length only to data objects
having certain file extensions. Enter the file extensions one per line that
should have limited number indexing. Any data object with a file
extension that is not listed will have all numbers indexed.
4. Configure Stop words. Stop words are common words found in data objects
that are not indexed in the full-text index. To index stop words like the or but
would compromise full-text indexing speed and the relevancy of the results. By
default, the following words are omitted from the full-text index: a, an, and,
are, as, at, be, but, by, for, if, in, into, is, it, no, not, of, on, or, such, that,
the, their, then, there, these, they, this, to, was, will, with.
v If you want the system to find a word listed, delete the word from the list to
allow the full-text index to find it. You must reindex to find the word.
v To add a stop word, enter one word per line, without punctuation, which
includes hyphens and apostrophes.
5. Configure Punctuation or special characters to index by entering punctuation
characters to be included in the index. Note that punctuation characters should
be entered without spaces separating them. By default, only some punctuation
is indexed as a letter in a word by default. Most punctuation is turned into a
space. To make certain that a specific punctuation character is indexed, add it
to this list.
6. To configure Include word stems in index, select whether or not to stem words
that have been indexed. For example, employ is the stem word of words such as
employed, employment, employs, and so on. If you utilize stemming and search
for the term employed, IBM StoredIQ Platform will denote any found instances
of employment, employ, employee, and so on when viewing the data object.
v Do not stem words that are indexed (faster indexing)By not stemming
indexed words, data sources will be indexed faster.
v Stem words that are indexed (improved searching)By stemming indexed
words, filters will be more precise, although slower. Without stemming, a
filter for trade would need to be written as trade or trades or trading or
traded to get the same effect, and even then a user may miss an interesting
variant.
Note: If stemming is enabled, the use of double quotes will return stemmed
terms in results. To find exact words with no stemmed terms, use single quotes.
7. To Optimize wildcard suffix searches, select whether or not to optimize
searches by suffix such as *ology, *tion, *ious, *ive, or *less.
v Do not optimize searches by suffix (faster indexing)By not optimizing
searches by suffix, data sources will be indexed faster.
Note: Stemming will not be performed for search terms with wildcards ("?",
"*") in them. This is true regardless of the term being placed within single
quotes.
8. Click OK.
Related concepts:
Configuring application configuration options on page 25
30 Administration Guide
email hash selections operate independently from the data object hash settings;
that is, a data object can have a binary hash or an email hash, but not both.
3. When hashing data objects, except for emails, specify the following options:
v Entire data object content (required for data object typing)
v Partial data object content
4. Click OK.
Related concepts:
Configuring application configuration options on page 25
To download the IBM Desktop Data Collector installer from the application:
1. Go to Administration > Configuration > Application > Desktop settings.
2. In the Download the Desktop Agent installer area, click Download the
desktop client installer.
3. Once the download is complete, click Save File.
Once the desktop client has been installed on a desktop and then connected to and
registered with the data server, that desktop is available as a data source within
the list of primary volumes. Connectivity and the correct IP address is required.
Related tasks:
Configuring desktop settings
32 Administration Guide
Creating volumes and data sources
A volume represents a data source or destination that is available on the network to
the IBM StoredIQ Platform appliance. A volume can be a disk partition or group of
partitions that is available to network users as a single designated drive or mount
point. IBM StoredIQ Platform volumes perform the same function as partitions on
a hard drive. When you format the hard drive on your PC into drive partitions A,
B, and C, you are creating three partitions that function like three separate physical
drives. Volumes behave the same way that hard drive disk partitions behave. You
can set up three separate volumes originating from the same server or across many
servers. Note that only Administrators can define, configure, and add or remove
volumes to IBM StoredIQ Platform.
Related reference:
Supported server platforms by volume type
Volume indexing
This topic describes volume indexing and the different depths at which volumes
can be indexed.
When defining volumes, you also determine the depth at which you want the
volume to be indexed. There are three levels of analysis:
v System metadata indexThis level of analysis runs with each data collection
cycle and provides only system metadata for system data objects in its results. It
is useful as a simple inventory of what data objects are present in the volumes
you have defined and for monitoring resource constraints (such as file size) or
prohibited file types (such as .MP3s).
v System metadata plus containersIn a simple system metadata index, container
data objects (.ZIP files, PSTs, emails with attachments, and the like) are not
included. This level of analysis provides container-level metadata in addition to
the system metadata for system data objects.
v Full-text and content taggingThis option provides the full native language
analysis that yields the more sophisticated entity tags. Naturally, completing a
full-text index requires more system resources than a metadata index. Users
must carefully design their volume structure and harvests so that the maximum
benefit of IBM StoredIQ Platform sophisticated analytics are used where
necessary, but not on resources that do not require them. Parameters and
limitations on full-text indexing are set when the system is configured.
Related concepts:
Creating volumes and data sources
To harvest and run policies on volumes on Windows Share (CIFS) servers, the user
must be in the backup operator group on the Windows Share server exposing the
shares on IBM StoredIQ Platform.
Related concepts:
Configuring server platforms on page 33
NFS
This topic describes the NFS server platform.
To harvest and run policies on NFS servers, you must enable root access on the
NFS server that is connected to IBM StoredIQ Platform.
Related concepts:
Configuring server platforms on page 33
Exchange servers
This topic describes Microsoft Exchange servers and various connections and
permissions.
v Secure connectionIf you want to connect to Exchange volumes over HTTPS,
you must add port number 443 after the server name, for example,
qa03exch2000.qaw2k.local:443. In some cases, this secure connection can result
in some performance degradation due to SSL overhead. If you enter the volume
information without the 443 suffix, the default connection will be over HTTP.
v Permissions for Exchange 2003The following permissions must be set on the
Exchange server to the mailbox store or the mailboxes from which you will
harvest.
Read
Execute
Read permissions
List contents
Read properties
List object
Receive as
v Permissions for Exchange 2007 and 2010The Full Access permissions must be
set on the Exchange server to the mailbox store or the mailboxes from which
you will harvest.
v Recoverable items in Exchange 2010To harvest the recoverable items folders
in Exchange 2010, you must be logged in as either an Administrator role.
v Deleted itemsTo harvest items that have been deleted from the Exchange
server, enable Exchanges transport dumpster settings. For more information,
refer to Microsoft Exchange Server 2010 Administrator's Pocket Consultant.
Configuration information is also available online at www.microsoft.com.
v Windows AuthenticationFor all supported versions, enable Integrated
Windows Authentication on each Exchange server.
Related concepts:
NFS
34 Administration Guide
Enabling integrated Windows authentication on Exchange
servers
This topic provides procedural information regarding how to integrate Windows
authentication on Exchange servers.
SharePoint
This topic describes SharePoint servers and various connections and privileges.
Secure Connection
If you wish to connect to SharePoint volumes over HTTPS, you may use the Use
SSL checkbox, or may add the port number 443 after the server name when setting
up the volume on StoredIQ. For a example, qa01.company.com:443. In some cases,
this secure connection can result in some performance degradation due to Secure
Socket Layer (SSL) overhead. If you enter the volume information without the 443
suffix, the default connection is over HTTP.
To run policies on SharePoint servers, you must use credentials with Full Control
privileges. We recommend using a Site collection administrator to harvest subsites
of a site collection.
Related concepts:
Configuring server platforms on page 33
IBM StoredIQ Platform is typically used with SharePoint for one of these instances:
to harvest and treat SharePoint as a source for policy actions or to use as a
destination for policy actions, meaning that you can write content into SharePoint
using IBM StoredIQ Platform. Consider these points:
v Attributes are not set/reset on a SharePoint harvest or if you copy from
SharePoint.
v Attributes are set only if you copy to SharePoint.
If you plan to only read from the SharePoint (harvest and source copies from),
then you must use user credentials with Read privileges on the site and on all of
the lists and data objects that you expect to process.
If you plan to use SharePoint as a destination for policies, you must use user
credentials with Contribute privileges on the site.
If you wish to index all the social data for a user profile in SharePoint 2010, then
the user credentials must own privileges to Manage Social Data as well.
Recommended Privileges
Alternate-access mappings
SharePoint 2007 and 2010 require the configuration of alternate-access mappings to
map IBM StoredIQ Platform requests to the correct Web sites.
36 Administration Guide
Example
You are accessing a SharePoint volume with the fully qualified domain name,
http://itweb.storediqexample.com, from the Intranet zone. An alternate-access
mapping for the public URL http://itweb.storediqexample.com for the Intranet
zone must be configured for the SharePoint 2007 or 2010 Web application hosting
the site to be accessed by the volume definition. If you are accessing the same
volume using SSL, the mapping added should be for the URL
https://itweb.storediqexample.com instead.
Note: When configuring SharePoint volumes using non-qualified names, you are
entering the URL for a SharePoint site collection/site that is leveraged by IBM
StoredIQ Platform in the volume definition. Consider the following:
v The URL must be valid with respect to the configured Alternate Access
Mappings configured in SharePoint.
v If the host name in the URL does not convey the fully qualified domain that
should be used to authenticate the configured user, an Active Directory server
must be specified. The specified Active Directory must be a fully qualified
domain name and will be used for authentication.
Related concepts:
SharePoint on page 35
Documentum
This topic describes requirements for Documentum servers.
To run harvests and copy from Documentum servers, you must use the
Contributor role.
Related concepts:
Configuring server platforms on page 33
Enterprise Vault
In order to configure an Enterprise Vault volume, you must first configure an
Enterprise Vault site. This topic provides procedural information regarding how to
configure an Enterprise Vault site.
Discovery Accelerator
This topic describes configurations that must be made to the Discovery Accelerator
server.
It is recommended that the credentials used for referencing the Enterprise Vault
Site are those of the Vault User or any other administrator. IBM StoredIQ Platform
will validate that the credentials are strong enough for it to:
v Login remotely to the specified server
v Perform DCOM operations over RPC remotely
38 Administration Guide
v Permission to set all of the review marks that will be selected for the volume
definition.
Related concepts:
Discovery Accelerator on page 38
NewsGator
This topic provides procedural information regarding required NewsGator
privileges and how they should be configured.
The user account used to harvest or copy from a NewsGator volume must have
the Legal Audit permission on the NewsGator Social Platform Services running on
the SharePoint farm.
40 Administration Guide
(E)xist
(W)rite
Retention
(H)old
v Centera PoolsIf you have an integrated Centera server, you need to create a
Centera pool before you can add a Centera volume. Unlike other volumes,
Centera servers are not placed into volume sets but into Centera Pools. Once
you have created a pool, the Centera pool will be in the list of available choices
when adding a Centera volume.
Prior to creating retention volumes using Enterprise Vault, you must configure
Discovery Accelerator customers and Enterprise Vault sites (customers first, then
sites) so that certain configuration items can appear in the retention volume
configuration lists. Before performing these tasks, you must log in to Discovery
Accelerator and run the ImportExport.exe tool to obtain the appropriate Customer
IDs and customer database names.
Note: If you use a system other than IBM StoredIQ Platform to ingest data into
Enterprise Vault but still want to use IBM StoredIQ Platform for exporting out of
Discovery Accelerator, you must define an Enterprise Vault site within IBM
StoredIQ Platform and then use that site to define a Discovery Accelerator volume.
When defining the Enterprise Vault site, DCOM configuration is a prerequisite.
Note: Remote DCOM is required on all Enterprise Vault servers and Discovery
Accelerator.
You can configure DCOM settings using the DCOM Config utility (DCOMCnfg.exe)
found in Administrative Tools in Control Panel as Component Services. This utility
exposes the settings that enable certain users to connect to the computer remotely
through DCOM. Members of the Administrators group are allowed to connect
remotely to the computer by default.
If the Enterprise Vault Service Account (or the user whose credentials are used to
define the Enterprise Vault Site IBM StoredIQ Platform) does not have permissions
to connect remotely for DCOM, then perform this procedure on the target server.
42 Administration Guide
3. Click the COM Security tab.
4. Under Launch and Activation Permissions, click Edit Limits.
5. In the Launch Permission dialog box, follow these steps if the user name does
not appear in the Groups or user names list:
v In the Launch Permission dialog box, click Add.
v In the Select Users, Computers, or Groups dialog box, add the username
and then click OK.
v In the Launch Permission dialog box, select your user and in the Allow
column under Permissions for User, select Remote Launch and select
Remote Activation, and then click OK.
To allow DCOM traffic over the network on the target server, the DCOM TCP port
(135) must be open on the Firewall. This command opens this port if it is closed:
netsh firewall add portopening protocol=tcp port=135
name=DCOM_TCP135
The port can also be opened using the Firewall User Interface.
Use the Tivoli Storage administrative client, dsmadmc, to enter commands. You
can use the administrative client in either interactive or batch mode. Consult the
Tivoli Storage Manager Administrator's Reference for more details about using the
administrative client.
Note: List NodesYou can also list nodes using the query node command.
Related concepts:
Configuring retention servers on page 40
NetApp SnapLock
This topic provides conceptual information regarding NetApp SnapLock retention
server configuration requirements.
When preparing to add SnapLock retention volumes, you must have licensed
SnapLock, created at least one SnapLock volume and shared it using either
44 Administration Guide
Windows Share or NFS (or both), and initialized the SnapLock compliance clock.
Consult the NetApp administrator documentation for specific instructions.
When configuring the server, set the retention period settings. IBM StoredIQ
Platform recommends setting the minimum setting to zero while not entering a
maximum setting. Access the SnapLock server and enter the following commands:
vol options <vol-name> SnapLock_minimum_period 0d
vol options <vol-name> SnapLock_default_period min
Related concepts:
Configuring retention servers on page 40
The IBM StoredIQ Platform application accesses the Hitachi HCAP server using
HTTP. Consequently, the HTTP gateway must be enabled on the server. Depending
on the current allow/deny lists for the HTTP gateway, you may need to add the
IBM StoredIQ Platform data server's IP addresses to the Allow IP addresses list.
Related concepts:
Configuring retention servers on page 40
Configuring FileNet
By providing the configuration values for a FileNet domain, you are supplying the
values needed to bootstrap into a domain.
There are certain administrative permissions that a user must have when that user
account is used in the Connect as text box in Chatter. When setting up a Chatter
Note: For Chatter administrators using the Auth token option, read how to set up
a sandbox account.
Related concepts:
Creating volumes
Creating volumes
This section describes the types of volumes in IBM StoredIQ Platform as well as
how to create them as data sources within the platform.
Related concepts:
Creating volumes and data sources on page 33
Related reference:
Appendix B, Supported server platforms and protocols, on page 129
Note: Case-sensitivity rules for each server type apply. Red asterisks within the
user interface denote required fields. For best-practice information regarding
editing volume definitions and system restarts, see Editing Volume Definitions.
46 Administration Guide
Applicable Volume
Field Name Required Action Special Notes Type
v Standard
Platform In the Platform list, select the
platform type. v Celerra
v NetApp
Vault store In the Vault store list, select the Enterprise Vault
desired vault store for the volume.
FileNet config Use the FileNet config list to select For more information, see FileNet. FileNet
the FileNet server you would like
to use for this configuration.
Mailbox server When configuring multiple client For Exchange primary volumes, this Exchange
access servers, enter the name of is the fully qualified domain name
one or more mailbox servers, where the mailbox(es) to be
separated by a comma. harvested reside.
v Exchange
Active Directory In the Active Directory server text This must be a fully qualified
server box, enter the name of the Active Active Directory server. v Discovery
Directory server. Accelerator
v SharePoint
v Exchange
Protocol To use SSL, select the Protocol API client uses HTTP over SSL to
check box. communicate with Discovery v Discovery
Accelerator Server configuration. Accelerator
v SharePoint
v NewsGator
Doc base In the Doc base text box, enter the A Documentum repository contains Documentum
name of the Documentum cabinets, and cabinets contain
repository. folders and/or documents.
48 Administration Guide
Applicable Volume
Field Name Required Action Special Notes Type
v Windows Share
Connect as In the Connect as text box, enter For Domino, select the user name
(CIFS)
the logon ID used to connect and for the primary user ID. The user id
mount the defined volume. must be configured on the System v Exchange
Configuration screen under the v SharePoint
Lotus Notes user administration v Documentum
link.
v Domino
v FileNet
v Discovery
Accelerator
v NewsGator
v Chatter
v Windows Share
Password In the Password text box, enter the For Domino, enter the password for
(CIFS)
password used to connect and the primary user ID.
mount the defined volume. v Exchange
v SharePoint
v Documentum
v Discovery
Accelerator
v Domino
v Celerra Windows
Share
v FileNet
v NewsGator
v Chatter
Auth token In the Auth token text box, enter The auth token must match the user Chatter
the token used to authenticate the name used in the Connect as field.
Chatter volume. Auth tokens can be generated
online on Salesforce. See
Configuring Chatter Volumes.
Object store In the Object store list, select the The object store must exist prior to FileNet
desired object store. the creation of a FileNet primary
volume.
Discovery In the Discovery Accelerator case This text box is populated from Discovery
Accelerator case text box, enter the name of the Discovery Accelerator once Accelerator
Discovery Accelerator. connection is established.
v Documentum
Harvest This is the Documentum harvest This option obtains the list of all
option: known Domino users and their v Domino
v To enable harvesting, select NSFs. It then harvests those
Harvest all document versions. mailboxes unless it was pointed to
a single mailbox using the initial
These are the Domino harvest directory.
options:
v To harvest mailboxes, select the
Harvest mailboxes option.
v To harvest mail journals, select
the Harvest mail journals option.
v To harvest all applications, select
the Harvest all applications
option.
Virtual root In Virtual root, change the default For Exchange, this option should be Exchange
name to match the Exchange server changed to match the server
designation. designation.
Personal archives To enable the collection of personal This option pertains only to Exchange 2010
archives, select the Harvest Exchange 2010 with SP1 applied.
personal archive check box.
50 Administration Guide
Applicable Volume
Field Name Required Action Special Notes Type
v Windows Share
Index options Select either or both of the Index These options are selected by
(CIFS)
options check boxes. default.
v NFS v2 and v3
v Include system metadata for
data objects within containers. v NetApp
v Include content tagging and v Exchange
full-text index. v SharePoint
v Documentum
v Celerra
v Enterprise Vault
v Discovery
Accelerator
v Domino
v FileNet
v NewsGator
v Chatter
v Desktop
v SharePoint
Subsites Select Recurse into subsites.
Versions Select Include all versions. IBM StoredIQ Platform supports SharePoint
indexing versions from SharePoint.
Versions For more information, see Special
Note: Adding SharePoint Volumes.
v Windows Share
Validation To validate volume accessibility, When selected (the default state),
(CIFS)
select Validation. IBM StoredIQ Platform tests to see
if the volume can be accessed. v NFS v2 and v3
v NetApp
v Exchange
v SharePoint
v Documentum
v Celerra
v Discovery
Accelerator
v Domino
v Chatter
v Windows Share
Start directory In Start directory, designate a start
(CIFS)
directory for the harvest. The start
directory involves volume v NFS v2 and v3
partitioning in order to break up a v NetApp
large volume. The start directory v SharePoint
must be underneath the initial
v Domino
directory, if an initial directory is
defined. For example, in the case of v Celerra
directories EH, E would be the v FileNet
Start directory and H would be the v Chatter
End directory.
v Windows Share
End directory In End directory, determine the end
(CIFS)
directory for the harvest. The end
directory is also part of volume v NFS v2 and v3
partitioning and is the last directory v NetApp
harvested. v SharePoint
v Domino
v Celerra
v FileNet
v Chatter
v Windows Share
Access Times In Access times, select one of these
(CIFS)
options:
v NFS v2 and v3
v Reset access times but do not
synchronize them. (This is the v NetApp
default setting.) v Domino
v Do not reset or synchronize v Celerra
access times.
v Reset and synchronize access
times on incremental harvests.
52 Administration Guide
Applicable Volume
Field Name Required Action Special Notes Type
v Windows Share
Volume In Constraints, select one of these The Scope harvests on these
(CIFS)
constraints options: volumes by extension option
pertains only to these volume types: v NFS v2 and v3
v Only use __ connection process
or v NetApp
(es)Specify a limit for the v Windows Share
Constraints number of harvest connections to v NFS v2 and v3 v Exchange
this volume. If the server is also v Domino
v NetApp
being accessed for attribute and
v Celerra v Celerra
full-text searches, you may want
to regulate the load on the server v Discovery Accelerator v SharePoint
by limiting the harvester v Documentum
processes. The maximum number v Enterprise Vault
of harvest processes is
automatically shown. This v Discovery
maximum number is set on the Accelerator
system configuration tab. v FileNet
v Control the number of parallel v Chatter
data object readsDesignate the v Desktop
number of parallel data object
reads.
v Scope harvests on these volumes
by extensionInclude or exclude
data objects based on extension.
IBM StoredIQ Platform supports using the entire sites portion of a Sharepoint URL
for the volume /sites/main_site/sub_site in the Volume field when adding a
new SharePoint volume. However, if the SharePoint volume URL contains spaces,
then you must also utilize the Server, Volume, and Initial directory fields in the
Add volume dialog box. In addition to the required fields Server type, Server,
Connect as, and Password. For example, the SharePoint volume with the URL
http://shpt2010.reglab5.local/sitestest/autoteamsite1/Attribute Harvest
WikiPages Library/ would require using the fields in the following table because
of the spaces in the URL.
Primary Volume
Field Example
Server shpt2010.reglab5.local
Volume /sitestest/autoteamsite1
When you add a primary volume, you define the volume by setting certain
properties. If a SharePoint volume is added, you have the option of indexing
different versions of data objects on that volume.
Since most versions of any given object will share full-text content and attributes,
there is a duplication of effort in processing them as well as the effort required to
maintain an updated context for the version history of an object in the index.
Additionally, if you enable versioning on a SharePoint volume, the API itself
causes extra overhead in fetching data and metadata for older versions.
v For each object, an additional API call must be made to get a list of all its
versions.
54 Administration Guide
v To fetch attributes for the older versions of an object, an API call must be made
for each attribute that needs to be indexed.
Related tasks:
Creating primary volumes on page 46
Whenever volume definitions are edited or modified, you should restart the
system.
Related tasks:
Creating primary volumes on page 46
Restarting and rebooting IBM StoredIQ Platform on page 15
Retention volumes store data objects that have been placed under retention,
meaning that the object will be retained.
This is the process for using retention volumes to store such data:
1. Configure your retention servers. See Configuring Retention Servers.
2. Create management or retention classes, or, if you are using Centera retention
servers, create Centera pools as described in Creating Centera Pools.
3. If you are using Enterprise Vault, ensure you have defined Enterprise Vault
Sites (see Discovery Accelerator). You may also want to define Retention
Categories on the Enterprise Vault server.
4. Create retention volumes, as detailed in Adding a Retention Volume.
Related concepts:
Creating volumes on page 46
Note: Case-sensitivity rules apply. Red asterisks within the user interface
denote required fields.
Applicable
Field Name Required Action Special Notes Volume Type
v Windows
Server type In the Server type list, select the
Share (CIFS)
server type.
v NFS v3
v Centera
v Hitachi
v IBM
Information
Archive
v Dell DX
Storage
v Enterprise
Vault
v NetApp
SnapLock
v NetApp
v FileNet
v Standard
Platform In the Platform list, select the
platform type.
v Centera
Pool In the Pool list, select the StoredIQ
pool profile name to provide access
to a specific Centera pool
56 Administration Guide
Applicable
Field Name Required Action Special Notes Volume Type
v Windows
Server In the Server text box, assign the
Share (CIFS)
server a name.
v NFS v3
v Hitachi
v IBM
Information
Archive
v NetApp
SnapLock
v IBM
Node name In the Node name text box, enter
Information
the name of the node.
Archive
v IBM
Node port In the Node port text box, enter the
Information
nodes port number.
Archive
v IBM
Node password In the Node password text box,
Information
enter the password for the node.
Archive
v FileNet
FileNet config In the FileNet config text box, enter
the name of the FileNet connection.
v Windows
Connect as In the Connect as text box, enter the
Share (CIFS)
login ID.
v IBM
Information
Archive
v FileNet
v Windows
Password In the Password text box, enter the
Share (CIFS)
password for the login ID.
v IBM
Information
Archive
v NetApp
SnapLock
v FileNet
v FileNet
Object store In the Object store list, select the The object store
desired object store. must exist prior
to the creation
of a FileNet
retention
volume.
v Windows
Volume In the Volume text box, enter the
Share (CIFS)
name or names of the volume to be
mounted. v NFS v3
v Centera
v Hitachi
v IBM
Information
Archive
v Dell DX
Storage
v NetApp
SnapLock
v Windows
Index options Select either or both of the Index These options
Share (CIFS)
options check boxes. are selected by
v NFS v3
v Include system metadata for data default.
objects within containers. v Centera
v Include content tagging and v Hitachi
full-text index. v IBM
Information
Archive
v NetApp
SnapLock
v Enterprise
Matter archive In the Matter archive name text box,
Vault
name enter the name of the matter archive.
v Enterprise
Description In the Description text box, enter a
Vault
description.
58 Administration Guide
Applicable
Field Name Required Action Special Notes Volume Type
Note: For v Windows
Constraints In Constraints, select either or both
FileNet retention Share (CIFS)
of these options: volumes, you
v NFS v3
v Only use __ connection can select only
process(es)Specify a limit for the Only use __ v Centera
the number of harvest connections connection v Hitachi
to this volume. If the server is process(es) v IBM
also being accessed for attribute option. Information
and full-text searches, you may Archive
want to regulate the load on the
v NetApp
server by limiting the harvester
SnapLock
processes. The maximum number
of harvest processes is v FileNet
automatically shown. This
maximum number is set on the
system configuration tab.
v Control the number of parallel
data object readsDesignate the
number of parallel data object
reads.
Related concepts:
Creating retention volumes on page 55
Discovery export volumes contain the data produced from a policy, which is kept
so that it can be exported as a load file and uploaded into a legal review tool.
Administrators can also configure discovery export volumes for managing harvest
results from cycles of a discovery export policy.
Note: Case-sensitivity rules for each server type apply. Red asterisks within the
user interface denote required fields.
Applicable
Special Volume
Field Name Required Action Notes Type
v Windows
Type Using the Type list, select the type of server.
Share
(CIFS)
v NFS v2,
v3
v Windows
Server In the Server text box, enter the name of the
Share
server where the volume is available for
(CIFS)
mounting.
v NFS v2,
v3
v Windows
Connect as In the Connect as text box, enter the logon ID
Share
used to connect and mount the defined
(CIFS)
volume.
v Windows
Password In the Password text box, enter the password
Share
used to connect and mount the defined
(CIFS)
volume.
v Windows
Volume In the Volume text box, enter the name of the
Share
volume to be mounted.
(CIFS)
v NFS v2,
v3
v Windows
Constraints To utilize Constraints, select this option:
Share
v Only use __ connection process (CIFS)
(es)Specify a limit for the number of
v NFS v2,
harvest connections to this volume. If the
v3
server is also being accessed for attribute
and full-text searches, you may want to
regulate the load on the server by limiting
the harvester processes. The maximum
number of harvest processes is
automatically shown. This maximum
number is set on the system Configuration
tab.
Related concepts:
Creating retention volumes on page 55
60 Administration Guide
Creating system volumes
This topic provides procedural information regarding how to create a system
volume.
System volumes support volume export and import. When you export a volume,
data is stored on the system volume. When you import a volume, data is imported
from the system volume.
Note: Case-sensitivity rules apply. Red asterisks within the user interface
denote required fields.
Applicable
Field Name Required Action Special Notes Volume Type
v Windows
Server type Using the Type list, select the type
Share (CIFS)
of server.
v NFS v2, v3
v Windows
Server In the Server text box, enter the
Share (CIFS)
name of the server where the
volume is available for mounting. v NFS v2, v3
v Windows
Connect as In the Connect as text box, enter the
Share (CIFS)
logon ID used to connect and
mount the defined volume.
v Windows
Password In the Password text box, enter the
Share (CIFS)
password used to connect and
mount the defined volume.
v Windows
Volume In the Volume text box, enter the
Share (CIFS)
name of the volume to be mounted.
v NFS v2, v3
v Windows
Constraints To utilize Constraints, select this
Share (CIFS)
option:
v NFS v2, v3
v Only use __ connection process
(es)Specify a limit for the
number of harvest connections to
this volume. If the server is also
being accessed for attribute and
full-text searches, you may want
to regulate the load on the server
by limiting the harvester
processes. The maximum number
of harvest processes is
automatically shown. This
maximum number is set on the
system Configuration tab.
Only primary and retention volume data can be exported or imported using the
export/import feature. Discovery export and system volumes cannot be imported
or exported. The target location of an export or the source location of an import is
always the IBM StoredIQ Platform system volume.
Export and import volume processes run as jobs in the background. These jobs are
placed into their prospective queues, and they are executed sequentially. When one
job completes, the next one automatically starts. These jobs can be cancelled at any
time while they are running. Cancelling one import or export job also cancels all
the jobs that come after the one cancelled. Because the export jobs and import jobs
reside in separate queues, cancelling one type of job does not cancel jobs in the
other queue. The jobs are not restartable.
The export process creates two files: a binary file and a metadata file, which
contains the exported data. These files' names contain the following information:
v Data server name and IP address
v Volume and server names
v Time stamp
The exported data consists of data from the selected volume and any related
information that describes that data with the exception of volume-specific audits.
The exported data has to be made available to the import data server before it can
be imported. This may require you to physically move the exported data to the
system volume of the import data server.
62 Administration Guide
Note: The job cannot be restarted.
Name Description
Export path (on system Where to save the data on the System volume. The default
volume) path is /exports. You can edit the export path. The specified
location will be automatically created if necessary.
Export full-text index Select this option to export the volumes full-text index.
(Available only if the volume has full-text index.)
An imported volume looks, acts, and is, just like a volume originally defined and
harvested on the data server. Any action or relationship that is valid for a
non-imported volume is valid for an imported volume, with a few exceptions:
v Logs and audit trails that capture the activity on the volume before the import
are not available. However, the import itself is audited. See Import Audits.
v The imported volume can be reharvested as long as the appliance has the proper
network access and rights to the original source server and volume.
v The imported volume can be reharvested as long as the data server has the
proper network access and rights to the original source server and volume.
v The data viewer works only if the appliance has the proper network access and
rights to the source server and volume. You must have access and permission on
export servers and volumes if the file you want to view has been migrated to a
secondary server at the time of the export.
Note: When a volume with a licensed feature is imported into a data server that
does not utilize licensing, the license is imported along with the volume. To see the
licensed features, users will need to log out and then log back in to the data server.
Name Description
Server and volume Server name and volume name where the data physically
resides
Exported from Server name and IP address of the server from which the
data was exported
Export date The day and time the data was exported
Total data objects Total number of data objects exported for the exported
volume
Contains full-text index Whether or not the full-text index option was chosen when
the data was exported
Related concepts:
Creating volumes on page 46
Deleting volumes
Administrators can delete volumes from the list of available data sources, provided
that the data server is connected to the gateway.
64 Administration Guide
v Object counts and sizes within user infosets will remain the same. Remember,
those user infosets were created at a specific point in time when this data source
was still available.
v Users exploring a specific data source and any generated reports will no longer
reference the deleted volume.
v No exceptions will be raised on previously executed actions. Instead, the data is
no longer available. For example, if an infoset is copied that contained data
objects from a volume that has been deleted, no exception is raised.
v If you mark a desktop volume for deletion, it will automatically be removed
from the Primary volume list; however, the status of that workstation is set to
uninstall in the background. When the desktop client next checks in, it will see
that change in status and uninstall itself.
Note: If retention volumes such as Centera, File Net, Hitachi, and so on contain
data, they cannot be deleted as IBM StoredIQ Platform is the source of record.
Instead, you will see the Under Management link.
To delete a volume:
1. Go to Administration > Data sources > Specify volumes > Volumes.
2. Click the tab of the volume type you would like to delete: Primary, Retention,
System, or Discovery export.
3. Click Delete, and in the confirmation dialog, click OK. The volume is deleted,
removing it from the list of available volumes.
StoredIQ imposes some policy limitations on volume types, which are identified in
this table.
Table 1. Policy limitations for volume types
Retention Volume
v Centera
v NetApp SnapLock (CIFS/NFS)
v CIFS Celerra FLR (CIFS/NFS)
v Dell DX Object Store
v CIFS
v NFS
v IBM FileNet
v Documentum /w RPS
v Hitachi HCAP
v IBM Information Archive
v Symantec Enterprise Vault
Copy to Centera, Windows Share SnapLock, NFS
(Retention) SnapLock, Hitachi, IBM Information Archive,
Windows Share, NFS, Documentum, SharePoint,
Enterprise Vault, Dell DX Storage, FileNet
66 Administration Guide
Table 1. Policy limitations for volume types (continued)
Related concepts:
Creating volumes on page 46
Understanding Harvests
This conceptual topic will help you to understand the harvest types available.
These options are selected when you create a job for the harvest. A harvest must be
run before you can start searching for data objects or textual content. An
Administrator initiates a harvest by including a harvest step in a job.
Most harvesting parameters are selected from the Configuration subtab (see
Configuring Application Settings); you can specify the number of processes to use
during a harvest, whether a harvest should continue where it left off if it has been
interrupted, as well as many other parameters. There are also several standard
harvesting-related jobs that are provided in the IBM StoredIQ Platform system.
You can separate harvesting activities into two steps: the initial harvest and harvest
post-processing. The separation of tasks gives Administrators the flexibility to
schedule the harvest or the post-process loading to run at times that do not impact
system performance for system users, who may, for example, be running queries.
These are examples of post-harvest activities:
v Loading all metadata for a volume.
v Computing all tags that are registered to a particular volume.
v Generating all reports for that volume.
v If configured, updating tags, and creating explorers in the harvest job.
Incremental harvests
Harvesting volumes takes time and taxes your organization's resources. You can
maintain the accuracy of the metadata repository quickly and easily with
incremental harvests. With both of these features, you can ensure that the
vocabulary for all volumes is consistent and up to date.
Once you have harvested a volume, you can speed up subsequent harvests by only
harvesting for data objects that have been changed or are new. An incremental
harvest indexes new, modified, and removed data objects on your volumes/file
servers.
Because the harvests are incremental, it takes less time to update the metadata
repository with the additional advantage of putting a lighter load on your systems
than the original harvests.
IBM StoredIQ Platform allows you to perform many types of harvests, depending
on your data needs. While in-depth harvests are common, there are also instances
where you need an overview of the data and a systemwide picture of files types
and sizes. For example, at the beginning of a deployment, you may want to obtain
a high-level view of a substantial amount of data so that you can make better,
more informed decisions about how you want to handle harvesting or other
policies going forward. This section provides possible system configurations that
will allow the system to process the volumes data in the quickest manner possible.
Volume Details
When configuring data sources for a lightweight harvest, you do not need to
include content tagging and full-text indices. By clearing this option, the system
70 Administration Guide
will index the files metadata, not the entire content of those files. The system can
then execute and complete harvests very quickly, and you will obtain a large
amount of information regarding file types, the number of files, the age of the files,
file ownership, and so on.
Harvesting data 71
v At a later date, you have the option of viewing the content of the skipped
files. This will create additional work, reharvesting these skipped files.
2. Determine which Locations to ignore.
There may be instances where large quantities of data are contained in
subdirectories, and that data is not relevant to your harvest strategy. For
example, you could have a directory with a tree of source code or software
archive that is not used as a companywide resource. In these cases, you can
eliminate these directories from the harvests by adding the directory to the
Locations to ignore. These locations are not specific to a volume, but can
instead be used for common directories across volumes.
3. Determine Limits.
v Maximum data object sizeThis setting is only relevant for full-text
harvests. In cases where there are a lot of very large files, you may want to
eliminate processing those files by setting the Maximum data object size to a
smaller number (the default value is 1,000,000,000). You will still collect the
metadata on the very large files, so you can search for them and determine
which files were missed due to the setting of this parameter.
4. Determine Binary Processing.
Binary processing is additional processing that can be performed if the
standard processing cannot index the contents of a file. For lightweight
harvests, the Run binary processing when text processing fails check box
should be cleared as this setting is only relevant for full-text harvests.
Note: Prior to modifying the full-text settings, see Configuring Full-text Index
Settings.
The full-text settings are only valid if you have full-text processing enabled for a
given volume. Consider these options for controlling the impact of a full-text
harvest on the systems performance.
1. Determine Limits. Limit the length of words to be harvested by selecting the
Limit the length of words index to __ characters option. The default value is
50, but you can reduce this number in order to reduce the quantity of indexed
words.
2. Determine Numbers. If there are large quantities of spreadsheet files, you can
control what numbers are indexed by the system.
Note: Prior to modifying the hash settings, see Configuring Hash Settings.
Determine the hash settings. A files hash is a unique, calculated number based on
the content of the file. By selecting Partial data object content, you reduce the
required processing to create the hash. Be forewarned, however, that two different
data objects could create the same hash. This is a small but potential risk. This is
only relevant for full-text harvests.
72 Administration Guide
Configuring jobs
This section provides procedural information about configuring jobs within IBM
StoredIQ Platform.
Jobs start tasks such as harvests. They can be run at the time of creation, or
scheduled to run at a designated future time and at regular intervals.
Jobs consist of either a single step or a series of steps. The actions available at each
step depend on the type of job being created. There are several of out-of-the-box
jobs included in IBM StoredIQ Platform; these and their locations in the interface
are described in this table.
Related concepts:
Configuring jobs on page 73
Creating a job
This topic provides procedural information regarding how to create a job.
To create a job:
1. From the Folders tab > Workspace folder, select New > Job.
2. Enter a unique job name.
3. In the Save in: list, select the appropriate folder, and the job is created.
4. Click OK. If you would like to view the job and add steps, click Yes.
5. On the View job page, click Add step and select a step type from the list.
6. For Run harvest jobs, on the Specify harvest and load options page, configure
the following options:
v Harvest these volumesSelect a volume from the list.
v Harvest typeSpecify the type of harvest:
Run a full harvest, meaning that all data objects on this volume will be
indexed.
Run an incremental harvest (default), meaning that only files or data
objects that have changed since the last harvest will indexed.
74 Administration Guide
v Harvest and load schedulingYou can separate harvest and load processes
to limit resource use. Select:
Load indexes when harvest completes.
Load indexes with next nightly system services job (This delays the
index loading to run with the next system-services job after the harvest
has completed. The system-services job is scheduled to run at midnight by
default.)
v Run harvest onlySelect this option if you plan to load harvested data into
indexes at a later time.
v Load indexes onlySelect this option to load previously harvested data into
indexes.
7. Harvest samplingSelect this option if you want to limit the harvest to a
smaller sample. This option skips every second, third, tenth, or other number
data object as entered in the text box.
8. Harvest limitsLimit the harvest by time or total number of data objects.
Enter the number of minutes or number of data objects.
9. Click OK.
Note: Prior to creating a job to discover retention volumes, you must have added
a retention volume.
Editing a job
This topic provides procedural information regarding editing an existing job.
Configuring jobs 75
1. From the Folders tab > Workspace folder, click the job you would like to edit.
The Job details page opens.
2. Click Edit job details, and the Edit job details dialog box appears, allowing
you to specify the time, date, and frequency for the job to run.
v In the Time: field, enter the time the job must start, or click Now to
populate the time field with the current time. You may want to add some
time if you have not specified all of the job steps.
v In the Date: field, enter the date on which to run the job, or click Today to
populate the date field with the current date.
3. Using the options in the Frequency field, specify how often the job must run.
If you select None, the job runs once, at the time and date provided.
4. Click OK.
5. To edit the job steps:
a. Add a step to by job by clicking Add step.
b. Edit an existing step by clicking Edit.
c. Remove an existing step by clicking Remove.
d. Change the order of existing steps by clicking Move up or Move down
icons.
6. Click OK.
Starting a job
This topic provides procedural information regarding starting a job.
To start a job:
1. To start a job, do either of the following:
v From the Folders tab > Workspace folder, click on the name of the job you
would like to start, and in the Job details page, click Start job.
v From the Folders tab > Workspace folder, right-click the job and select Start
job.
2. In the Schedule area, a started job will be displayed as This job is running
now.
3. Click OK to return to the Folders tab. The started job will display Running in
the Status column.
Saving a job
This topic provides procedural information regarding saving a job.
76 Administration Guide
Running a predefined job
This topic provides procedural information regarding running a predefined job.
Deleting a job
This topic provides procedural information regarding deleting a job.
To delete a job:
1. From the Folders tab > Workspace folder, select Filter by... Jobs.
2. Click the job name to open the job details.
3. Click Delete in the lower-left hand corner of the screen. Click OK.
Monitoring processing
You can track the systems processing on your harvest/policy and discovery export
tasks using the View cache details feature. The appliance gathers data in
increments and caches the data as it gathers it. If a collection is interrupted, the
appliance can resume collection at the point that it was interrupted, instead of
starting over from the beginning of the task.
To monitor processing:
From Administration > Dashboard, in the Appliance status pane, click View
cache details. The View cache details page appears. To see harvest/policy progress,
click the Volume cache tab, or to see discovery export job progress, click the
Discovery export cache tab.
Note: Note that information for a job is only available while the job is running.
Once a task is completed, the job disappears from the list.
Table 2. Harvest/Volume Cache Details
v Copy
Type Type of job being executed
v Harvestfull
v Harvestincremental
Configuring jobs 77
Table 2. Harvest/Volume Cache Details (continued)
v CachingThe volume
State Status of the process
cache is currently being
created/updated by a
harvest or policy
v CachedCreation or
update of volume cache is
complete (harvest only)
v LoadingVolume cache
contents are being
successfully loaded into
the volume cluster
CachedCreation or update
of volume cache is complete
(harvest only).
LoadingVolume cache
contents are being
successfully loaded into the
volume cluster
Related concepts:
78 Administration Guide
Configuring jobs on page 73
To troubleshoot a harvest:
1. Click Administration > Dashboard > Jobs in Progress to verify that your job
continues to run.
2. In Jobs in Progress, note the Total data objects encountered number.
3. Wait 15 minutes, letting the harvest continue to run.
4. Note the new value the Total data objects encountered, and then compare it to
that value denoted previously.
5. Go to Question 1: Is the Total data objects encountered counter increasing?
Question 1: Is the Total data objects encountered counter increasing?
v YesIf the number of encountered data objects continues to increase, then
the harvest is running correctly.
v NoIf the number of encountered objects remains the same, then go to
Question 2: Is the load average up?
Question 2: Is the load average up?
a. To view load averages, on Appliance status > About appliance > View
details > System services, look at the load averages in the Basic system
information area.
v YesIf the load averages number is up, the harvest may be stuck. Call
technical support to report that the harvest is stuck on files.
v NoThe job is not really running, meaning that the job must be restarted.
Go to Question 3: Did the job complete on the second pass?
Question 3: Did the job complete on the second pass?
v YesIf the job completed successfully after it was restarted, then the harvest
is not stuck.
v NoThe job did not complete successfully. Call technical support to report a
job that does not complete.
Configuring jobs 79
80 Administration Guide
Utilizing desktop collection
When configuring desktop settings, you are enabling or disabling encryption
within IBM StoredIQ Platform. The IBM Desktop Data Collector (desktop client or
client) enables desktops as a volume type or data source, allowing them to be used
just as other types of added data sources. The IBM Desktop Data Collector is
provided as a standard MSI file and is installed according to the typical method
(such as Microsoft Systems Management Service (SMS)) used within your
organization. The IBM Desktop Data Collector can collect PSTs and .ZIP files as
well as other data objects and is capable of removing itself once its work is
completed.
Once the desktop client has been installed on a desktop and then connects and
registers with the data server, that desktop is available as a data source within the
list of primary volumes. Additionally, while the snippet support and the snippet
step-up action are supported by IBM Desktop Data Collector, it should be noted
that a desktop cannot be the target or destination of an action.
Prior to using the IBM Desktop Data Collector, the Administrator may want to
notify end users that desktop collection is going to be performed and make them
aware of the following:
v That the desktop must be connected over the network during data collection. If
the connection is interrupted, IBM Desktop Data Collector will resume its work
from the point at which it stopped.
v That users might notice a slight change in performance speed, but that they can
continue working normally. Desktop collection will not interfere with work
processes.
v That certain actions can be taken from the tray icon: Right-click for About,
Restart, Status, and Email Logs (which packages logs into single file and
launches the email client so that the user can mail them to the IBM StoredIQ
Platform Administrator).
Note that all communications are outbound from the client. The appliance never
pushes data or requests to the desktop. The IBM Desktop Data Collector pings the
One can download the installer application from the application in the
Configuration tab. Also, the Administrator can temporarily disable the client
service on all desktops registered to the data server from the Configuration tab.
Required
v SERVERACTIONNODEADDRESSIP address or hostname for the Action
node. When the install is not silent, the user is prompted for IP address or
hostname. The default is the value of this argument. This field must be entered
accurately or manual correction will be required in the desktop config file.
Optional
v SERVERACTIONNODEPORTPort number for the Agent on the Action node.
Defaults to 21000, and should only be changed when the agent will connect on a
different port that is subsequently mapped to 21000.
v NOTRAYICONSpecifies whether the agent displays the IBM Desktop Data
Collector tray icon while running. Changing the setting to 1 forces the agent to
run silently and not display a tray icon.
v MSI
v Emailing linksSend a link within an email such as file:\\g:\group\install\
Client-install.vbs. The link may be to any executable file format such as .BAT,
.VBS, or .MSI. The .BAT/.VBS formats can be used to pass client arguments to
an .MSI. The user executing the link must have administrative privileges.
v NT Logon Script, in which a .BAT file or .VBS script invokes msiexec. Examples
are given here:
/iinstall
/x {7E9E08F1-571B-4888-AC08-CEA8A076F5F9}uninstall the agent. The
product code must be present.
/quietinstall/uninstall runs silently. When specifying this option,
SERVERACTIONNODEADDRESS must be supplied as an argument.
Set WshShell = CreateObject("WScript.Shell")
Batch file
msiexec /i G:\group\install\desktopclient.msi NOTRAYICON=1
SERVERACTIONNODEADDRESS=clust017.test.local /q
82 Administration Guide
Related concepts:
Installing the IBM Desktop Data Collector in stealth mode
Desktop collection processes on page 81
Using the delete policy with the IBM Desktop Data Collector: special
notes
When you use the IBM Desktop Data Collector to delete files from a desktop, they
are removed permanently. They are not transferred to the appliance or backed up
to any other location. Consequently, you must carefully review the infoset of
affected data objects prior to executing a delete action. Your organization may use
custom applications or other files that you may not want to delete. In reviewing
the returned list, do not allow the following to be deleted.
v Anything under this directory: c:\Windows
v Anything under this directory: Documents and Settings, with these extensions:
c:\Documents and Settings\<username>\UserData\ and extension *.xml
c:\Documents and Settings\<username>\Cookies\ and extension *.txt
c:\Documents and Settings\<username>\Start Menu\Programs\ and extension *.lnk
v Executable files
Utilizing desktop collection 83
*.dll
*.exe
*.ocx
v Drivers
*.sys
*.inf
*.pnf
v Installers
*.msi
*.mst
v Important data files
*.dat
*.ini
*.old
*.cat
v These file names
desktop.ini
ntuser.dat
index.dat
ntuser.pol
ntuser.dat.log
84 Administration Guide
Using Folders
This section provides both conceptual and procedural information regarding
folders and their usage.
The Folders tab displays two types of folders: Library and Workspace.
Library folder
Workspace folder
The Workspace folder is a custom folder that reflects your use of the system. By
default, it contains a folder entitled Templates. If you are using the system for IT
purposes, you may want to create folders for each locale or function.
Note: These folders can be renamed, moved, or deleted, and you also have the
options of setting folder security.
Creating a folder
To create a folder:
1. From the Folders tab, select New > New Folder. The Create new folder dialog
appears.
2. In the Name: field, give a name that represents the folders purpose (legal
matter, local, business unit, or the like).
3. In the Description: field, type a description for the folder.
4. In the Create in: field, use the list to select a place for the folder. Note that all
custom folders must be placed in the Workspace or a Workspace subdirectory;
you cannot add folders to the Library folder.
5. Click OK. If you wish to open the folder, click OK in the dialog that appears.
Related concepts:
Understanding folder types
Deleting a folder
When deleting folders, note that only empty folders can be deleted.
To delete a folder:
1. From within the Folders tab, Workspace folder, do either of the following:
a. Select the check box next to the folder you want to delete, and in the
Actions list, select Delete.
Moving a folder
To move a folder:
1. From within the Folders tab, Workspace folder, do either of the following:
a. Select the check box next to the folder you want to move, and in the
Actions list, select Move.
b. Right-click on the folder name and select Move.
2. Click OK, and in the Move items dialog, select the new location from the list.
3. Click OK.
Related concepts:
Understanding folder types on page 85
Renaming a folder
To rename a folder:
1. Right-click on the folder name and select Rename.
2. In the Rename folder dialog, change the Name and/or the Description.
3. Click OK.
Related concepts:
Understanding folder types on page 85
86 Administration Guide
5. Click Save.
Related concepts:
Understanding folder types on page 85
Using Folders 87
88 Administration Guide
Using audits and logs
This section describes the audit and log categories in the IBM StoredIQ Platform
system, including descriptions of the various audit types as well as how to view
and download details.
Data objects can be skipped during a harvest for a variety of reasons such as the
object being unavailable or a selected user option that excludes the data object
from the harvest. The Harvest details page lists all skipped data objects based on
file-system metadata level and content level.
All skipped harvest audit data and other files that have not been processed can be
downloaded for analysis.
This section lists the different fields seen when harvesting audits.
v Harvest audit by volume
v Harvest audit by time
v Harvest audit overview, including summary options, results options, and
detailed results
v Skipped data objects details
90 Administration Guide
Harvest Overview Detailed Results: Fields and Descriptions
Content type known, partial Content type known, but Content type known, but
processing complete error processing content cannot extract content
v This is the number of data v This is the number of data v This is the number of data
objects for which the objects for which the objects for which the
content type is known and content type is known, but content type is known, but
partial processing is an error was thrown while the content could not be
complete. processing content. extracted.
Content type unknown, not Binary text extracted, full Binary text extracted, partial
processed processing complete processing complete
v This is the number of data v This is the number of data v This is the number of data
objects for which the objects for which the objects for which the
content type is unknown binary text has been binary text has been
and has not been extracted and full extracted and partial
processed. processing has been processing has been
completed. completed.
Error processing binary Total
content v This lists the total number
v This is the number of data of data objects.
objects for which an error
was thrown while
processing binary content.
Related concepts:
Using audits and logs on page 89
This section lists the different fields seen when viewing import audits.
92 Administration Guide
Imports by Volumes Details: Fields and Descriptions
Total data objects imported Overwrite existing Status
v This is the total number of v If the import overwrote an existing v This is the status of the import:
imported data objects. volume, the status is Yes. If the Complete or Incomplete.
import did not overwrite an
existing volume, the status is No.
Related concepts:
Using audits and logs on page 89
Every action taken by the system and its users is captured by the event logs. These
actions include creating draft and published queries and tags, running policies,
publishing queries, deleting objects, configuring settings, and any other action
taken through the IBM StoredIQ Platform interface. A detailed of list of log entries
is provided in the event log messages.
You can view event logs for the current day or review saved logs from previous
days, and up to 30 days worth of logs can be viewed through the interface. If you
select and clear a day of logs, those logs are removed from the system.
Related reference:
Appendix C, Event log messages, on page 133
Subscribing to an event
To subscribe to an event:
1. Go to Audit > Event logs.
2. Click View all event logs, and the Event log for today page opens.
3. To the right of the event log to which you would like to subscribe, click
Subscribe. The Edit notification page appears.
4. In Destination, select the method by which you would like to be notified of
this event log. If you select Email address, be certain to use commas to
separate multiple email addresses.
5. Click OK.
Note: You can also subscribe to an event on the Dashboard. In the Event log
area, click Subscribe to the right of the event.
On the Administration > Dashboard, click Clear for the current view.
Policy audits provide a detailed history of the policy, including type of action, date
last executed, start and end dates with times, average speed, total data objects, and
data object counts. They can be viewed by name, volume, time, and by discovery
export.
This section lists the different fields seen when harvesting audits.
94 Administration Guide
Policy Audit by Volume: Fields and Descriptions
Volume Most recent date a policy Number of policies executed
v This is the name of the was executed v This is the number of
volume on which the v This is the most recent policies that were
policy was executed. date on which the policy executed.
was last executed.
Note: A warning in a policy audit trail is a success with the following conditions:
v If you copy an Exchange item such as re:, the re is copied, not the :. This will
generate a warning.
v The copied file is renamed.
v The file system to which you are copying does not accept characters in the file
name.
Note: To view the list of data objects, click on the [#] data objects link. To
create a report, click Create XML or Create PDF.
v Click Volume to open the policy audit by volume page.
v Click on a volume link to go to the Policy audit by time page.
v Click Time to see Audit by time page for the policy.
v On the Policy audit by time page, click the policy name to open the Policy
execution results page.
Note: To view the list of data objects, click the [#] data objects link. To create
a report, click Create XML or Create PDF.
v Click Discovery export.
v On the Policy audit by discovery export page, click the discovery export
name to open the Discovery export runs by production page. The page
details further information according to the incremental runs of the policy.
v Click a policy name to open the Policy executions by time page.
v Click on a policy name to open the Policy execution results page.
Note: To view the list of data objects, click the [#] data objects link. To create
a report, click Create XML or Create PDF.
As you review audit results through these pages, you can continue clicking
through to review various levels of information, from the volume and policy
execution level down to the data objects. To view more policy execution details,
click on the policy name in the execution summary page, which can be
accessed by any of the above policy views. As you continue browsing, IBM
StoredIQ Platform provides more detailed information such as:
v Source and destination settings
96 Administration Guide
v Policy optionsDetails of the policy actionThis section reflects the options
selected when creating the policy. Most attributes that appear depend upon
the type of policy run and the options available in the policy editor.
v Query (eitherIBM StoredIQ Platform or user-defined)
v View metadata linkThe view metadata page describes security details for
source and destination locations of the policy action.
Related concepts:
Understanding policy audits on page 94
Policy details
Execution details
Related concepts:
Understanding policy audits on page 94
98 Administration Guide
Saving results from an audit
You can save the results of policy executions into PDF and XML files. The
information can be saved as PDF and XML files. The exporting of information
appears as a running job on the dashboard until completed.
Success Data object is a duplicate of Data object skipped but Data object is a duplicate
[object name] will be loaded in load file. produced in a previous run
This applies to intermediate (discovery export only)
and files archives produced
during a discovery export
policy.
Data objects can receive a warning during a policy action if they fail to do any of
the following:
Set directory attributes Reset time stamps Set attributes Set time stamps
Set security descriptor Set access modes (Windows Set owner information Set group information
(Windows Share) Share) (NFS)
Set security permissions Create a link after a Find template to create a Extract text for the object
migrate (Windows Share, shortcut (Windows Share) (Discovery export policy)
NFS)
Data objects can receive one of the following failures during a policy action.
Data-Object Failures
Failed to create target Source does not exist Failed to find a new name
directory structure for the incoming object
Error copying data to target Could not copy due to Could not delete source after
network errors move
All modify actions failed File timed out waiting in the File under retention; cannot
pipeline be deleted (retention server)
Data objects are categorized in the other category during a discovery export policy
when:
v A data object is a member that makes its container responsive.
v A data object is a non-responsive member of a container.
Related concepts:
Understanding policy audits on page 94
3.07.0
Japanese
9.0
9.014.0
2002
2004
2005
NIST 3.0
X3
4.0
4.5
4.0
4.5
4.0
4.5
2002
3.x
6.0
8.013.0
2004
1.1
2.0
1998
2001
2004
1998
2001
2.0
2.0
2.0
2.0
X3
6.x
7.x
8.0
6.x
7.x
8.0
6.x
7.x
8.0
6.x
7.x
8.0
2000
2002
Related reference:
Appendix A, Supported file types, on page 101
MS Binder 7.01997
RAR RAR
Winzip ZIP
9.014.0
2002
2004
2005
dBXL 1.3
ENABLE 3.0
4.0
4.5
FoxBase 2.1
Framework 3.0
3.07.0
Japanese
9.0
NIST 3.0
3.x
6.0
2.0
PostScript PS level II
6.x
7.x
8.0
Visio (preview) 4
Visio 2003 5
2000
2002
WBMP n/a
10
MS Binder 7.01997
RAR RAR
Winzip ZIP
9.014.0
2002
2004
2005
dBXL 1.3
ENABLE 3.0
4.0
4.5
FoxBase 2.1
Framework 3.0
3.07.0
Japanese
9.0
NIST 3.0
3.x
6.0
2.0
PostScript PS level II
6.x
7.x
8.0
Visio (preview) 4
Visio 2003 5
2000
2002
WBMP n/a
10
X3
2.0
6.x
7.x
8.0
4.0
4.5
Framework 3.0
1.1
2.0
1998
2001
2004
MS MultiPlan 4.0
2.0
X3
6.x
7.x
8.0
VP Planner 3D 1.0
System
WML 5.2
ENABLE 3.0
4.0
4.5
Framework 3.0
2002
6.0
8.013.0
2004
MacWrite II 1.1
1998
2001
2.0
PFS: Write A
SAMNA Word IV
Sprint 1.0
Related reference:
Appendix A, Supported file types, on page 101
SharePoint attributes
This section describes the various SharePoint data object types and their properties
currently supported by IBM StoredIQ Platform.
Calendar
Recurring calendar events are indexed as a single object in IBM StoredIQ Platform.
Each recurring calendar events will have multiple Event Date and End Date
attribute values, one pair per recurrence. For instance, if there is an event defined
for American Independence Day and is set to recur yearly, it will be indexed with
Event Dates 2010-07-04, 2011-07-04, 2012-07-04, and so on.
Survey
Surveys have no full-text indexable body, and they are always indexed with
size=0.
The hash of a full-text indexed object is generally computed using the full-text
indexable body of the object. However, in the case of SharePoint list item objects
(excluding documents and pictures), the full-text indexable body might be empty
or too simplistic, meaning that you could easily obtain duplicate items across
otherwise two completely different objects. For this reason, other attributes are
included in the hash computation algorithm.
These attributes are included while computing the hash for the SharePoint data
objects, excluding documents and pictures.
Table 4. Hash-computation attributes
Attributes Types
Generic attributes v Title (SharePoint)
v Content Type (SharePoint)
v Description (SharePoint)
Blog post attributes v Post category (SharePoint)
Wiki page attributes v Wiki page comment
Calendar event attributes v Event category (SharePoint)
v Event date (SharePoint)
v Event end date (SharePoint)
v Event location (SharePoint)
Task or project task attributes v Task start date (SharePoint)
v Task due date (SharePoint)
v Task assigned to (SharePoint)
Contact attributes v Contact full name (SharePoint)
v Contact email (SharePoint)
v Contact job title (SharePoint)
v Contact work address (SharePoint)
v Contact work phone (SharePoint)
v Contact home phone (SharePoint)
v Contact mobile phone (SharePoint)
Link attributes v Link URL (SharePoint)
Survey attributes All survey questions and answers in the response are
included in the hash.
Related reference:
Appendix A, Supported file types, on page 101
Windows 2000
Windows 2003
Windows 2008
EMC Celerra Windows Share V1.0 DART 5.5 or later 4.4.2 When back-up
operator is set,
Celerra does not reset
access times.
NetApp Filer Windows Share V1.0 ONTAP 7.0.0 or later 4.4.2 Tested against 7.x
servers. NetApp 6.x
support until proven
otherwise.
Email Archives
Desktops
Event
Type Number Reason Sample Message Customer Action
ERROR 1001 Harvester was unable to Harvester could not Log into UTIL and restart
open a socket for listening allocate listen port after the application server.
to child processes. <number> attempts. Restart the data server.
Cannot kickstart Contact Customer Support.
interrogators. (1001)
ERROR 9083 Unexpected error while Exporting volume Contact Customer Support.
exporting a volume. 'dataserver:/mnt/demo-A'
(1357) has failed (9083)
ERROR 9086 Unexpected error while Importing volume Contact Customer Support.
importing a volume 'dataserver:/mnt/demo-A'
(1357) failed (9086)
ERROR 15001 No volumes are able to be No volumes harvested. Make sure IBM StoredIQ
harvested in a given job. (15001) Platform still has
For instance, all of the appropriate permissions to
mounts fail due to a a volume. Verify there is
network issue. network connectivity
between the data server
and your volume. Contact
Customer Support.
ERROR 15002 Could not mount the Error mounting volume Make sure the data server
volume. Check <share> <startdir> on still has appropriate
permissions and network server <server-name>. permissions to a volume.
settings. Reported <reason>. (15002) Verify there is network
connectivity between the
data server and your
volume. Contact Customer
Support.
ERROR 15021 Error saving harvest Failed to save Contact Customer Support.
record HarvestRecord for This message occurs due
qa1:auto-A (15021) to a database error.
ERROR 17001 Unhandled fatal exception Centera Harvester fatal Contact Customer Support.
in Centera Discovery. failure: <exception
description> (17001)
ERROR 17012 Error while trying to create Unable to create Centera Contact Customer Support.
a volume during Centera Volume This message occurs due
Discovery Company_jpool_2009_ to a database error.
FEB_1 in pool jpool.
Error:<database error
description> (17012)
ERROR 17505 Unable to query object Unable to determine object Contact Customer Support.
count for a discovered count for <server>:<share>
volume. This is likely due
to a database error.
ERROR 17506 Generic retention Error creating volume Contact Customer Support.
discovery could not create <server>:<share:>
discovered volume.
ERROR 18001 SMB connection fails. Windows Share Protocol Make sure IBM StoredIQ
Exception when Platform still has
connecting to the server appropriate permissions to
<server-name> : <reason>. a volume. Verify there is
(18001) network connectivity
between the data server
and your volume. Contact
Customer Support.
ERROR 18002 SMB volume mount failed. Windows Share Protocol Verify the name of the
Check the share name. Exception when server and volume to
connecting to the share make sure they are correct.
<share-name> on If this message persists,
<server-name> : <reason>. then Contact Customer
(18002) Support.
ERROR 18003 There is no volume Windows Share Protocol Contact Customer Support.
manager. Exception while
initializing the data object
manager: <reason>.
(18003)
ERROR 18006 Grazer volume crawl Grazer._run : Unknown Verify the user that
threw an exception. error during walk. (18006) mounted the specified
volume has permissions
equivalent to your current
backup solution. If this
message continues, contact
Customer Support.
ERROR 18021 An unexpected error from Unable to fetch trailing Check to ensure the
the server prevented the activity stream from NewsGator server has
harvest to reach the end of NewsGator volume. Will sufficient resources (disk
the activity stream on the retry in next harvest. space, memory, and so on).
NewsGator data source (18021) It is very likely that this
being harvested. The next error is transient. If the
incremental harvest will error persists across
attempt to pick up from multiple harvests, contact
where the current harvest Customer Support.
was interrupted.
ERROR 18018 Start directory has escape Cannot graze the volume, Consider turning off
characters, and the data root directory Nez has escape character checking.
server is configured to escape characters (18018)
skip them.
ERROR 33003 Could not mount the Unable to mount the Verify user name and
volume. Check volume: <error reason>. password used for
permissions and network (33003) mounting the volume are
settings. accurate. Check the user
data object for appropriate
permissions to the volume.
Make sure the volume is
accessible via one of our
built-in protocols (NFS,
Windows Share, or
Exchange). Verify that the
network is properly
configured for the data
server to reach the
volume. Verify the data
server has appropriate
DNS settings to resolve the
server name.
ERROR 33004 Volume could not be Unmounting volume failed Reboot the data server. If
unmounted. from mount point : the problem persists, then
<mount point>. (33004) Contact Customer Support.
ERROR 33005 Data server was unable to Unable to create Reboot the data server. If
create a local mounting mount_point using the problem persists, then
point for the volume. primitive.threadSafe contact Customer Support.
Makedirs(). (33005)
ERROR 33010 Failed to make SMB Mounting Windows Share Verify user name and
connection to Windows volume failed with the password used for
Share server. error : <system error mounting the volume are
message>. (33010) accurate. Check the user
data object for appropriate
permissions to the volume.
Make sure the volume is
accessible via one of our
built-in protocols
(Windows Share). Verify
that the network is
properly configured for
the data server to reach
the volume. Verify the
data server has
appropriate DNS settings
to resolve the server name.
ERROR 33011 Internal error. Problem Unable to open Reboot the data server. If
accessing local /proc/mounts. Cannot test the problem persists, then
/proc/mounts if volume was already contact Customer Support.
mounted. (33011)
ERROR 33012 Database problems when An exception occurred Contact Customer Support.
deleting a Volume while working with
HARVESTS_TABLE in
Volume._delete(). (33012)
ERROR 33013 No volume set was found Unable to load volume set Contact Customer Support.
for the given volumes set by its name. (33013)
name.
ERROR 33014 System could not An error occurred while Contact Customer Support.
determine when this performing the
volume was last harvested. last_harvest operation.
(33014)
ERROR 33018 An error occurred Mounting Exchange Server Verify user name and
mounting the Exchange failed : <reason>. (33018) password used for
share. mounting the share are
accurate. Check for
appropriate permissions to
the share. Make sure the
share is accessible. Verify
that the network is
properly configured for
the data server to reach
the share. Verify the data
server has appropriate
DNS settings to resolve the
server name.
ERROR 33019 Failed to connect and Mounting HCAP volume Ensure the connectivity,
authenticate to the Hitachi failed : Cannot connect to credentials and
Archivas Content Archive HCAP share. (33019) permissions to the Hitachi
server. volume and retry.
ERROR 33020 Failed to connect and Mounting IBM Information Ensure the connectivity,
authenticate to the IBM Archive volume failed credentials and
Information Archive with the error: Server permissions to the IBM
retention volume. unreachable. (33020) Information Archive
volume and retry.
ERROR 33022 Failed to connect to the Mounting Discovery Verify the information
Discovery Accelerator Accelerator volume failed used to add the volume
using the information for with the error: insufficient and ensure all details have
the Volume. permissions to review been entered correctly
CaseOne (33022) before retrying. If the error
points to network issues
with connectivity, address
them and retry.
ERROR 33027 The attempt to connect Mounting IBM FileNet Ensure the connectivity,
and authenticate to the retention volume failed : credentials, and
IBM FileNet server failed. <reason>. (33027) permissions to the FileNet
volume and retry.
ERROR 34002 Could not complete the Copy Action aborted as Verify there is space
copy action because the the target disk has run out available on your policy
target disk was full. of space (34002) destination and try again.
ERROR 34009 Could not complete the Move Action aborted as Verify there is space
move action due to full the target disk has run out available on your policy
target disk. of space. (34009) destination, then run
another harvest before
executing your policy.
When the harvest
completes, try running the
policy again.
ERROR 34015 The policy audit could not Error Deleting Policy Contact Customer Support.
be deleted for some Audit: <error message>
reason. (34016)
ERROR 34020 The copy to Centera action Copy to Centera failed as Check permissions on the
could not be executed we do not have read/write access profile provided for
because of insufficient permissions on the access the Centera pool on which
permissions. profile used. (34020) the volume has been
defined, and check if the
proper permissions have
been provided.
ERROR 34021 The move to Centera Move to Centera failed as Check permissions on the
action could not be we do not have read/write access profile provided for
executed because of permissions on the access the Centera pool on which
insufficient permissions. profile used. (34020) the volume has been
defined, and check if the
proper permissions have
been provided.
ERROR 34026 HSM action cannot recall a HSM Stub action aborted Verify there is space
file because the primary because the primary disk available on the volume
volume does not have has run out of space that the file is being
enough free space. (34026) recalled to and run the
HSM action again.
ERROR 34030 Discovery export policy is Production Run action Create sufficient space on
aborted since it detected aborted because the target target disk and run
the target disk is full. disk has run out of space. discovery export policy
(34030) again.
ERROR 34034 The target volume for the Copy objects failed, unable Ensure the connectivity,
policy could not be to mount volume: login credentials and
mounted. The policy will QA1.COMPANY. permissions to the target
be aborted. COM:SHARE. (34034) volume for the policy and
retry.
ERROR 41004 The job terminated <job-name> ended Try to run the job again. If
abnormally. unexpectedly. (41004) it fails again, contact
Customer Support.
ERROR 41007 Job has failed. [Job name] has failed Look at previous messages
(41007). to see why it failed and
refer to that message ID to
pinpoint the error. Contact
Customer Support.
ERROR 42001 The copy action could not Copy data objects did not Contact Customer Support.
run because of parameter run. Errors
errors. occurred:<error-
description>. (42001)
ERROR 42002 The copy action was Copy data objects failed, Check permissions on the
unable to create a target unable to create target target. Make sure the
directory. dir:<target- permissions that are
directory-name>. (42002) configured to mount the
target volume have write
access to the volume.
ERROR 42004 An unexpected error Copy data objects Contact Customer Support.
occurred. terminated abnormally.
(42004)
ERROR 42006 The move action could not Move data objects did not Contact Customer Support.
run because of parameter run. Errors
errors. occurred:<error-
description>. (42006)
ERROR 42007 The move action was Move data objects failed, Check permissions on the
unable to create a target unable to create target target. Make sure the
directory. dir:<target-directory- permissions that are
name>. (42007) configured to mount the
target volume have write
access to the volume.
ERROR 42009 An unexpected error Move data objects Contact Customer Support.
occurred. terminated abnormally.
(42009)
ERROR 42017 An unexpected error Delete data objects Contact Customer Support.
occurred. terminated abnormally.
(42017)
ERROR 42025 The policy action could Policy cannot execute. Contact Customer Support.
not run because of Attribute verification
parameter errors. failed. (42025)
ERROR 42050 The data synchronizer Content Data Synchronizer Contact Customer Support.
could not run because of synchronization of
an unexpected error. <server-name>:
<volume-name> failed
fatally.
ERROR 42059 Illegal set of parameters Production Run on objects Contact Customer Support.
passed to discovery export did not run. Errors
policy. occurred: The following
parameters are missing:
action_limit. (42059)
ERROR 42060 Discovery export policy Production Run on objects Verify the Discovery
failed to create target (Copying native objects) export volume has write
directory for the export. failed, unable to create permission and re-execute
target dir: production/10. policy.
(42060)
ERROR 42062 Discovery export policy Production Run on objects Contact Customer Support.
was terminated (Copying native objects)
abnormally. terminated abnormally.
(42062)
ERROR 42088 The full-text optimization Full-text optimization Contact Customer Support.
process failed; however, failed on volume
the index is most likely <volume-name> (42088)
still usable for queries.
ERROR 45802 A full-text index is already Time allocated to gain Contact Customer Support.
being modified. exclusive access to
in-memory index for
volume= 1357 has expired
(45802)
ERROR 45803 The index for the specified Index '/deepfs/fulltext/ No user intervention is
volume does not exist. volume_index/ required.
This message may occur volume_1357' not found
under normal conditions. (45803)
ERROR 45805 The query has not been Query ID: 123 does not No user intervention is
started or has expired The exist (45805) required.
forme.r is a programming
error, the latter is normal.
ERROR 45806 The query expression is Failed to parse 'dog pre\3 Revise your full-text query.
invalid or not supported. bar' (45806)
ERROR 45808 A transaction has never No transaction for client: No user intervention is
been started or has node.client.com_FINDEX_ required. The system
expired. QUEUE_1357_1172515222_3_2 handles this condition
(45808) internally.
ERROR 45812 A File I/O error occurred Failed to write disk Try your query again.
while accessing index data. (45812). Contact Customer Support
for additional assistance if
necessary.
ERROR 45814 The query expression is Query: 'a* b* c* d* e*' is Refine your full-text query.
too long. too complex (45814)
ERROR 45815 The file that is being Java heap exhausted while Check the skipped file list
indexed is too large or the indexing node with ID: in the audit log for files
query expression is too '10f4179cd5ff22f2a6b that failed to load due to
complex. The engine has 79a1bc3aef247fd94ccff' their sizes. Revise your
temporarily run out of (45815) query expression and retry.
memory.
ERROR 46023 Tar command failed while Failed to back up fulltext Check disk space and
persisting full-text data to data for server:share. permissions.
Windows Share or NFS Reason: <reason>. (46023)
share.
ERROR 46024 Unhandled fatal exception Exception <exception> Contact Customer Support.
while persisting full-text while backing up fulltext
data into a .tgz file. data for server:share
(46024)
ERROR 46025 Was not able to delete Failed to unlink Check permissions.
partial .tgz file after a incomplete backup image.
failed full-text backup. Reason: <reason>. (46025)
ERROR 47002 Synchronization failed on Synchronization failed for Contact Customer Support.
a query. query '<query-name>' on
volume
'<server-and-volume>
(47002)
ERROR 47101 An error occurred during Cannot process full-text Restart services and
the query of a full-text expression (Failed to read contact Customer Support.
expression. from disk (45812) (47101)
ERROR 47207 User is running out of disk Disk usage exceeds Contact Customer Support.
space. threshold. (%d) In rare cases, this message
can indicate a program
error leaking disk space. In
most cases, however, disk
space is almost full, and
additional storage is
required.
ERROR 47212 Interrogator crashed while Harvester 1 Does not exist. If the problem persists
processing a file. The Action taken : restart. (that is, the system crashes
current file will be missing (47212) on the same file or type of
from the volume cluster. files), contact Customer
Support.
ERROR 47214 SNMP notification sender Unable to resolve host Check spelling and DNS
is unable to resolve the name setup.
trap host name. nomachine.nowhere.com
(47214)
ERROR 50011 The DDL/DML files Database version control Contact Customer Support.
required for the database SQL file not found. (50011)
versioning were not found
in the expected location on
the data server.
ERROR 50018 Indicates that the Database restore is Contact Customer Support.
pre-upgrade database unsuccessful. Contact
restoration failed, which Customer Support. (50018)
was attempted as a result
of a database upgrade
failure.
ERROR 50020 Indicates the current Versions do not match! Contact Customer Support.
database requirements do Expected current database
not meet those specified version: <dbversion>.
for the upgrade and (50020)
cannot proceed with the
upgrade.
ERROR 50021 Indicates that the full Database backup failed. Contact Customer Support.
database backup failed (50021)
when attempting a
data-object level database
backup.
ERROR 61005 The discovery export load Production load file Contact Customer Support.
file generation fails generation failed. Load
unexpectedly. The load files may be produced, but
files may be produced post-processing may be
correctly, but incomplete. (61005)
post-processing actions
like updating audit trails
and generating report files
may not have completed.
ERROR 61006 The discovery export load Production load file Free up space on the target
file generation was generation interrupted. disk, void the discovery
interrupted because the Target disk full. (61006) export run and run the
target disk is full. policy once more.
ERROR 68001 The gateway and data Gateway connection failed Update your data server to
server must be on the due to unsupported data the same build number as
same version in order to server version. the gateway and restart
connect. services. If your encounter
issues, contact Customer
Support.
ERROR 68003 The data server has failed The data-server connection Contact Customer Support.
to connect to the gateway to the gateway cannot be
over an extended period of established.
time.
ERROR 80002 The system failed to open Failed to connect to the The "maximum database
a connection to the database (80002). connections" configuration
database. parameter of the Database
engine may need to be
increased. Contact
Customer Support.
Related reference:
Appendix C, Event log messages, on page 133
Event
Type Number Reason Sample Message Customer Action
INFO 9001 No conditions have been Harvester: Query <query Add conditions to the
added to a query. name> cannot be inferred specified query.
because no condition for it
has been defined (9001).
INFO 9002 One or more conditions in Harvester: Query < query Verify that regular
a query were incorrect. name> cannot be inferred expressions are properly
because of regular formed.
expression or other
condition error (9002).
INFO 9012 Indicates the end of Dump of Volume cache(s) No user intervention is
dumping the content of completed (9012). required.
the volume cache.
INFO 9013 Indicates the beginning of Postprocessing for volume No user intervention is
the load process. 'Company Data required.
Server:/mnt/demo-A'
started (9013).
INFO 9067 Indicates load progress. System metadata and No user intervention is
tagged values were required.
successfully loaded for
volume 'server:volume'
(9067).
INFO 9069 Indicates load progress. Volume 'data server: No user intervention is
/mnt/demo-A': System required.
metadata, tagged values
and full-text index were
successfully loaded (9069).
INFO 9084 The volume export Exporting volume 'data No user intervention is
finished. server:/mnt/demo-A' required.
(1357) completed (9084)
INFO 9091 The load process has been Load aborted due to user No user intervention is
aborted by the user. request (9091). required.
INFO 15008 The volume load step has Post processing skipped No user intervention is
been skipped, per user for volume required.
request. <server>:<volume>.
(15008)
INFO 15009 The volume load step has Harvest skipped for No user intervention is
been executed but the volume required.
harvest step has been <server>:<volume>.
skipped, per user request. (15009)
INFO 15012 The policy running on the Volume <volume> on No user intervention is
volume has completed and server <server> is free required.
the volume load can now now. Proceeding with load.
proceed. (15012)
INFO 15013 The configured time limit Harvest time limit reached No user intervention is
on a harvest was reached. for server:share. Ending required.
harvest now. (15013)
INFO 15014 The configured object Object count limit reached No user intervention is
count limit on a harvest for server:share. Ending required.
was reached. harvest now. (15014)
INFO 15017 Check box selected for Deferring post processing No user intervention is
nightly load job. for volume server:vol required.
(15017)
INFO 15018 Harvest size and/or time Harvest limit reached on No user intervention is
limit is reached. server:volume. Synthetic required.
deletes will not be
computed. (15018)
INFO 15019 User stops harvest process. Harvest stopped by user No user intervention is
while processing volume required.
dpfsvr:vol1. Rest of
volumes will be skipped.
(15019)
INFO 15020 Harvest vocabulary has Vocabulary for Full harvest should be run
changed. Full harvest dpfsvr:jhaide-A has instead of an incremental
should run instead of changed. A full harvest is harvest.
incremental. recommended (15020).
INFO 15022 The user is trying to Permission-only harvest: No action needed as the
execute an ACL-only permission checks not volume is skipped.
harvest on a volume that supported for
is not a Windows Share or <server>:<share>
SharePoint volume.
INFO 15023 The user is trying to Permission-only harvest: No action needed as the
execute an ACL-only volume <server>:<share> volume is skipped.
harvest on a volume that has no associated user list.
has not been assigned a
user list.
INFO 17002 Sent when Centera Centera External Iterator : No user intervention is
Discovery sends the query Starting to populate using required.
to the Centera server pool QAPOOL. (17002)
INFO 17004 A Centera Discovery Object limit reached for No user intervention is
auto-created volume has _QAPOOL_2009_JAN_1, required.
reached the preconfigured starting new volume
limit, starting a new one. (17004)
INFO 17007 Pending data return from Centera Harvester: No Check if a Centera node is
Centera. items returned from down. A Centera cluster
Centera for over 5 may be overloaded.
minutes. Still waiting.
(17007)
INFO 17009 Configured time limit Centera Discovery : time No user intervention is
reached for Centera limit for discovery required.
discovery step. reached. Ending this run.
(17009)
INFO 17507 Limit (time or object Retention discovery limit Contact Customer Support.
count) reached for generic reached for
retention discovery. <server>:<share>
INFO 18016 Displays the list of top Choosing top level No user intervention is
level directories selected directories: <directories> required.
by matching the start (18016)
directory regular
expression. Displays at the
beginning of a harvest.
INFO 34001 Marks current progress of <volume>: <count> data No user intervention is
a copy action. objects processed by copy required.
action. (34001)
INFO 34004 Marks current progress of <volume>: <count> data No user intervention is
a delete action. objects processed by delete required.
action. (34004)
INFO 34008 Marks current progress of <volume>: <count> data No user intervention is
a move action. objects processed by move required.
action. (34008)
INFO 34014 A policy audit was Deleting Policy Audit # No user intervention is
deleted. <audit id> <policy name> required.
<start time> (34014)
INFO 34015 A policy audit was Deleted Policy Audit # No user intervention is
deleted. <audit id> <policy name> required.
<start time> (34015)
INFO 41001 A job was started either <jobname> started. (41001) No user intervention is
manually or was required.
scheduled.
INFO 41002 The user stopped a job <jobname> stopped at user No user intervention is
that was running. request (41002) required.
INFO 41006 Rebooting or restarting Service shutdown. Rerun jobs after reboot or
services on the controller Stopping outstanding jobs. restart if you want the jobs
or compute node causes (41006) to complete.
all jobs to stop.
INFO 41008 Database compactor Database compactor was Set the database
(vacuum) job can not run not run because other jobs compactor's job schedule
while there is database are active (41008). so that it doesnt conflict
activity. with long-running jobs.
INFO 42005 The action completed or Copy complete: <number> No user intervention is
was aborted. Shows results data objects required.
of copy action. copied,<number>
collisions found. (42005)
INFO 42010 The action completed or Move complete: <number> No user intervention is
was aborted. Shows results data objects required.
of move action. moved,<number>
collisions found. (42010)
INFO 42018 The action completed or Copy data objects No user intervention is
was aborted. Shows results complete: <number> data required.
of deleted action. objects copied,<number>
collisions found. (42018)
INFO 42028 The action completed or Policy completed (42028). No user intervention is
was aborted. Shows results required.
of policy action.
INFO 42032 The action completed or <report name> completed No user intervention is
was aborted. Shows results (42032). required.
of report action.
INFO 42033 The synchronizer started Content Data Synchronizer No user intervention is
automatically or manually started. (42033) required.
with the GUI button.
INFO 42048 Reports that the Content Data Synchronizer No user intervention is
synchronizer is skipping a skipping required.
volume if synchronization <server-name>:<volume-
is determined not to be name> as it does not need
required. synchronization. (42048)
INFO 42049 Reports that the Content Data Synchronizer No user intervention is
synchronizer has started starting synchronization required.
synchronization of a for volume
volume <server-name>:<volume-
name>
INFO 42053 The policy, that was Proceeding with execution No user intervention is
waiting for participant of <policy-name>. required.
volumes to be loaded
before continuing, is now
starting.
INFO 42065 A discovery export policy Proceeding with execution No user intervention is
that was held up for want of 'Production case One'. required.
of resources, is now done (42065)
waiting, and will begin
execution.
INFO 42066 A new discovery export New run number 10 Note the new run number
run has started. started for production in order to tie the current
Production Case 23221. run with the
(42066) corresponding audit trail.
INFO 42067 Discovery export policy is Production Run producing No user intervention is
preparing the audit trail in Audit Trail XML. (42067) required.
XML format. This may
take a few minutes.
INFO 42074 A query or tag has been Successfully sent query No user intervention is
replicated to a member 'Custodian: Joe' to member required.
data server successfully. data server San Jose Office
(42074)
INFO 46001 The backup process has Backup Process Started. No user intervention is
begun. Any selected (46001) required.
backups in the system
configuration screen will
be run if necessary.
INFO 46002 The backup process did Backup Process Failed: Check your backup
not complete all its tasks <error-description>. volume.
successfully. One or more (46002)
backup types did not
occur.
INFO 46003 The backup process Backup Process Finished. No user intervention is
successfully completed (46003) required.
attempting all the
necessary tasks. Any parts
of the overall process add
their own log entries.
INFO 46004 The Application Data Application Data backup Check your backup
backup, as part of the failed. (46004) volume. Look at the setup
overall backup process, for the Application Data
needed to run but did not backup. If backups
succeed. continue to fail, Contact
Customer Support.
INFO 46005 The Application Data Application Data backup No user intervention is
backup, as part of the finished. (46005) required.
overall backup process,
needed to run and
succeeded.
INFO 46006 The Application Data Application Data backup No user intervention is
backup, as part of the not configured, skipped. required.
overall backup process, (46006)
was not configured.
INFO 46007 The Harvested Volume Harvested Volume Data Check your backup
Data backup, as part of the backup failed. (46007) volume. Look at the setup
overall backup process, for the Harvested Volume
needed to run but did not Data backup. If backups
succeed. continue to fail, contact
Customer Support.
INFO 46008 The Harvested Volume Harvested Volume Data No user intervention is
Data backup, as part of the backup finished. (46008) required.
overall backup process,
needed to run and
succeeded.
INFO 46009 The Harvested Volume Harvested Volume Data No user intervention is
Data backup, as part of the backup not configured, required.
overall backup process, skipped. (46009)
was not configured
INFO 46010 The System Configuration System Configuration Check your backup
backup, as part of the backup failed. (46010) volume. Look at the setup
overall backup process, for the System
needed to run but did not Configuration backup. If
succeed. backups continue to fail,
contact Customer Support.
INFO 46013 The Audit Trail backup, as Policy Audit Trail backup Check your backup
part of the overall backup failed. (46013) volume. Look at the setup
process, needed to run but for the Audit Trail backup.
did not succeed. If backups continue to
fail, Contact Customer
Support.
INFO 46014 The Audit Trail backup, as Policy Audit Trail backup No user intervention is
part of the overall backup finished. (46014) required.
process, needed to run and
succeeded.
INFO 46015 The Audit Trail backup, as Policy Audit Trail backup No user intervention is
part of the overall backup not configured, skipped. required.
process was not (46015)
configured.
INFO 46019 Volume cluster backup Indexed Data backup Contact Customer Support.
failed. failed: <specific error>
(46019)
INFO 46020 Volume cluster backup Indexed Data backup No user intervention is
finished. finished. (46020) required.
INFO 46021 Volume is not configured Indexed Data backup not No user intervention is
for indexed data backups. configured, skipped. required.
(46021)
INFO 60001 The user updates an object Query cities was updated No user intervention is
on the system. This by the administrator required.
includes any object type account (60001).
on the data server,
including the updating of
volumes.
INFO 60002 The user creates an object. Query cities was created No user intervention is
This includes any object by the administrator required.
type on the data server, account (60002).
including the creation of
volumes.
INFO 60003 The user deletes an object. Query cities was deleted No user intervention is
This includes any object by The administrator required.
type on the data server, account (60003).
including the deletion of
volumes.
INFO 60004 The user publishes a Query cities draft was No user intervention is
full-text query set or a published by the required.
query. administrator account
(60004).
INFO 60005 The user tags an object. Query tagging for cities No user intervention is
This includes a published class was started by the required.
query, a draft query, or administrator account
tag. (60005).
INFO 60006 A user restarted services Application services restart No user intervention is
on the data server. for all data servers was required.
requested by the
administrator account
(60006).
INFO 61002 Concordance discovery Load file(s) ready for No user intervention is
export is ready to upload upload. (61002) required.
the load file(s).
INFO 65000 The log file has finished Log file download No user intervention is
downloading complete (65000) required.
Related reference:
Appendix C, Event log messages, on page 133
Event
Type Number Reason Sample Message Customer Action
WARN 1002 An Interrogator process Processing could not be Classify the document
died because of an completed on object, manually and Contact
unknown error. The data interrogator died : <data Customer Support.
object it was processing object name>. (1002)
will be skipped. A new
process will be created to
replace it.
WARN 1003 Interrogator child process Interrogator terminated Try to re-add the volume
did not properly get before accessing data being harvested. If that
started. There may have objects. (1003) fails, Contact Customer
been problems accessing Support.
the Volume to be
harvested.
WARN 1004 Interrogator child process Processing was not Contact Customer Support.
was terminated because it completed on object,
was no longer responding. interrogator killed : <data
The data object it was object name>. (1004)
processing will be skipped.
A new process will be
created to replace it.
WARN 6001 A user email could not be Failed to send an email to Verify that your SMTP
sent. The mail server user <email address>; server is configured
settings are incorrect. check mail server correctly. Make sure the IP
configuration settings address configured for the
(6001). data server is permitted to
relay on the configured
SMTP server.
WARN 8001 The database needs to be The Database is Run the Database
vacuumed. approaching an maintenance task to
operational limit. Please vacuum the database.
run the Database
maintenance task using the
Console interface (8001)
WARN 9068 Tagged values were System metadata and Contact Customer Support.
loaded, but full-text index tagged values were loaded
loading failed. successfully for volume
'server:volume', but
loading the full-text index
failed (9068)
WARN 9070 Tagged values and full-text Loading system metadata, Contact Customer Support.
index loading failed. tagged values and the
full-text index failed for
volume 'server:volume'
(9070)
WARN 15003 The volume mount Volume <volume name> Contact Customer Support.
appeared to succeed, but on server <server name> is
the test for mount failed. not mounted. Skipping.
(15003)
WARN 15005 There was a component [<component>] Run Contact Customer Support.
run failure. failure. (15005)
WARN 15006 Cleanup failed for [<component>] Cleanup Contact Customer Support.
component after a run failure on abort. (15006)
failure.
WARN 15007 A component timed out Component Try your action again. If
and needs to be stopped. [<component>] this error continues,
unresponsive; contact Customer Support.
autostopping triggered.
(15007)
WARN 15010 The same volume cannot Volume <volume-name> No user intervention is
be harvested in parallel. on server <server-name> is required. You may wish to
The harvest will be already being harvested. verify that the volume
skipped and the next one, Skipping. (15010) harvest has completed.
if any are in queue,
started.
WARN 15015 Configured harvest time Time limit for harvest Reconfigure harvest time
limit reached. reached. Skipping Volume limit.
v1 on server s1. 1 (15015)
WARN 15016 Configured harvest object Object count limit for Reconfigure harvest data
count limit reached. harvest reached. Skipping object limit.
Volume v1 on server s1
(15016)
WARN 17008 Query ran to discover Centera External Iterator : Contact Customer Support.
Centera items terminated Centera Query terminated
unexpectedly. unexpectedly (<error
description>). (17008)
WARN 17011 Running discovery on the Pool Jpool appears to have Make sure that two jobs
same pool in parallel is another discovery running. are not running at the
not allowed. Skipping. (17011). same time that discover
the same pool.
WARN 17504 Sent when a retention Volume <server>:<share> Contact Customer Support.
discovery is run on any is not supported for
volume other than a discovery. Skipping.
Windows Share retention
volume.
WARN 18007 Directory listing or Walker._walktree: OSError Make sure the appliance
processing of data object - <path><reason> (18007) still has appropriate
failed in Grazer. permissions to a volume.
Verify there is network
connectivity between the
appliance and your
volume. Contact Customer
Support.
WARN 18010 The skipdirs file is either Unable to open skipdirs Contact Customer Support.
not present or not readable file: <filename>. Cannot
by root. skip directories as
configured. (18010)
WARN 18011 An error occurred reading Grazer._run: couldn't read Contact Customer Support.
the known extensions list extensions - <reason>.
from the database. (18011)
WARN 18012 An unknown error Grazer._run: couldn't read Contact Customer Support.
occurred reading the extensions. (18012)
known extensions list from
the database.
WARN 18015 NFS initialization warning NIS Mapping not User name and group
that NIS is not available. available. (18015) names may be inaccurate.
Please check that your NIS
server is available and
properly configured in the
data server.
WARN 18017 A folder in an Enterprise Skipping Folder (ID=3) in If all folders were expected
Vault case being harvested volume to be harvested in the
was skipped because of evdiscaccel.company.com: volume, verify that the
insufficient permissions. CaseOne due to username used to add the
insufficient permissions. Volume has Folder Review
Check Review permissions permission on all folders.
on the folder. (18017)
WARN 18019 The checkpoint saved from Unable to load checkpoint If the message repeats in
the last harvest of the for NewsGator volume. A subsequent harvests,
NewsGator data source full harvest will be contact Customer Support.
failed to load. Instead of performed instead. (18019)
performing an incremental
harvest, a full harvest will
be executed.
WARN 18020 The checkpoint noted for Unable to save checkpoint If the message repeats in
the current harvest of the for NewsGator harvest of subsequent harvests,
NewsGator data source volume. (18020) contact Customer Support.
could not be saved. The
next incremental harvest of
the data source will not be
able to pick up from this
checkpoint.
WARN 33016 System could not unmount Windows Share Protocol Server administrators may
this volume. Session teardown failed. see connections left
(33016) hanging for a predefined
period of time. These
connections will drop off
after they time out. No
user intervention required.
WARN 33017 System encountered an An error occurred while Contact Customer Support.
error while trying to figure retrieving the query
out what query use this instances pointing to a
volume. volume. (33017)
WARN 34003 Skipped a copy data object Copy action error :- Target Please verify there is space
because disk full error. disk full, skipping copy : available on your policy
<source volume> to destination and try again.
<target volume>. (34003)
WARN 34010 Skipped a move data Move action error :- Target Please verify there is space
object because disk full disk full, skipping copy : available on your policy
error. <source volume> to destination. After verifying
<target volume>. (34010) space is available, please
run another harvest before
executing your policy.
Upon harvest completion,
please try running the
policy again.
WARN 34029 Discovery export policy Discovery export Run Create sufficient space on
detects the target disk is action error: Target disk target disk and run
full and skips production full, skipping discovery discovery export policy
of an object. export: again.
share-1/saved/years.pdf
to production/10/
documents/1/0x0866e
5d6c898d9ffdbea720b0
90a6f46d3058605.txt.
(34029)
WARN 34032 The policy being executed No volumes in scope for Check policy query and
has no volumes in scope policy. Skipping policy scoping configuration, and
based on the configured execution. (34032) re-execute policy.
query and scoping. The
policy will not be
executed.
WARN 34033 Celerra data mover error. Data mover returned Consult the Celerra
There are a large number NO_MATCHING_CONNECTION Administrator Manual.
of possible causes; the : The specified
actual one is listed in the OFFLINE_PATH was not
error message. The sample able to map to an existing
provides the most connection to a secondary
common one. server. (34033)
WARN 34035 If the global hash setting Copy objects : Target hash If target hashes need to be
for the system is set to not will not be computed computed for the policy
compute data object hash, because Hashing is audit trail, turn on the
no hash can be computed disabled for system. global hash setting before
for the target objects (34035) executing the policy.
during a policy action.
WARN 34036 The policy has no source The policy has no source Confirm that the query
volume(s) in scope, volume(s) in scope. Wait used by the policy has one
meaning that the policy for the query to update or more volumes in scope.
cannot be executed. before executing the policy.
(34036)
WARN 42003 The job containing this Copy data objects stopped No user intervention is
action is stopped by the at user request. (42003) required.
user.
WARN 42008 The job containing this Move data objects stopped No user intervention is
action is stopped by the at user request. (42008) required.
user.
WARN 42016 The job containing this Delete data objects No user intervention is
action is stopped by the stopped at user request. required.
user. (42016)
WARN 42026 The job containing this Policy stopped at user No user intervention is
action is stopped by the request. (42026) required.
user.
WARN 42035 When the job containing Set security for data No user intervention is
this action is stopped by objects stopped at user required.
the user. request. (42035)
WARN 42051 Two instances of the same Policy <policy-name> is No user intervention is
policy cannot run at the already running. Skipping. required.
same time. (42051)
WARN 42052 Policies will wait to One or more volume(s) No user intervention is
execute until after volumes needed by policy required.
have loaded, if those <policy-name> are being
volumes are participants to loaded. Waiting for the
the policy by virtue of bulk load(s) to finish.
being in the query. (42052)
WARN 42061 Discovery export policy Discovery export run on No user intervention is
was stopped by user. objects (Copying native required.
objects) stopped at user
request. (42061)
WARN 42064 Discovery export policy A Discovery export run The discovery export
execution has been related to policy policy execution is held up
delayed because a 'Discovery export case for required resources.
conflicting discovery One' is in progress. Execution should begin as
export run is in progress. Waiting for it to finish. soon as resource becomes
(42064) available.
WARN 42068 Policy failed to set Copy objects warning, Note that policy may not
appropriate permissions unable to set permissions be able to set appropriate
on the target directory. on target directory: permissions on the objects
Objects created from the share-1/saved. (42068) it creates. If this is not
policy may not have acceptable, verify that
appropriate permissions target volume has proper
set. write permissions and
re-execute.
WARN 42069 If the Copy data objects Discovery export If the modified objects
modified since last DAT_Export is configured need to be acted upon,
harvest option is selected to act on members of either use a discovery
for a discovery export containers, and cannot act export acting only on the
policy, it is only valid if on objects modified after original file/email archive,
the discovery export itself the last harvest. Discovery or perform an incremental
is defined to act on the export run X will skip harvest on the source
original file/email archive, modified objects. (42069) volume(s).
as opposed to their
members. If this is not
the case, the warning lets
the user know that
modified objects will still
be skipped.
WARN 46026 Volume is being harvested Volume volume:share is in Rerun backup when
or policies are running use. Unable to back up volume is not in use.
against it. If there are other full-text index. Will retry
full-text indexes to be later. (46026)
backed up, the system
works on those and retries
this volume later.
WARN 47202 The system is starting to Database connections Contact Customer Support.
run low on database usage seems excessive
connections. This is (512/415) (47202)
abnormal, and an
indication of process
restarts and connections
not being cleared.
WARN 47215 Someone internally or SSHD: Failed password for Contact your local IT
externally is trying (and root from 172.17.18.185 manager. This could be
failing) to SSH into the port 57982. (47125) either a mistyped
data server. password by a legitimate
user or in the worst case
scenario, a genuine
break-in attempt.
WARN 61003 One of the load files Failed to mount Some of the load files will
cannot be uploaded transaction cache dump be missing after the
because the compute node '/deepfs/postgres/ discovery export
could not be accessed to production_cache'. (61003) completes. These load files
obtain. will be reproduced on a
new run. If problem
persists across runs,
Contact Customer Support.
WARN 61004 Warns the user that one of Transaction Cache Dump Run the discovery export
the transaction cache failed with error - policy that saw the error
dump processes Validation failed during again. If the error persists,
encountered an error. In creation of load file. and you cannot find any
the case of a discovery (61004) cluster/data server
export run, this means that configuration issues,
the discovery export will contact Customer Support.
fail to produce one of the
load files.
Related reference:
Appendix C, Event log messages, on page 133
IBM may not offer the products, services, or features discussed in this document in
other countries. Consult your local IBM representative for information on the
products and services currently available in your area. Any reference to an IBM
product, program, or service is not intended to state or imply that only that IBM
product, program, or service may be used. Any functionally equivalent product,
program, or service that does not infringe any IBM intellectual property right may
be used instead. However, it is the user's responsibility to evaluate and verify the
operation of any non-IBM product, program, or service.
IBM may have patents or pending patent applications covering subject matter
described in this document. The furnishing of this document does not grant you
any license to these patents. You can send license inquiries, in writing, to:
The following paragraph does not apply to the United Kingdom or any other
country where such provisions are inconsistent with local law:
INTERNATIONAL BUSINESS MACHINES CORPORATION PROVIDES THIS
PUBLICATION "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER
EXPRESS OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
WARRANTIES OF NON-INFRINGEMENT, MERCHANTABILITY OR FITNESS
FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer of express or
implied warranties in certain transactions, therefore, this statement may not apply
to you.
Any references in this information to non-IBM Web sites are provided for
convenience only and do not in any manner serve as an endorsement of those Web
sites. The materials at those Web sites are not part of the materials for this IBM
product and use of those Web sites is at your own risk.
Licensees of this program who wish to have information about it for the purpose
of enabling: (i) the exchange of information between independently created
programs and other programs (including this one) and (ii) the mutual use of the
information which has been exchanged, should contact:
The licensed program described in this document and all licensed material
available for it are provided by IBM under terms of the IBM Customer Agreement,
IBM International Program License Agreement or any equivalent agreement
between us.
This information contains examples of data and reports used in daily business
operations. To illustrate them as completely as possible, the examples include the
names of individuals, companies, brands, and products. All of these names are
fictitious and any similarity to the names and addresses used by an actual business
enterprise is entirely coincidental.
COPYRIGHT LICENSE:
Each copy or any portion of these sample programs or any derivative work, must
include a copyright notice as follows:
If you are viewing this information softcopy, the photographs and color
illustrations may not appear.
Trademark
IBM, the IBM logo, and ibm.com are trademarks or registered trademarks of
International Business Machines Corp., registered in many jurisdictions worldwide.
Other product and service names might be trademarks of IBM or other companies.
A current list of IBM trademarks is available on the Web at "Copyright and
trademark information" at http://www.ibm.com/legal/copytrade.shtml.
Adobe, the Adobe logo, PostScript, and the PostScript logo are either registered
trademarks or trademarks of Adobe Systems Incorporated in the United States,
and/or other countries.
Java and all Java-based trademarks and logos are trademarks or registered
trademarks of Oracle and/or its affiliates.
UNIX is a registered trademark of The Open Group in the United States and other
countries.
Other company, product, and service names may be trademarks or service marks
of others.
This Software Offering does not use cookies or other technologies to collect
personally identifiable information.
If the configurations deployed for this Software Offering provide you as customer
the ability to collect personally identifiable information from end users via cookies
and other technologies, you should seek your own legal advice about any laws
applicable to such data collection, including any requirements for notice and
consent.
For more information about the use of various technologies, including cookies, for
these purposes, See IBMs Privacy Policy at http://www.ibm.com/privacy and
IBMs Online Privacy Statement at http://www.ibm.com/privacy/details the
Notices 163
section entitled Cookies, Web Beacons and Other Technologies and the IBM
Software Products and Software-as-a-Service Privacy Statement at
http://www.ibm.com/software/info/product-privacy.
O
OCR image processing 26 U
OCR processing 27 user
Optical Character Recognition 27 creating new 23
editing 24
user account
P deleting 24
policy audit locking 24
viewing by discovery export 97 unlocking 24
viewing by name 97
viewing by time 97
viewing by volume 97 V
viewing details 96 volume
policy audit failure messages 100 adding primary 46
policy audit success messages 99 discovery export 60
policy audit warning messages 99 system 61
policy audits 94 volume cache
predefined job deleting 79
running 77 volume data
processing exporting 62
monitoring 77 exporting to a system volume 62
importing 62
importing to a system volume 63
R volume definitions
recovery agent 32 editing 55
remote DCOM volume import audit
enabling 42 viewing 93
retention volume volume indexing 33
adding 56 volume-import audits 92
Enterprise Vault 59 volume, deleting 65
retention volumes 55 volumes
Enterprise Vault 42 policy limitations by type 65
retention 55
S
search depth W
volume indexing 33 warning messages
server platforms policy audit 99
supported by volume type 129 Windows authentication
services enabling integration on Exchange
restarting 15 servers 35
SharePoint Windows Share
alternate-access mappings 36 server 34
Privileges 36
privileges for social data 36
Secure Connection 35
servers 35
SharePoint objects
supported types 126
SharePoint volume
performance considerations with
versioning 54
SC27-5692-00