You are on page 1of 4

EMC Data Archive POC Notes Architecture: The initial architecture was to use Rainfinity to write data to a Celerra

and then Kazeon would read the target Celerra for search capability.

Legal Data Drag and drop

Kazeon

Production SM Source Celerra

EMC SM POC Target Celerra

Archive Data Drag and drop

Rainfinity

During the POC the architecture changed due to the limitation of Kazeon and Rainfinitys ability to archive data from a non-shared system. So the new architecture was to use the Celerra to present a public share which uses could access to drag and drop the files for archive to. After the files has been moved to the Celerra share Kazeon would then index the data for search capability and lastly Rainfinity would archive the data to another filesystem for long term retention.

Legal Data

Rainfinity

EMC SM POC Target Celerra

Archive Data

Kazeon

Installation: Kazeon The installation of Kazeon was very simple and quick after the system was racked we were active for testing within a few hours. Celerra The NAS was setup and ready by the end of the day for configuration. There were no major problems with this setup. Rainfinity There was not SE that was available for the installation. Once we did get someone to install the RF appliance we still needed another person to configure the appliance. The appliance took 1.5 days to get configured this seems a lot more complex then I would have figured. The one thing to keep in mind though is that this was a one-time setup, after the setup the policy to archive the data was easy to configure. Also to note was that the architecture was modified during the setup of this process of how we would archive. POC Test: Objective testing This test is simply a test of the objectives outlined in the RFI. Objective 1 - We need a solution that must provide the means to automate the archiving of structured and unstructured data to a centralized storage location based on IT corporate policies. Objective Achieved and Tested We could move unstructured data to a NAS share where it was then archived for long term retention. I was unable to test structured data due to lack of time and equipment availability. Objective 2 - We need a solution that must provide the ability to collect and track all incremental changes of structured and unstructured active data. The collected data must then be stored on a centralized secondary storage location while preserving the file date information and preventing the destruction of data for means of long term legal retention. Objective Achieved and Tested Using Kazeon we were able to scan file shares for data requested for long term retention. After which a copy of the found data could be copied and retention applied for long term retention on an incremental basis. Objective 3 - Provide the capability to search through collected data as identified in the 2nd bullet with a future e-discovery process. Objective Achieved and Tested Using Kazeon to index data we were able to search data at both metadata level and a document contents level searching. Objective 4 - Capability to assign ownership of data when archived.

Metrics: Copy of data from PC to archive file system - 21GB 2 hours Kazeon basic scan of files on archive file system 21GB 2 min Kazeon deep scan of files on archive file system 21GB Rainfinity archive to legal file system 21GB 30 min

RFI Testing Criteria: 1. Can your solution de-duplicate against all the data that it stores? The Celerra filesystem can de-duplicate. This a post process 2. Does your solution support Oracle, MS SQL and/or Unix (Redhat Linux and HP-UX)? The data written can be either NFS or CIFS data so as long as the data to be archived can write the data to this location it can be recovered at a later date. But it should be noted that this whole process would be manual it is unable to do this via tools or automation. 3. Can the network be throttled during ingestion of archive data? No, this is not done with any tool set with this EMC solution. You would have to utilize Qos rules to make any sort of bandwidth limitations. 4. Can replication solution support a remote site greater than 500 miles away? Yes the Celerra can replicate to another Celerra. 5. Can the data in the repository be searched? Yes, Kazeon can index both source and destination data. There is a basic index which pulls only meta data and there is also a deep index which will search inside documents. The searching with Kazeon if very good and provides a lot of information very easily. 6. Are servers required for your solution? No, the equipment needed at the production site are 1 Celerra (NAS), Kazeon appliance and Rainfinity appliance, all of which are self contained in a rack. 7. Can your solution achieve all the project objectives? Yes all project objectives can be met. 8. Does your solution require an agent? No agents are required data is accessed by being shared to the network. 9. Can you set quotas on storage? Yes you can set quotas on the Celerra for storage requirements. 10. Can identified data for recovery be written to removable medium? Yes identified data can be copied down to the desktop or exported into a legal extraction csv file.
Support Matrix Description

EMC x

solution that must provide the means to automate the archiving of structured and unstructured data to a centralized storage location based on IT corporate policies.
solution that must provide the ability to collect and track all incremental changes of structured and unstructured active data. The collected data must then be stored on a centralized secondary storage location while preserving the file date information and preventing the destruction of data for means of long term legal retention.

Provide the capability to search through collected data as identified in the 2nd bullet with a future e-discovery process Capability to assign ownership of data when archived Can your solution de-duplicate against all the data that it stores?

x x x

Does your solution support Oracle, MS SQL and/or Unix (Redhat Linux and HP-UX)? Can the network be throttled during ingestion of archive data? Can replication solution support a remote site greater than 500 miles away? Can the data in the repository be searched? Are servers required for your solution? Can your solution achieve all the project objectives? Does your solution require an agent? Can you set quotas on storage? Can identified data for recovery be written to removable medium?

x x x x x

Thoughts & Perceptions:

You might also like