Datadobi Announces DobiMigrate 5.8 – Introduces Chain of Custody

Datadobi recently announced version 5.8 of its DobiMigrate software and introduced a “Chain of Custody” feature. I had the opportunity to speak to Carl D’Halluin and Michael Jack about the announcement and thought I’d share some thoughts on it here.

 

Don’t They Do File Migration?

If you’re unfamiliar with Datadobi, it’s a company that specialises in NAS migration software. It tends to get used a lot by the major NAS vendors as rock solid method of moving data of a competitor’s box and onto theirs. Datadobi has been around for quite a while, and a lot of the founders have heritage with EMC Centera.

Chain of Custody?

So what exactly does the Chain of Custody feature offer?

  • Tracking files and objects throughout an entire migration
  • Full photo-finish of source and destination system at cutover time
  • Forensic input which can serve as future evidence of tampering
  • Available for all migrations.
    • No performance hit.
    • No enlarged maintenance window.

[image courtesy of Datadobi]

Why Is This Important?

Organisations are subject to a variety of legislative requirements the word over to ensure that the data presented as evidence in courts of law hasn’t been tampered with. Some of them spend an inordinate amount of money ensuring that the document management systems (and the hardware those systems reside on) offer all kinds of compliance and governance features that ensure that you can reliably get up in front of a judge and say that nothing has been messed with. Or you can reliably say that it has been messed with. Either way though, it’s reliable. Unfortunately, nothing lasts forever (not even those Centera cubes we put in years ago).

So what do you do when you have to migrate your data from one platform to another? If you’ve just used rsync or robocopy to get the data from one share to another, how can you reliably prove that you’ve done so, without corrupting or otherwise tampering with the data? Logs are just files, after all, so what’s to stop someone “losing” some data. along the way?

It turns out that a lot of folks in the legal profession have been aware that this was a problem for a while, but they’ve looked the other way. I am no lawyer, but as it was explained to me, if you introduce some doubt into the reliability of the migration process, it’s easy enough for the other side to counter that your stuff may not have been so reliable either, and the whole thing becomes something of a shambles. Of course, there’s likely a more coherent way to explain this, but this is tech blog and I’m being lazy.

 

Thoughts

I’ve done all kinds of data migrations over the years. I think I’ve been fortunate that I’ve never specifically had to deal with a system that was being relied on seriously for legislative reasons, because I’m sure that some of those migrations were done more by the seat of my pants than anything else. Usually the last thing on the organisation’s mind (?) was whether the migration activity was compliant or not. Instead, the focus of the project manager was normally to get the data from the old box to the new box as quickly as possible and with as little drama / downtime as possible.

If you’re working on this stuff in a large financial institution though, you’ll likely have a different focus. And I’m sure the last thing your corporate counsel want to hear is that you’ve been playing a little fast and loose with data over the years. I anticipate this announcement will be greeted with some happiness by people who’ve been saddled with these kinds of daunting tasks in the past. As we move to a more and more digital world, we need to carry some of the concepts from the physical world across. It strikes me that Datadobi has every reason to be excited about this announcement. You can read the press release here.

 

EMC – Configuring NTP for the Centera

We recently installed Governance Edition on our Centera clusters, and needed to get time in sync between the clusters. I have NFI why ntp wasn’t configured previously, but here’re the steps. Special thanks to Mat for bothering to take screenshots of everything while it was happening. I know it’s a long article, and I don’t want to give too much away, but everyone gets the right time in the end.

Config# show cluster time
Cluster: Mon Aug 12 23:49:12 2013 UTC (External NTP disabled)
Config# set ntp configuration
External NTP Enabled? (yes, no) [no]: yes
Network association (access, management) [access]: access
IP address of NTP server: 1.1.1.1
Issue the command? (yes, no) [no]: yes
Command failed: External NTP is not enabled. The cluster is more than 30s off with the NTP server. Please contact EMC Customer Support for assistance in enabling external NTP.
Config# set cluster time
You cannot set the cluster time on a cluster is not in Basic mode. Please contact EMC Customer Support.

 

At this point you’ll need to contact EMC Support, as you won’t be able to run the commands yourself. You can either opt to drift the time in (this is very slow) or you can have an outage on the cluster and force the time to be set (we opted to drift the time in).

Config# set cluster time
Current cluster time is: 13 Aug 2013 02:32:53 UTC
New UTC time [DD-MMM-YYYY HH:MM:SS or IP address of NTP server]: 1.1.1.1
New cluster time will be set to 13 Aug 2013 02:41:12 UTC
Stop cluster services to change time? (yes, no) [no]: no
Cluster time will be adjusted forward with approximately 8 minutes 19 seconds.
The adjustment will take approximately 11 days 14 hours 28 minutes to complete.
Issue the command? (yes, no) [no]: yes
Cluster started to adjust cluster time with 8 minutes 19 seconds.

 

You can check the progress of the time sync by using the “show cluster time” command,  it will give you an approximate amount of time remaining till the adjustment process has completed

Config# show cluster time
Warning: Cluster time adjustment is in progress. 8 minutes 18 seconds left to adjust forward
Cluster: Tue Aug 13 02:44:06 2013 UTC (External NTP disabled)
Config# show cluster time
Warning: Cluster time adjustment is in progress. 1 minute 17 seconds left to adjust forward
Cluster: Thu Aug 22 21:10:50 2013 UTC (External NTP disabled)
Config# show cluster time
Cluster: Sun Aug 25 22:47:58 2013 UTC (External NTP disabled)

Once the time has been adjusted you can then set up NTP

Config# set ntp configuration
External NTP Enabled? (yes, no) [no]: yes
Network association (access, management) [access]: access
IP address of NTP server: 1.1.1.1
Issue the command? (yes, no) [no]: yes
External NTP is successfully enabled.

If you have more than one NTP server you can add it to the list of server to sync against

Config# set ntp addserver
IP address of NTP server: 2.2.2.2
Issue the command? (yes, no) [no]: yes
The specified NTP server is successfully added.

Show the current NTP configuration and status

Config# show ntp configuration
Cluster NTP configuration:
External NTP is enabled
Network association: access
NTP servers:
# IP
---------------
1 1.1.1.1
2 2.2.2.2
---------------
Config# show ntp status all
Node name: centera1
Node roles: access,replication,management,storage,blc
# IP           Status Offset
-----------------------------------
1 1.1.1.1 good   0s
2 2.2.2.2 good   0s
-----------------------------------
Node name: centera2
Node roles: access,replication,management,storage,blc
# IP           Status Offset
-----------------------------------
1 1.1.1.1 good   0s
2 2.2.2.2 good   0s
-----------------------------------

Now when you issue the “show cluster time” command it states that it is synced via NTP.

Config# show cluster time
Cluster: Mon Aug 26 05:00:22 2013 UTC (Synchronized with external NTP)

EMC – Maximum Pool LUN Size

Mat has been trying to create a 42TB LUN to use temporarily for Centera backups. I don’t want to go into why we’re doing Centera backups, but let’s just say we need the space. He created a Storage Pool on one of the CX4-960s, using 28 2TB spindles and 6+1 private RAID Groups. However, when he tried to bind the LUN, he got the following error.

err1

Weird. So what if we set the size to 44000GB?

err2

No, that doesn’t work either. Turns out, I should really read some of the stuff that I post here, like my article entitled “EMC CLARiiON VNX7500 Configuration guidelines – Part 1“, where I mention that the maximum size of a Pool LUN is 16TB. I was wrong in any case, as it looks more like it’s 14TB. Seems like we’ll be using RAID Groups and MetaLUNs to get over the line on this one.

Update articles page

I’ve added a brief article covering the steps involved in installing the EMC Centera Tools software suite – used for management and maintenance of Centera clusters. I’m hopeful that I’ll have time to do some more articles covering basic usage of the Centera CLI in the near future. Feel free to look at some of the other articles I’ve published as well – hopefully you’ll find someting useful in there.