Datadobi opens vast file system window to clients – Blocks and Files
Datadobi has opened up its petabyte-scale file mapping technology to customers with a Dobi query language.
Datadobi is a provider of data migration technology, with DobiMigrate moving NAS files and objects, and DobiProtect protecting and recovering them. Both products are powered by DataDobi’s data mobility engine and are designed to analyze large file systems (data lakes) containing billions of files. This produces a catalog containing huge lists of file paths and their metadata in a proprietary format.
Historically, these scan files were only used to perform data migration or data protection by Datadobi, but now customers can query them directly using Dobi Query Language (DQL) as part of a file rating service.
Carl D’Halluin, CTO of Datadobi, wrote in a blog: âThe volume of data is only expected to increase over the next few years. IT administrators need a data management solution that can turn data into digestible material to enable organized decisions about storage options for migration and protection.
Step forward DQL. It is a query framework that can search inside data lakes for:
- Identify cold datasets – data that is infrequently accessed;
- Identify old data sets – data that was created or modified some time ago;
- Identify data sets belonging to a specific user or group, for example by users who no longer work in the company;
- Identify which shares, exports, or directory trees are homogeneous (cold, old, owner, file types) and can be treated as a single set of data, for example, on which to take specific lifecycle actions.
Datadobi created the File System Assessment as a customer service offering last year to help plan for a data migration or reorganization. DQL allows you to customize these assessments.
DQL is a form of file system analysis. It opens the door to future possibilities such as an alerting software layer that can run DQL to know the state of the filesystem and the subset, and suggest actions, such as a file / subset. object moves to faster or slower / cheaper storage depending on its activity level. A step beyond that would be to put policies in place and then have an additional layer of software take automatic action if threshold values ââare reached.
For example, move all files or objects of a particular type to archive or cold storage if certain access rate requirements are met (for example, if the data has not been accessed within 60 days) . Datadobi has some interesting possibilities here.