Validate the integrity of backup data, detect corruption due to ransomware and support streamlined recovery to minimize business interruption.

Learn More >>

About Our Technology
Unmatched enterprise integration and high-speed indexing
Enterprise Integration

The Index Engines platform seamlessly integrates into existing data infrastructures to perform fast, scalable indexing. Each engine understands common network and storage protocols. The diagram below shows an example implementation of Index Engines for comprehensive data discovery.


Unified integration and data collection from the following environments is currently supported: Legacy Backup Tapes, Virtual Tape Libraries, NAS, Desktops, Network Servers, Hard Drives, USB Drives, Forensic Images.

As Index Engines processes data, a searchable index is generated containing rich metadata and content for all common unstructured files and email.  This unified index contains information about data residing on backup tapes, networks and large storage file systems to allow for search and extraction of unique content through a single interface.

The data is not copied during the indexing process, therefore the index footprint is remarkably small - only 4-8% of the original data size. No other solution has the depth and breadth of discovery across the enterprise. One Index Engine can index over a billion objects per system.
Unmatched Speed
The Index Engines platform indexes unstructured and email data unprecedented speeds and with extreme scalability. Architected to keep up with backup speeds and high speed networks, Index Engines makes discovery of hundreds of TBs of enterprise data achievable. The intelligent engine can process multiple indexing streams in parallel, at speeds of 1TB/hour using a single node.   

The Index Engines platform is designed with maximum flexibility allowing the platform to index content from any source including LAN, Tape, VTL, and D2D systems.  In addition to the complex unstructured and semi-structured formats like Exchange and Lotus Notes, the Index Engines platform understands and can directly process NDMP input and most backup formats. How Index Engines achieves these capabilities is explained below.

Enterprise Class Indexing

All components of the platform, including the database, word scraping, and query engine have been developed by Index Engines in order to meet Enterprise speed and scalability requirements.  Traditional indexing products that use third party and open source components, typically designed for Internet-class indexing, fail when faced with terabytes of data using a single node. 

The success of the Index Engines architecture has been proved based on the efficient processing of enterprise data.  Full content and metadata indexing is performed at 1TB/hour/node.  The resulting index footprint is only 4 to 8% of the original data. 

Sequential Processing

As data streams to Index Engines it is scanned and processed sequentially in order to maintain high speed throughput.  Sequential processing of streaming data is a core component of Index Engines intellectual property and thus a unique advantage of the platform.  Traditional indexing requires random disk access in order to process data.  Random access to disk is slow because of unavoidable hardware limitations.

Without sequential processing of streaming data enterprise-class indexing cannot occur within reasonable timeframe and with a realistic number of processing nodes.  Traditional vendors will claim to meet these requirements; however they will be forced to utilize dozens of processing nodes just to keep up.  1TB/hour per node is not a speed metric that random disk access can achieve.

Leveraging Purpose Built Protocols

Data can only be processed as quickly as it can be scanned.  Index Engines has taken advantage of higher speed protocols for access to enterprise data.  NDMP was developed as a high speed protocol to efficiently dump bulk data from a NAS device over the network.  NDMP is significantly faster than CIFS or NFS for accessing bulk data over the network. Using NDMP, data can stream at speeds of up to 100 MB/second per GigE link.  Multiple streams can be configured to process 1TB/hour.  Traditional network protocols, such as NFS and CIFS, are supported by Index Engines but more simultaneous streams are required for fast processing speeds.  Index Engines platform supports sustained indexing using NFS/CIFS of 800GB/hour/ node, 80% of the NDMP indexing performance.

Index Engines knowledge of proprietary backup formats and the proprietary sequential processing technology make it possible to directly process Tapes, VTLs, and D2D subsystems directly at these high speeds.  With the knowledge of backup formats these subsystems can be read and managed directly to create detailed knowledge of backup content.

Evaluating Index Engines

The Index Engines platform is focused on providing comprehensive and efficient indexing of data across the enterprise. When choosing an enterprise indexing strategy, you will need to evaluate the following criteria:


Index Engines

Collecting Data

How to find what to index

Just as Internet search engines crawl the web to find data, these solutions crawl desktops, servers and archives, making for a slow and resource intensive process.

Our engine does not burden the infrastructure with new processes or agents; it indexes data where it resides or is stored.

Indexing Speed

Large volumes require fast indexing

Traditional indexing solutions are bogged down by slow network protocols during crawling and the speed of random disk access during indexing. 

Our proprietary indexing technology indexes data at up to 1 TB/hr.

Learn more about our speed here.


Is sensitive data protected?

Some indexing solutions are not built for the enterprise; they generate cache copies of your data and don't integrate into existing permission schemas. 

Our product is designed from the ground up for the enterprise - no copies of data are created and we inherit your existing security environment.

Index Footprint

How much storage will I need?

Typical indexing products generate bloated indexes often two to three times the size of the original data, creating a storage nightmare.

Our optimized index is only 4-8% of the original data size, a more manageable size which also results in faster queries.


How much stuff can I index?

Due to slow indexing speeds, and large storage requirements, these solutions are often relegated to small projects.

Our index economically processes billions, not millions, of files supporting enterprise-wide engagements.


What does it take to get going?

Other indexers require that data be configured in a specific manner so the indexer can understand it, a time consuming process.

Our solution is designed to be plug and play, no changes to your data environment are necessary.


What's it going to cost?

Due to the inefficiencies mentioned above indexing data enterprise-wide is cost prohibitive.

Index Engines provides the most economical and scalable solution on the market today.

Detect527 Detect

Analytics and machine learning detect corruption with 99.5% confidence.

diagnose527 Diagnose

Forensic reports identify the impact of the attack and the vector utilized.

recover527 Recover

Detailed listing of backup sets with most recent pre-attack version of files.

 Power Over Information™

Index Engines has been providing organizations with power over information since 2004. Our scalable, high performance indexing engine has been uniquely architected to support both primary and backup data and is unmatched in the industry.

CyberSense® is the latest offering that combines analytics and machine learning to detect corruption due to a cyberattack with 99.5% accuracy. Beyond CyberSense, Index Engines delivers an extensive information management product family that supports management and governance of content with petabyte-class scalability and performance.

Contact Us            Our Technology

Manage thin small

Management Edition

Understand what data exists on your network and manage it more effectively. The Management Edition gives you an affordable, quick and intuitive way to execute your data policies.

Analysis sma

Analysis Edition

The Analysis Edition combines the indexing, reporting and tiering of the Management Edition and adds advanced capabilities including pattern and sensitive data search.

Governance small

Governance Edition

The Governance Edition provides comprehensive indexing, search and management of sensitive and personal data across primary storage and backup data to support critical initiatives.