Interacting with billions of rows of data in seconds from a single source is exciting, but big data exploration gets really interesting when you work with data across multiple sources.
All of these scenarios require an agile and flexible approach to multisource analytics to get the best insights.
Data is everywhere, and it’s your job to make sense of it -- wherever it resides. Zoomdata offers practical approaches to multi-source analysis. They are multisource dashboards, Zoomdata Fusion, cross-source filtering, and keysets.
Few traditional BI vendors offer multi-pass and cross-source set analysis, and those that do require time-consuming involvement from DBAs or IT.
With Zoomdata, users enjoy fast, on-the-fly analysis across a wide variety of data, no matter where it resides.
Zoomdata Fusion makes multiple data sources appear as a single source to the business user. Fusion can present data from relational and non-relational sources, including structured and unstructured sources. (No audio.)
Without requiring any semantic modeling or advanced setup, users can rapidly explore and visualize the impact of applying common filter criteria across all elements on a dashboard, regardless of where the data comes from.
Keysets are ideal for multi-pass use cases, and for rapid data exploration and set analysis across a wide variety of data platforms. Keyset Analytics is especially powerful when related data sources are too big to join into a single data source, are streaming live, or are too sensitive to move into curated systems because of security concerns.
As markets change and technology evolves, so does our data. Our data accumulates faster and grows ever bigger. It will never slow down. We’ll never generate less data.
Putting this data to work requires a variety of data platforms.
Business requirements dictate that some data is so big that it needs to be managed in infinitely scalable systems like Hadoop. Near-real-time analysis is very doable for organizations that land data in modern fast data sinks. Some data isn’t needed for real-time analysis, but needs to be “hot” and return results super quickly. Search engine databases are flexible and allow qualitative analysis through free text search. Some data is stored on-prem, some is in the cloud, and some straddles both through a hybrid-cloud.
And, of course, all these data platforms come with their own tradeoffs, including cost. For all the advances we enjoy, data will persist in a wide variety of systems for a long time to come.
DBAs and IT professionals, want to learn more? Learn about pushdown processing
Be wary of traditional BI vendors that encourage you to combine and manage all your data into their proprietary “single stack” or “full stack” environment. Look for words like import, ingest, extract, and so on. What might sound like data centralization is far too often data duplication and “black box” data transformations. These are avoidable data governance and lineage problems.
For modern data sources, the problem is even worse.
Above all -- copying data out of one data source into another is inherently risky, and not always in the best interest of the business analyst.
Maintaining appropriate security privileges across all data extracts is difficult. Removing context such as columns and rows may be the right thing to do for some use cases, but it’s generally better to keep everything in place and hide or filter unauthorized data based on user privileges.
Then there’s the problem of data transformation. When the same data is transformed into multiple secondary databases, which is the source of the truth?
These problems are real-life and non-trivial. There are even products you can buy to track and manage this mess of data duplication. But instead of buying yet another IT product to manage copies of data all over the enterprise, we at Zoomdata encourage you to:
Interacting with billions of rows of data in seconds from a single source is exciting, but big data exploration gets really interesting when you work with data across multiple sources.