Dear Forum Members,
We are planning to use cloverDX graphs to ingest files approx. 40-45 files on a daily basis, each files may contains 1000-30000 rows. Can CloverDX handles this? Is there any documentation where you can direct me to?
Kind regards,
Hi,
This is a tricky question because there are many variables that influence the duration of processing the data.
The computing power mostly depends on the used Hardware, but in the mentioned use case, if there is a reasonable number of columns, the 45 * 30 000 records are about 1 350 000 records, this can be processed in seconds.
Other Variables are the number of jobs, graphs, components in them, and time distribution. But generally, hundreds of thousands of records can be processed in minutes even in fairly complicated graphs.
A very common bottleneck is caused by a lack of system memory, but we are talking about millions and millions of records in very complicated processes.
So, based on your specifications of the example your use case should be a piece of cake for CloverDX on any hardware.
Keep in mind that we always recommend trying everything in a test environment first.
Best regards, Tom.