Re: Changing large in-memory data to file based operation
Yes if you store to file and set the body as java.io.File or
java.io.FileInputStream in the bean then the the sftp producer should
use that as-is and upload the file in a streaming fashion.
On Sat, Mar 10, 2018 at 6:47 AM, Goyal, Arpit <arpit.goyal@xxxxxxx> wrote:
> Hello Camel Experts,
> Currently we have something of below route (as simplified version), where we keep reading records 1000 at a time and keep formatting and holding in memory.
> Current Simplified Solution:
> Generally the data is small and we kept in-memory, we want to improve the same now to safe-guard any in-memory problem. In our landscape we can use 'tmp' storage to hold the file. So we plan to change it like this
> Possible Modified Solution:
> When locally tried the solution, we see it works - Camel understand 'java.io.File' object & writes to SFTP.
> Question 1 - Would camel load it in memory completely before writing to SFTP?
> Question 2 - If yes, how can we solve this problem - does some way of streaming works?
> Checked : http://www.catify.com/2012/07/09/parsing-large-files-with-apache-camel/ already but they are holding in memory.
> Limitation: We don't want to write in chunks to sftp because of the complexity of some data have to written out-of-order.
Camel in Action 2: https://www.manning.com/ibsen2