[英]Sqoop incremental load using Informatica BDM
I am new to Informatica BDM.I have a use case in which I have to import the data incrementally (100 tables) from RDBMS into Hive on daily basis. 我是Informatica BDM的新手,我有一个用例,每天必须将RDBMS的数据(100个表)以增量方式导入到Hive中。 Can someone please guide me with the best possible approach to achieve this? 有人可以指导我采用最好的方法来实现这一目标吗?
Thanks, Sumit 谢谢Sumit
Hadoop is write onces read many (WORM) approach and the incremental load is not easy stuff. Hadoop是一次写入多次读取(WORM)的方法,而增量负载并非易事。 There are following guideline you can follow and validate your current requirement 您可以遵循以下准则,并验证您当前的要求
If the delete operation is not coming as change, then you also need to think how to act on it and in such case, you need to get the full refresh. 如果删除操作不是更改,那么您还需要考虑如何操作,在这种情况下,您需要进行完全刷新。
声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.