Forum Discussion

Siva_Venna's avatar
Siva_Venna
Contributor
5 years ago

Read historical data in the first run

I have a pipeline that extracts data from Oracle table and write data to destination. Oracle select snap will read current day’s date based on a time field. When I am migrating this pipeline to PROD environment in the first run it should read entire data(historical data) and from second run it should read current day data. how to design this ?Thanks

5 Replies

  • Supratim's avatar
    Supratim
    Contributor III

    @Siva_Venna You should have one file with date filed in SLDB. your pipeline should read the file and pass the date value to oracle snap and do the comparison like- date_field > oracle_date_field . After completion of your pipeline you should calculate the max value of oracle_date_field and save to the same file in SLDB.It will help you to extract the delta record from Oracle any time you want.

    • Siva_Venna's avatar
      Siva_Venna
      Contributor

      thank you @Supratim, as you have suggested that’s doable in DEV. I am concerned how do I achieve this in PROD environment where I won’t be able to modify the file or snap settings . I am looking for some configuration/design that doesn’t need manual intervention.

      • Supratim's avatar
        Supratim
        Contributor III

        You don’t need to change the manually for every single execution.
        When you migrate to prod first time also move this file with some old dates , rest of the time pipeline will update the latest time stamp on this file.

    • Supratim's avatar
      Supratim
      Contributor III

      [full_incremental_load_2020_09_25.slp|attachment]
      Try this way

      (upload://mpHeJ7IYL1IB9KyKwa2VlgD2FtY.slp) (19.7 KB)