This post is part of an entire series
- Hana Smart Data Integration – Adapters
- Hana Smart Data Integration – Batch Dataflows
- Hana Smart Data Integration – Realtime Table Replication
- Hana Smart Data Integration – Realtime Sources with Transformations
- Hana Smart Data Integration – Realtime Sources with History Preserving
- Hana Smart Data Integration – Architecture
- Hana Smart Data Integration – Fun with Transformation Services
One of the more basic use cases is to copy data from a source database into Hana. Either just once to try out Hana or doing an initial load and setup realtime replication for all changes in order to have Hana side by side with the source system, e.g. for reporting.
Using the Smart Data Integration technology that can be done, you have to create virtual tables, write the SQL or dataflow for the initial load – one per source table – and create the realtime subscriptions. Or in short, use the Hana WebIDE to create a hdbreptask object that does all of that for you.
But let’s do something different instead of replicating a source database into Hana this time. Just for fun, a database would work as well of course. We want to replicate the Twitter!
For that we need an Adapter to connect to Twitter – no problem, shipped together with the Agent – and create a remote source with the Twitter login credentials. Everything else is straight forward:
Open the hdbreptask editor in the WebIDE by creating a new file of that ending
Select the remote source, the target schema and all remote tables – well in Twitter there are just two but in reality you can add hundreds of tables at once
Optionally specify filters
Execute the task
Actually, there is a nice video of all of this