Working with the Embedded Data Lake in SAP Data Warehouse Cloud
In this blog post you will learn how to realize a Cold-To-Hot data tiering scenario in SAP Data Warehouse Cloud and SAP HANA Cloud data lake. In this scenario, data will be loaded from a S3 Bucket in AWS into the SAP HANA Cloud data lake using a Data Flow in SAP Data Warehouse Cloud. Part of this data will then be snapshotted and stored as hot in-memory data. The Hot-To-Cold scenario will be covered in a different blog.
The figure below shows the overall approach for this scenario. First, we will create a new table in the SAP HANA Cloud data lake and a SAP HANA Cloud virtual table which is mapped to that table. The virtual table is used in SAP Data Warehouse Cloud to insert the S3 data into the data lake. Once the data is inserted four views are created one of which contains snapshotted in-memory data. These views will then be consumed by SAP Analytics Cloud to visualize the data and show performance differences in query times. In this example, space names such as “UK_SPACE” are used, however, feel free to set the space names according to your naming convention.
The following steps will guide you through the scenario to set up a Cold-To-Hot data tiering scenario in SAP Data Warehouse Cloud:
1. As a first step, the embedded Data Lake in flexible tenant configuration needs to be configured. Check out more in the following blog.
2. Next, a space needs to be selected that connects to the embedded Data Lake. This can be done in the Space Management in the Overview Tab. Here the checkbox “Use this space to access the data lake” needs to be activated. Note that only one space can be assigned to access the SAP HANA Cloud data lake.
3. A database user needs to be created so that you can access the underlying SAP HANA Cloud database and create the tables. This option can also be found in the Space Management. Create a new database user with read & write privileges and click on “Save”.
4. Once this is done you can use your preferred SQL tool to create tables in the data lake and access those tables via SAP HANA virtual tables in your open SQL schema. In this blog, the SAP HANA Database Explorer will be used. It can directly be opened from the Space Management via “Open Database Explorer”. In the explorer click on the corresponding instance and enter your credentials to connect to the database.
5. SAP Data Warehouse Cloud offers two stored procedures that you can use to easily create and access the tables. To create tables in the data lake you can use the stored procedure “DWC_GLOBAL”.”DATA_LAKE_EXECUTE”. Now create a table in the data lake and make sure that the columns’ data types match to the respective S3 data you are going to use.
CALL "DWC_GLOBAL"."DATA_LAKE_EXECUTE"('CREATE TABLE UK_SALES_ACQUISITION ( Location VARCHAR(40), Product VARCHAR(40), Time_ VARCHAR(8), Sales_Unit DECIMAL(9,1), Year VARCHAR(4) )');
6. Next you can create a virtual table in your open SQL Schema that refers to the table in the data lake. This can be done with the following procedure:
CALL "DWC_GLOBAL"."DATA_LAKE_CREATE_VIRTUAL_TABLE" ( VIRTUAL_TABLE_NAME => '0_UK_SALES_ACQUISITION', DATA_LAKE_TABLE_NAME => 'UK_SALES_ACQUISITION' );
7. Another important step is to grant the space in SAP Data Warehouse Cloud privileges to insert and update the virtual table. Otherwise, the data flow would not be able to insert data into the table. The following procedure will grant the space all privileges:
GRANT ALL PRIVILEGES on "AASPACE_W21_20220921#ONBOARDING"."0_UK_SALES_ACQUISITION" to AASPACE_W21_20220921 with grant option
8. Now that the virtual table is created, you can go back to SAP Data Warehouse Cloud into your Space with data lake access and create a data flow. To choose the S3 Bucket go to “Sources” -> “Connections” and to your S3 connection. Choose your source file from the S3 bucket and drop it into the canvas of the data flow.
9. The virtual table previously created in the open SQl schema is also available in the sources panel. Drag and drop it onto the canvas of the data flow and click on import and deploy to make it useable as a native SAP Data Warehouse Cloud artefact.
Also make sure that it is connected to the source S3 data. A projection node is inserted into the data flow to filter and remove columns that are not needed. In this example, Location, Time, Product and the KPI sales unit are retained. Also, a calculated column Year is created to make it easier to filter on different time periods.
Make sure that the table is set as the target table and all columns are mapped. In the end, the data flow should look like this:
10. You can now save and deploy the data flow.
11. Once finished the data flow can be started. Run the data flow and review the status in the Data Integration Monitor. The Data Integration Monitor provides a good overview of all data flow runs. Here you can find information such as last run status, the frequency, duration, start and end timestamps and if set next scheduled runs. In the Record Count you can track how much data has already been transferred to the target.
12. After the status of the data flow changed to completed and all the data is transferred into the data lake, share the table into UK_SPACE.
13. The next step is to switch to the UK_SPACE and create a new view based on the shared table. The shared table can be found under Shared Objects. As a first view we want to create a snapshotted view for the Year 2021 which will contain about 50 million rows of data. For this purpose, a filter Year = ‘2021’ is inserted on the Year column. In the properties set ‘Analytical Dataset’ as a semantic usage, toggle the switch ‘Expose for Consumption’ and set Sales Unit as a measure. Save and deploy the View and name it 2_UK_SALES_2021. When the view is deployed scroll down in the properties panel to the persistency area and create a snapshot. The details of the snapshot you can review in the Data Integration Monitor (View Persistency Monitor).
14. Create View for 2020 Data (non-persisted)
The second view we want to create is a non-persisted view for year 2020 which will contain approx. 53 million rows. To do this, follow the same steps as before, except that no snapshot is created. The name of the view is set to 2_UK_SALES_2020.
15. Create View with Union of 2021 snapshotted and 2020 virtual data
Thirdly, create a view to combine the 2021 snapshotted data and the 2020 virtual data, which we created in steps 13 & 14. The union can be created by dragging the second view on top of the first view and choosing the union option. Set the same properties as in the previously created views. A snapshot is not created. The name of the view is set to 3_UK_SALES_2020_2021.
16. Create View to combine 2020 and 2021 virtual data
We also want to create a view to get 2020 and 2021 data from the data lake directly and compare the query times to the 3_UK_SALES_2020_2021 view. Create a view and filter on the year 2020 and 2021.
17. Consumption in SAP Analytics Cloud
Next a story in SAP Analytics Cloud is created which loads the data from SAP Data Warehouse Cloud via a live connection and displays it on the canvas. If you have not yet created a live connection please refer to this document.
Create a new Canvas Page and click on Add data. Now choose SAP Data Warehouse Cloud as the data source. Select the connection and choose the respective UK_SPACE. First, select the 2020 Sales data. Add a chart to display the 2020 data.
Next add the 2021 Sales data similar to the previous step and set the charts data source to 2021. Select the measure Sales_Unit again so that the chart displays data. Notice that the data is loaded much faster than before because this view has been persisted in SAP Data Warehouse Cloud and resides in memory unlike 2020 data, which is stored in the data lake.
You can also test the query times for the union view which combines data from the data lake with snapshotted data. Compared to the view were all data resides in the data lake this is about 1.5 times faster. The figure below displays how the data flows through SAP Data Warehouse Cloud to SAP Analytics Cloud and shows the query time differences.
To sum up, this blog has shown how to carry out a Cold-To-Hot data tiering scenario in SAP Data Warehouse Cloud and the embedded HANA Cloud data lake. It was shown that to insert data into the data lake a data lake table and a SAP HANA virtual table in the open SQL schema needs to be created which can then be used in the data flow. After that several views were built on top, one of which was snapshotted and therefore stored as ‘hot’ data. Consuming the views in SAP Analytics Cloud made clear that query times for persisted data were much faster compared to data which resided in the data lake. But of course, it can also be beneficial to store the data in the data lake, e.g. depending on the frequency of data access, volume and how well this data is structured. Especially in times of ever-growing data footprint this provides a great opportunity to better utilize existing resources and optimize your total cost of ownership.
Feel free to also check out this blog which provides valuable insights into SAP Data Warehouse Cloud and the SAP HANA data lake. Also thanks to Nicolas Gabelmann, Philipp Plazibat, Amogh Kulkarni and Oliver Huth for their contribution to this blog.
Thanks for sharing blog, it very clear with detailed steps. Just one query on performance of data load of approx 103 million rows and 5 column into Data lake virtual table in your scenario from S3 as source? From snapshot of Data flow monitor as it's 28 sec (In progress state). Also it will it be great if you tell what size Memory, cpu your DWC tenant?