I am excited to share about  our new mission published this week in the SAP Discovery Center,  “Access, Share and Monetize Data with SAP Data Warehouse Cloud”. SAP BTP Startup program “Data-to-Value Track”, an initiative from SAP.iO is supporting startups to become a data provider by leveraging the Data Marketplace for SAP Data Warehouse Cloud. And this mission will guide startups as we provide a detailed step-by-step process on how to replicate/federate data from Cloud storages. As we see more adoptions and understand the different data sources that will be used to ingest data into our SAP Data Warehouse Cloud Data Marketplace, we will update the mission accordingly.

The relevance of external data has been more significant with the enterprises especially with the rise of COVID-19 crisis. Be it the external COVID-19 metrics that will be relevant for Human Resources Team for understanding the impact of employee productivity, geo-spatial data for performing advanced analytics for Logistics/Real estates, external Industry specific data for identifying trends and forecasting, the data providers has been a catalyst for enterprises to plan proactively and take necessary actions in this dynamic environment.

Solution%20Architecture%20for%20Data%20replication%20from%20Cloud%20Storages

Solution Architecture for Data replication from Cloud Storages

Although this discovery mission focuses on startups for onboarding data providers to SAP Data Warehouse Cloud, we encourage cloud enthusiasts & technical consultants to sign up for the missions to understand the following

  1. Establishing connections to Cloud Storages[Amazon S3/Azure Blob Storage/ Google Cloud Storage] and create data flows to replicate the data using embedded data pipelines.
  2. Creating Database Access thereby providing you access to open sql schema and replicate  the data from cloud storages using python scripts. You can enable to read-write access to the  open sql schema and use the python scripts provided as part of the discovery mission.
  3. Create Data Profile, data products as a data provider.
  4. Publishing data with necessary licenses in Data marketplace.
  5. Validating the data as a data consumer from Data Marketplace.
  6. Finally loading the external data into your space and blending it with your internal data & visualize using SAP Analytics Cloud.

Additionally we will be providing python scripts focusing on parallel inserts into open sql schemas and will updating the mission in the next few weeks. You can also refer the GitHub repository shared as part of the mission.

Based upon the feedback and consumption, we will add additional data source connections for federation/replication. Here is the link for the mission .

We are excited to hear your feedback. Please feel free to comment on the Discovery mission here or you can reach out to Alina Novosolova  or myself  if you have issues accessing the mission.

 

Sara Sampaio

Sara Sampaio

Author Since: March 10, 2022

0 0 votes
Article Rating
Subscribe
Notify of
0 Comments
Inline Feedbacks
View all comments
0
Would love your thoughts, please comment.x
()
x