Before you start the extraction of data from your SAP system, you need to prepare the initial steps in the cloud, to receive the data.
The cloud is organized in SYSTEMS, DATASOURCES and PROJECTS.
The cloud regularly receives information from the extractors which you have installed in your SAP systems. But 2 things needs to happen before the data is displayed in the cloud. 1) setup of the cloud to recognise your SAP environment and the data from those systems, and 2) initial setup of the extractors in your SAP systems including reports variant and batchjob timings and frequency.
In order for Gekkocloud to recognise your SAP systems they need to be created in the cloud.
Navigate to the menu on the left called "New System", its visible after you press Systems.
|System ID||DEV||It makes sense to choose your SYSID here|
|Description||Development||A meaningful description to yourself|
|System Type||OPS||Either DEV or OPS|
Note: Please create your Prod (OPS) system before creating your Dev system in order for you be able to link the OPS system during the Dev system creation.
Proceed to create as many Dev and Ops systems as you want to monitor. If you have more than one Dev system pointing to Production create both the Dev systems in the cloud.
Below is an example of a normal system setup. Remember that its easiest if you create the Ops system first because you need to point a Dev system to an Productive system.
example: A typical Ops/prod system
example: A typical Dev system that connects to production
For a given Dev-Ops system chain you need to set up the required datasources for both your Dev and your Ops system.
The datasources represents a "data-area" in the cloud and they are fed by extractors that you have installed in your systems.
Navigate to the menu on the left called "New Datasource". Its visible after you press Data Sources.
|System ID||DEV||This is a drop-down of the systems you have created|
|Template||TRANSPORT||This is a drop down of the datasources available to you|
|Description||TMS data(Dev)||A meaningful description to yourself|
In a DevOps scenario you need to setup these datasources:
Exactly these datasources and their corresponding SAP extractors needs to be active for you to benefit from a complete DevOps scenario. This model example suggests the frequency and timings for the batch-jobs.
|TRANSPORT||Hourly (delta), One daily full load||Hourly (delta), One daily full load(*)|
|CROSSDEPEND||Daily, some time after midnight||-|
|DUMP||-||Daily, some time after midnight|
|APPSTATS||-||Daily, some time after midnight|
|WEBSTATS||-||Daily, some time after midnight|
|SQLM||-||Daily, some time after midnight|
|HANAREADY||Daily, some time after midnight||-|
|PERFORM||Daily, some time after midnight||-|
(*) If the import to the system does not happen more than once a day the Production TRANSPORT extractor obviously only needs to run after the import has run.