Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

Version 1 Next »

Introduction

This plugin would allow SAP ECC enterprise users to specify one of the Inventory Management data sources as source in Cloud Data Fusion. Typical pipeline scenario would include specifying one of the data sources as source (under Sources plugins) and BigQuery as sink. For the sake of simplicity, this document will cover one data source (Material Movements from Inventory Management - 2LIS_03_BF).

User Experience and Configuration


  1. SAP ECC Setup

    1. SAP ECC users log into SAP Gateway Service Builder (segw) to create and expose the data source as SAP OData service. SAP has published following articles around this step:

      1. OData service introduction 

      2. Step-by-step guide to build OData service

    2. User creates OData service for the data source (2LIS_03_BF). Output is RESTful OData service that can be accessed by Cloud Data Fusion.

    3. User can also use SAP Netweaver Gateway client to test the service (with HTTP request and response)

  2. Cloud Data Fusion - Source configuration

    1. User would log into their CDF instance and click on “Hub”.

    2. Under “Plugins” section, user should be able to find “SAP ECC Source”.

    3. This integration would provide a single JAR files for SAP ECC data sources. User would upload the JAR file and this would deploy the plugin using the JAR file.

    4. Once uploaded, the user is prompted to provide following configuration information. Please note that all the fields below should be macro enabled.


Label

Label Description

User Widget

Name

Name of the source

Textbox

Description

Description of the service and the data source it is going to connect to (e.g. 2LIS_03_BF)

Textbox

OData service URL

OData service URL exposed via SAP ECC

Textbox

User name

Login user name for authenticating API call

Textbox

Password

Password for authenticating API call

Textbox


    1. The same dialog will have “Test connection” button to allow users to validate a validation API call. If failed, the error message will be presented to the user in red next to “Test connection” button.

    2. Once connection has been established successfully to this OData service, it can be used as a source for creating data pipelines using Cloud Data Fusion.

  1. Cloud Data Fusion - Sink setup (e.g. BigQuery for this specific example)

    1. User would create a BigQuery table inside a dataset where the data extracted through OData service API will be ingested.

    2. User needs to ensure that the table schema includes data types that have been mapped as part of this integration (see below for integration considerations section for more details)

Integration considerations


  • Data type mismatch - OData output has data types (section 6) different than the data types available in BigQuery. Some sort of data type conversion needs to happen before the extracted data can be ingested into BigQuery. Below is the suggested mapping of OData data types to CDAP schema data types:

OData data type

CDAP schema data type

Edm.Binary

Schema.Type.String

Edm.Boolean

Schema.Type.Bool

Edm.Byte

Schema.Type.String

Edm.DateTime

Schema.Type.DateTime

Edm.Decimal

Schema.Type.Float

Edm.Double

Schema.Type.Double

Edm.Single

Schema.Type.Float

Edm.Guid

Schema.Type.String

Edm.Int16

Schema.Type.Int

Edm.Int32

Schema.Type.Int

Edm.Int64

Schema.Type.Long

Edm.SByte

Schema.Type.Bytes

Edm.String

Schema.Type.String

Edm.Time

Schema.Type.Time

Edm.DateTimeOffset

Schema.Type.Timestamp

  • Field Level Lineage (FLL): Field level lineage should be available for all the sources and sinks.

References


Plugin Type

  • Batch Source
  • Batch Sink 
  • Real-time Source
  • Real-time Sink
  • Action
  • Post-Run Action
  • Aggregate
  • Join
  • Spark Model
  • Spark Compute


  • No labels