Skip to main content

Capturing Data with the SAP Extractor

The SAP Extractor is a program in SAP ERP that can be both customized or taken from a standard Data Source. It prepares and captures data through an extract structure that can be transferred to the Business Warehouse of SAP. Both the options of the SAP Extractor help to describe a delta load process or various types of full load. The SAP BW can remotely access the various data transfer activities of the SAP Extractor.



For more on SAP Extractor, click here.

SAP Extractor executes SAP data extraction in three ways.

The first is Content Extraction used to extract BW content, FI, HR, CO, SAP CRM, and LO cockpit. The second is Customer-Generated Extraction where the SAP Extractor is used for LIS, FI-SL, CO-PA. The third is Generic Extraction which is based on DB View, Infoset, Function Modules. The SAP Extractor used for a specific extraction activity depends on the particular needs of an organization.

Data capturing and extraction with the SAP Extractor is initiated with the help of several application-specific extractors. After being hard-coded for the Data Source, these are then delivered together with the BI Content for the BW. The SAP Extractor is so designed that it exactly matches the structure of the Data Source.      

Moreover, there are several generic extractors that are used for extracting SAP data from source systems and transferring to the SAP BW. As the complete data capturing process is automated, the SAP Extractor will automatically know which data has to be extracted and from which tables the data has to be read. A major advantage here is that SAP data extracted can be moved directly to Excel sheets with users having the option to generically extract SAP data.     

Comments

Popular posts from this blog

Snowflake Data Lake – A Cloud-based Data Storage Solution

  Data lakes are data storage repositories that ensure massive volumes of data in its native form – structured, semi-structured, or unstructured – can be stored for processing at a later date. In the past, data storage solutions had various components like data warehouses, data marts, and more. But now, with cloud-based platforms like the Snowflake data lake, all these are not required. Snowflake data lake   is a high-performing cloud-based solution that provides unlimited storage and computing capabilities. Users have the option of scaling up or down in data storage usage and pay only for the resources used. This is critical for businesses as there is no need to invest in additional hardware and software to meet extra storage requirements whenever there is a spike in demand. Another advantage of the Snowflake data lake   is that it has optimized computing capabilities. Even when multiple users simultaneously execute several intricate queries, there is hardly any drop in speed or perf

The Evolution of Technology of Oracle Change Data Capture

Oracle change data capture ( CDC) was first launched with the 9i version as an in-built tool of the Oracle database. It was a tool that recorded and monitored all changes made in the user tables in a database. These changes were then stored in change tables and used in ETL applications for later processing and transferring to other data warehouses and databases. The release version of Oracle change data capture   had triggers placed in the source database. However, database administrators found this technology very invasive and did not favor it. Ultimately, Oracle changed the Oracle change data capture   technology and released it with the 10g version after naming it Oracle Streams.  The working of this release was different. Oracle change data capture   used the redo logs of the source database along with a replication tool of Oracle Streams. This technology turned out to be very successful and a highly optimized method to identify and move change data to a target database without af