ETL Work Flow for Extract Transform Loading?Journal: International Journal of Computer Science and Mobile Computing - IJCSMC (Vol.3, No. 6)
Publication Date: 2014-06-30
Authors : J.Anitha; Prasad Babu;
Page : 610-617
Keywords : ETL; ETL process; ETL tool; Data warehouse refreshment;
Extraction, Transformation and Loading (ETL) processes are responsible for the operations taking place in the back stage of a data warehouse architecture. In a high level description of an ETL process, Initially the data is extracted from the source data stores that can be On-Line Transaction Processing (OLTP) or legacy systems files under any format, web pages. various kinds of documents (e.g., spreadsheets and text documents) or even data coming in a streaming fashion. Typically, only the data that are different from the previous execution of an ETL process (newly inserted, updated, and deleted information) should be extracted from the sources. After this phase, the extracted data are propagated to a special purpose area of the warehouse, called the Data Staging Area (DSA), where their transformation, homogenization and cleaning take place. The most frequently used transformations include filters and checks to ensure that the data propagated to the warehouse respect business rules and integrity constraints, as well as schema transformations that ensure that data fit the target data warehouse schema. Finally, the data are loaded to the central data warehouse (D W) and all its counterparts (e.g., data marts and views). In a traditional data warehousing setting, the ETL process periodically refreshes the data warehouse during idle or low-load, periods of its operation ( e.g., every night) and has a specific timewindow to complete. Now a days, Business necessities and demands require near real time data warehouse refreshment and significant attention is drawn to this kind of technological advancement.
Other Latest Articles
Last modified: 2014-06-26 19:38:57