Share the link to this page
Copied

About the Product

Vertica is a SQL analytics database built to manage rapidly growing volumes of data in Data Warehouse. Many times CSV files, received from internal upstream or external vendor system, required to be cleansed, consumed, integrated, and reported in Data Warehouse. Data integration complexity increase to a greater extent if an incoming CSV file consists of dissimilar patterns of record and header layout.

This white paper will show SQL only practical solution approach for consumption and integration of CSV file with similar data layout in the Vertica analytics database without using user-defined function/shell scrips/programming script etc.

What will you learn in this course?

  • Business scenario
  • Header row specification
    • Header row in CSV format
    • Header row in tabular format
  • Data row specification
    • Standard data row in CSV format
    • Detail data row in CSV format
    • Data row in tabular format
  • Sample data files
  • Business requirement
    • Load above data files merging header row information in data rows and store in a raw table in integration schema
    • Load final integration table with clean customer data - update existing and insert new records
  • Prerequisites
    • Copy files in the database server folder
    • Create schema in the database
    • Create staging and integration tables
  • Load and transform staging raw table
    • Load file data
    • Remove null, double quotes, non-breaking space, and folder name from file path with an empty string
    • Split fields based on comma delimiter and trim
    • Copy standard header row to standard data row based on file name join
    • Copy detail header row to detail data row based on file name join
    • Delete header rows
    • Preview of loaded table data
  • Load integration raw table
    • Preview of loaded table data
  • Load integration master table
    • Update existing records based on the key column
    • Insert new records based on the new key column
    • Preview of loaded table data
  • Scope of enhancement adding reject logic and reject reprocessing

Requirements

  • You should have a basic understanding of database and SQL
  • You should be familiar with data engineering concepts
  • You should be working or planning to work as an ETL developer or data engineer in OLTP or DW environment

Authors

School

DataPad OÜ's School

One-time Fee
$25
€23.70
£19.81
CA$35.25
A$38.69
S$33.59
HK$194.63
CHF 22.19
NOK kr277.23
DKK kr176.95
NZ$43.24
د.إ91.82
৳2,996.53
₹2,111.94
RM111.76
₦41,660.11
₨6,962.14
฿869.94
₺860.47
B$144.92
R430.66
Лв46.43
₩34,898.12
₪93.78
₱1,468.60
¥3,858.37
MX$508.62
QR91.41
P342.80
KSh3,247.35
E£1,232.42
ብር3,036.63
Kz22,800
CLP$24,381.72
CN¥180.81
RD$1,510.83
DA3,330.47
FJ$56.89
Q193.66
GY$5,243.70
ISK kr3,458.82
DH249.98
L455.62
ден1,460.57
MOP$201.06
N$456.19
C$922.66
रु3,370.82
S/95.19
K100.82
SAR93.90
ZK688.44
L118.13
Kč599.14
Ft9,680.62
SEK kr274.46
ARS$24,930.61
Bs173.27
COP$112,476.88
₡12,761.36
L633.25
₲195,651.70
$U1,072.18
zł102.41

What's Included

File Size: 533K
Pages: 10
Language: English
Level: Advanced
Skills: Data Transportation, Vertica, Data Massaging, Batch Integration, Data Engineering, CSV File, White Paper
Age groups: 25-34 years, 35-44 years, 45-54 years, 55-64 years

Sign Up

Share

Share with friends, get 20% off
Invite your friends to LearnDesk learning marketplace. For each purchase they make, you get 20% off (upto $10) on your next purchase.