Position:Senior Azure Data Engineer with Retail & Python Programming Experience
Location:Positions might go HYBRID next year
Duration: 12+ Months
Job Description:
Must have 15+ years of Total IT experience
Calgary
700 2nd Street SW
Toronto
134 Peter Street
San Francisco
350 Bush Street
Arlington
1515 North Courthouse Road
Westminster
11030 Circle Point Road
Chicago
35 West Wacker Drive
Boston
40 Water Street
Birmingham
205 Hamilton Row
Houston
1111 Bagby Street
Atlanta
384 Northyards Boulevard NW
Miami
2911 Grand Avenue
New York
375 Hudson Street
Minneapolis
500 North 3rd Street
Los Angeles
13031 West Jefferson Boulevard
Seattle
US WA Seattle 1448 NW Market Street
Dallas/Irving
6021 Connection Drive
Irvine
5301 California Ave
Required Skills and Qualifications:
Strong proficiency in Python programming language.
Experience in building data pipelines and ETL processes.
Familiarity with Snowflake and SAP systems, including data extraction methods (e.g., APIs, database connectors).
Knowledge of data transformation techniques and tools.
Proficiency in Python libraries and frameworks for data manipulation (e.g., Pandas, NumPy, PySpark).
Understanding of database systems and SQL queries.
Experience with data integration and synchronization.
Familiarity with data governance and compliance principles.
Strong problem-solving and analytical skills.
Excellent communication and collaboration abilities.
Attention to detail and ability to work with large datasets efficiently.
Job Description:
Data Pipeline Development: Design, develop, and maintain robust data pipelines using Python programming language to extract data from Snowflake and SAP systems and transform it into the desired format for the target system.
Implement efficient data ingestion, transformation, and loading processes to ensure accurate and reliable data transfer between systems.
Collaborate with stakeholders to understand data requirements, source systems, and target systems, and design appropriate data pipelines accordingly.
Write clean, optimized, and scalable code to handle large volumes of data and ensure efficient data flow throughout the pipeline.
Monitor and optimize data pipeline performance, identifying and resolving issues, bottlenecks, and data quality problems.
Data Transformation: Define and implement data transformation rules and logic to clean, filter, aggregate, and transform data from Snowflake and SAP into the required format for the target system.
Leverage Python libraries and frameworks such as Pandas, NumPy, or PySpark to manipulate and process data efficiently during the transformation process.
Ensure data quality and integrity by applying data validation, normalization, and standardization techniques.
Develop data mapping and conversion scripts to handle schema differences and ensure data consistency across systems.
Collaborate with data analysts and business stakeholders to understand data semantics and requirements for accurate transformations.
Data Integration and System Connectivity: Establish connectivity with Snowflake and SAP systems, extracting data through APIs, database connectors, or other relevant methods.
Integrate and synchronize data from multiple sources, ensuring data consistency and coherence.
Collaborate with IT teams to implement secure and efficient data transfer mechanisms, adhering to data governance and compliance policies.
Develop error handling and exception management strategies to handle data transfer failures and ensure data integrity during the integration process.
Documentation and Collaboration: Document the data pipeline design, architecture, and implementation details, including data source specifications, transformation rules, and target system requirements.
Collaborate with cross-functional teams, including data analysts, data scientists, and business stakeholders, to understand their data needs and provide necessary support.
Participate in meetings and discussions to align data engineering initiatives with business goals.
Stay up-to-date with emerging technologies, tools, and best practices in data engineering and make recommendations for process improvements.