We are looking for a passionate certified Data Engineer. The successful candidate will turn data into information, information into insight and insight into business decisions.
Data analyst responsibilities include conducting full lifecycle analysis to include requirements, activities and design. Data analysts will develop analysis and reporting capabilities. They will also monitor performance and quality control plans to identify improvements.
Primary skillset: Python/Pyspark & Azure, ADF, SQL, SQL Server, Data Warehousing, ETL
Secondary: Databricks
Nice to have: Informatica/ETL.
Responsibilities
- Design and develop ETL processes based on functional and non-functional requirements in python / pyspark within Azure platform
- Understand the full end to end development activities from design to go live for ETL development and Azure platform
Recommend and execute improvements - Document component design for developers and for broader communication.
- Understand and adopt an Agile (SCRUM like) software development mindset
- Follow established processes/standards, business technology architecture for development, release management and deployment process
- Execute and provide support during testing cycles and post-production deployment, engage in peer code reviews.
- Elicit, analyze, interpret business and data requirements to develop complete business solutions, includes data models (entity relationship diagrams, dimensional data models), ETL and business rules, data life cycle management, governance, lineage, metadata and reporting elements.
- Apply automation and innovation on new and on-going data platforms for those development projects aligned to business or organizational strategies.
- Design, develop and implement reporting platforms (e.g. modeling, ETL, BI framework) and complex ETL frameworks that meet business requirements.
- Deliver business or enterprise data deliverables (that adhere to enterprise frameworks) for various platforms/servers/applications/systems.
Requirements
- Proven working experience as a data engineer
- Bachelor degree or equivalent in Computer Science
- Skilled in Python object-oriented programming
- Skilled in AWS Compute such as EC2, Lambda, Beanstalk, or ECS
- Skilled in AWS Database products such as Neptune, RDS, Redshift, or Aurora
- Skilled in AWS Management and Governance suite of products such as CloudTrail, CloudWatch, or Systems Manager
- Skilled in Amazon Web Services (AWS) offerings, development, and networking platforms
- Skilled in SQL
- Skilled in Jenkins
- Skilled in JSON
- Skilled in discovering patterns in large data sets with the use of relevant software such as Oracle Data Mining or Informatica
- Skilled in cloud technologies and cloud computing
- Experience using software and computer systems' architectural principles to integrate enterprise computer applications such as xMatters, AWS Application Integration, or WebSphere
- Determining causes of operating errors and taking corrective action
- Experience in the process of analyzing data to identify trends or relationships to inform conclusions about the data
- Skilled in creating and managing databases with the use of relevant software such as MySQL, Hadoop, or MongoDB
- Programming including coding, debugging, and using relevant programming languages
- Communication including communicating in writing or verbally, copywriting, planning and distributing communication, etc.
- Ability to frame ideas as systems and analyzing the inputs, outputs, and process
- Experience helping an organization to plan and manage change in effort to meet strategic objectives
- Adept at managing project plans, resources, and people to ensure successful project completion
- Working with people with different functional expertise respectfully and cooperatively to work toward a common goal