As a Data Engineer, you will work with a Data Engineering team and will be responsible for the development and maintenance of data integration processes, ensuring the efficient and accurate extraction, transformation, and loading of data from diverse sources into our data warehouse. You will collaborate with data analysts, Data Engineers, and other stakeholders to support the organization's data needs and ensure data quality and availability.
Our Data Engineer position will be critical to the delivery and implementation of data transformation tools throughout the organization. As such, we are seeking candidates with advanced communication skills who enjoy the delivery of Data Engineering services within the organization.
Key Responsibilities:
- Develop Data Pipelines: Build and maintain scalable and efficient data pipelines to ingest, process, and transform large volumes of data from various sources.
- Data Extraction and Integration: Extract data from various source systems, such as databases, APIs, flat files, and external data feeds, while ensuring data integrity and security.
- Data Transformation: Design and implement data transformation processes to cleanse, reshape, and enrich data, making it suitable for analytical and reporting purposes.
- Performance Optimization: Optimize data processing and storage for performance, scalability, and cost-effectiveness. Identify and resolve performance bottlenecks in data pipelines and systems.
- Data Quality and Governance: Implement data quality checks, validation rules, and data governance policies to ensure high-quality, reliable data. Develop and maintain data documentation and metadata repositories.
- Collaboration and Communication: Collaborate with data scientists, analysts, software engineers, and other stakeholders to understand data requirements and deliver solutions that meet business objectives. Communicate effectively with technical and non-technical stakeholders.
- Troubleshooting and Support: Provide support for data-related issues, troubleshoot problems, and ensure timely resolution of incidents. Participate in on-call rotation as needed.
- Data Loading: Develop ETL jobs to load transformed data into the target data warehouse or data repository, ensuring data consistency and performance.
- Error Handling: Develop error handling and exception management processes to address data issues and maintain data pipeline reliability.
- Documentation: Create and maintain documentation for ETL processes, data lineage, and data transformation rules for knowledge sharing and data governance.
- Data Security: Ensure data security and compliance with relevant data protection and privacy regulations.
Qualifications:
- 4+ years of experience in Data Engineering or related roles, with a proven track record of designing and implementing scalable data solutions.
- 3+ years experience with Python
- Expertise in SQL and database technologies (e.g., PostgreSQL, MySQL, etc.), experience in data manipulation and transformation.
- Hands-on experience with cloud platforms, preferably AWS (especially serverless (lambda, API gateway, eventbridge, secrets manager, RDS, SNS). Will consider Azure, or Google Cloud Platform experience.
- Knowledge of database systems, data modeling, and data warehousing concepts.
- Familiarity with data integration techniques, data profiling, and data quality.
- Strong problem-solving and analytical skills.
- Excellent communication and collaboration skills.
- Attention to detail and a commitment to data accuracy.
Job Types: Full-time, Contract
Pay: $45.00 - $55.00 per hour
Expected hours: 40 per week
Experience level:
- 4 years
Schedule:
- 8 hour shift
Work Location: Remote