Job description
Extracts, cleanses, and validates data; troubleshoots and proposes fixes;Writes code that matches provided design and assists with ETL processesManage alerting and monitoring on jobs and systemsProvides technical consulting to users of data warehouse; ensures data availability and accessibility for end users by developing and maintaining the
appropriate tools and processesRequest
Strong communicator – able to communicate clearly and proactively and
manage stakeholder expectations.Understands database technologies and team's tech stack (AWS, Spark, Scala,
Java, Python)Strong analytical problem solving skills; ability to maintain and debug data
systems with little oversightStrong business domain knowledge, ability to translate business
requirements/requests to technical tasksDevelops advanced, efficient, and accurate logic using 1 programming or ETL
language (Python, SQL, Java)Spark, working in RDDs and DataFrames/Datasets API (with emphasis on
DataFrames) to query and perform data manipulationSpark Structured StreamingExperience with any version control, preferably GithubExperience with cloud computing platforms, we use AWS (Kinesis, S3, Lambda,
DynamoDB, terraform)Has experience with ANSI SQL relational database (Oracle, SQL, Postgres,
MySQL)Reliably delivers assigned tasks with limited supervision, may begin to deliver
tightly scoped small projects independentlyCollaborates with immediate team to support broader team initiativesCollaborates with business end users and engineering team to maintain data
warehouse and debug data issuesAt least 3 years of experience working on a software or data engineering team
OR relevant BS degree in Computer Science or Mathematics