Share
Key job responsibilities
You'll solve data warehousing problems on a massive scale and apply cloud-based AWS services to solve challenging problems around: big data processing, data warehouse design, self-service data access, automated data quality detection and building infrastructure as a code. You'll be part of the team that focuses on automation and optimization for all areas of DW/ETL maintenance and deployment.
- Experience in at least one modern scripting or programming language, such as Python, Java, Scala, or NodeJS
- Experience with big data technologies such as: Hadoop, Hive, Spark, EMR
- Experience building/operating highly available, distributed systems of data extraction, ingestion, and processing of large data sets
- Experience with data modeling, warehousing and building ETL pipelines
- Knowledge of distributed systems as it pertains to data storage and computing
- Knowledge of professional software engineering & best practices for full software development life cycle, including coding standards, software architectures, code reviews, source control management, continuous deployments, testing, and operational excellence
- Experience as a Data Engineer or in a similar role
- Experience with SQL
- Experience with AWS technologies like Redshift, S3, AWS Glue, EMR, Kinesis, FireHose, Lambda, and IAM roles and permissions
- Experience working on and delivering end to end projects independently
These jobs might be a good fit