Role: Data Engineer
Location: SFO, CA (Remote)
Duration: 6 Months
Must haves
· Relevant experience of atleast 5 years
· Good knowledge on Python, especially for data processing
· Very good knowledge of SQL and experience with writing complex SQL on PostgreSQL
· Experience in data pipeline orchestration tools, preferably Airflow
· Basic understanding of containers and familiarity with docker commands
· Working knowledge any distributed systems like Spark, Hadoop etc.
· Have been part of data engineering engagements involving developing complex data pipelines or ETL/ELT processes involving ingestion & processing of data
· Very good debugging skills
· Flexible to learn new technologies and adapt to dynamic environment
Nice to have
· Exposure to cloud (preferably AWS)
· Working experience on Snowflake
· Basic understanding and usage of Jenkins for continuous deployment
· Understanding of ML model lifecycle and pipelines
- provided by Dice