Requirements
● We are looking for a candidate with 1+ years of experience in a Data Engineer role, who has Bachelor’s in Computer Science or another quantitative field. She/He should also have experience using the following software/tools:
○ Big data tools: Hadoop (YARN, HDFS), Spark, Kafka, etc.
○ Relational SQL and NoSQL databases, including Postgres and Cassandra.
○ Data pipeline and workflow management tools: Airflow, Luigi, etc.
○ Stream-processing systems: Storm, Spark-Streaming, etc.
○ Object-oriented/object function scripting languages: Python, etc.
● Experience building and optimizing data ingestion pipelines
● Being comfortable with working on Linux operated machines.
● A successful history of manipulating, processing and extracting value from large disconnected datasets.
● Strong project management and organizational skills.
● Experience supporting and working with cross-functional teams in a dynamic environment.