Các Phúc Lợi Dành Cho Bạn
Mô Tả Công Việc
- Create and maintain optimal data pipeline architecture;
- Identify, design, and implement internal process enhancement: automating manual processes, optimizing data delivery & infrastructure for greater scalability, etc.
- Assemble large, complex datasets that meet functional or non-functional business requirements;
- Develop the solution required for optimal ETL of data from a wide variety of data sources;
- Work with stakeholders to assist with data-related technical issues and support their data needs;
- Work closely with Data Science team to build data analytics products by utilizing the data pipeline and implementing necessary services to provide actionable insights into operational efficiency and other key performance metrics
Yêu Cầu Công Việc
+ Big data tools: Hadoop, Spark, Kafka, etc.
+ Relational SQL and NoSQL databases, including MySQL, MongoDB & ArangoDB;
+ Data pipeline and workflow management tools: Luigi, Airflow, etc.
+ Stream-processing systems: Spark-Streaming, etc.
+ Object-oriented/object function scripting languages: Python, Java, C++, Scala, etc.
- Experience in building and optimizing ‘big data’ data pipelines, architectures and data sets;
- Experience in performing root cause analysis on internal, external data and processes to answer specific business questions and identify opportunities for improvement;
- Experience in manipulating, processing and extracting value from large disconnected datasets;
- Experience in supporting and working with cross-functional teams in a dynamic & fast-paced environment.