Develop ETL processes that convert data into formats through a team of data analysts and dashboard charts.
Be responsible for performance, speed, scalability, and extensibility of any application requiring usage of the pipeline.
Collaborate with analytics and business teams to improve data models that feed business intelligence tools, increase data accessibility and foster data-driven decision making across the organization.
Qualifications
Bachelor’s Degree in Computer Science, Software Engineering, Information Technology, or equivalent industry experience.
Experience in programming languages such as Python, SQL, or Scala.
Understand the concepts of Data Lake, Data Warehouse, and Data Mart.
Experience in Big Data technologies and their ecosystem such as Hadoop, Spark and Airflow.
Experience in building and maintaining reliable and scalable ETL on big data platforms as well as experience working with varied forms of data as structured, semi-structured or unstructured data.
Familiar with Cloud Computing Services such as AWS, Azure, or GCP.
Understand the process of data pipeline such as collecting, transforming, and publishing data.
Understand the tools and design platforms that allow processing of data from multiple heterogeneous sources with different frequencies (batch/real-time).
Innovative problem-solving skills with the ability to identify and resolve complex architectural issues.
Ability to translate and clearly formulate technical issues Project.