Create and develop data ingestion from various sources: RDBMS, Rest API, Kafka, Text File, and Spreadsheet. Design, develop, optimize, and maintain data architecture and pipelines.
Work with Core Data Engineering / Data Warehousing team to utilize existing frameworks for the implementation of these data pipelines.
Drive the prioritization, strategy, and focus to solve user problems.
Maintain and optimize data pipelines.
Participate in code reviews and follow best practices for development and documentation of data pipelines.
Continuously learn and adapt to new technologies and methodologies within the data engineering landscape.
Qualifications
Experience as a Data Engineer or Data Analyst is preferable.
Excellent command of programming languages, preferably in Python.
Familiarity with managing a serverless data warehouse like BigQuery or Redshift.
Familiarity with schedulers like Airflow & Airbyte.
Experience working with GitHub and Docker.
Deep knowledge of SQL database design (MySQL, Redshift, PostgreSQL).
Understanding of how to optimize data retrieval and develop dashboards, reports, and other visualizations for stakeholders.
Good communication skills to work across departments.