· Collaborate with cross-functional teams to understand data requirements and design efficient data pipelines.
· Develop, implement, and maintain ETL (Extract, Transform and Load) processes to collect, clean, and store data from various sources.
· Optimize and enhance existing data pipelines for better performance, scalability, and reliability.
· Ensure data quality and integrity through data validation, monitoring, and error handling.
· Troubleshoot and resolve data-related issues in a timely manner.
· Collaborate with data scientists and analysts to provide them with the necessary data sets and tools to perform advanced analytics and reporting.
· Implement data security and privacy best practices to protect sensitive information.
· Stay updated with the latest trends and technologies in data engineering and contribute to the continuous improvement of our data infrastructure.
Skills & Qualifications:
· Bachelor's degree in Computer Science, Information Technology, or a related field (or equivalent work experience).
· 6+ years of experience in data engineering or a related role.
· Proficiency in programming languages such as Python, Java, or Scala.
· Strong experience with data warehouse technologies (e.g., AWS Redshift, Google BigQuery, Snowflake).
· Hands-on experience with ETL tools and frameworks (e.g., Apache Spark, Apache NiFi).
· Knowledge of database systems (SQL and NoSQL) and data modeling.
· Familiarity with data integration and orchestration tools (e.g., Apache Airflow).
· Understanding of data security and compliance standards.
· Excellent problem-solving and communication skills.
· Ability to work independently and collaboratively in a fast-paced environment.
· Strong attention to detail and a commitment to delivering high-quality work.
· Experience with version control systems (e.g., Git).
· Certification in relevant data engineering technologies.