This range is provided by Cititec Talent. Your actual pay will be based on your skills and experience — talk with your recruiter to learn more.
Base pay range
Our client, a leading commodities trading firm, is going through a huge technology and data transformation and is seeking an experienced Senior Data Architect to design and implement a greenfield data ecosystem across on-premise and cloud for trading technology.
Key Responsibilities:
- Data Lake Design & Architecture
- Design and implement a scalable data lake using Apache Iceberg for efficient data storage and processing.
- Architect hybrid on-premise and cloud database solutions that optimise performance and cost.
- Define data partitioning, schema evolution, and versioning strategies to enable efficient query execution.
Data Ingestion & Processing:
- Implement high-performance data ingestion pipelines for structured and unstructured data.
- Utilise Apache Arrow Flight for optimised in-memory data transport and real-time analytics.
- Work with technologies like Kafka, Spark, and Airflow for streaming and batch processing.
Data Governance & Quality:
- Establish data quality frameworks to ensure accuracy, completeness, and consistency.
- Implement metadata management, lineage tracking, and audit logging.
- Define security and compliance policies for sensitive financial and trading data.
- Leverage cloud-native services (AWS/GCP/Azure) and on-premise databases to maximise efficiency.
- Collaborate with DevOps teams to enable CI/CD for data pipelines.
Key Skills & Experience:
- Strong experience in Data Architecture for large-scale financial or commodities trading environments.
- Deep understanding of Apache Iceberg (table format for big data).
- Expertise in Apache Arrow & Arrow Flight for high-speed in-memory data transfer.
- Hands-on experience with cloud (AWS/GCP/Azure) and on-prem databases (PostgreSQL, MySQL, Oracle, etc.).
- Strong background in data ingestion frameworks (Kafka, Flink, Spark, Airflow).
- Knowledge of data governance, security, and regulatory compliance.
- Proficiency in Python, SQL, and distributed data processing frameworks.
- Previous experience in commodities trading or capital markets.
Preferred Technologies & Tools:
- Big Data & Storage: Apache Iceberg, Delta Lake, Hadoop, Parquet, ORC
- Data Processing: Apache Spark, Flink, Dask
- Data Quality & Governance: Great Expectations, dbt, Collibra
Seniority level
Director
Employment type
Full-time
Job function
Information Technology
Industries
Oil and Gas, Investment Management, and Investment Banking