Senior Data Engineer
Senior Data Engineer
CancelledOn site
Industry : Financial Services
Location:
Work Experience: 5 years
Job Description
● Design, develop, and maintain scalable, secure, and cost-effective data infrastructure on cloud platforms such as AWS, GCP, or Azure. ● Implement cloud-native solutions like data lakes, warehouses (e.g., Snowflake, BigQuery, Redshift), and serverless computing. Build, deploy, and manage real-time and batch data pipelines using tools such as Apache Airflow, Apache Kafka, or cloud- native orchestration solutions. ● Optimize ETL/ELT processes for seamless data ingestion, transformation, and integration from diverse sources. Ensure high performance, scalability, and cost-efficiency of cloud data solutions through tuning and capacity planning. ● Implement caching strategies and data partitioning techniques for large-scale datasets. Enforce best practices for data privacy, security, and regulatory compliance (e.g., GDPR, PCI DSS). ● Implement identity management, encryption, and monitoring tools to safeguard sensitive data. Work closely with cross-functional teams to understand business requirements and deliver cloud-based data solutions. ● Mentor junior data engineers and contribute to team knowledge sharing. Evaluate emerging cloud and big data technologies to recommend enhancements to our data infrastructure. ● Drive automation in deployment, testing, and monitoring using CI/CD pipelines and Infrastructure-as-Code (e.g., Terraform, CloudFormation). Requirements ● Bachelor’s or Master’s degree in Computer Science, Engineering, Data Science, or a related field. 5+ years of data engineering experience with at least 3 years working on cloud-based solutions. ● Proven experience in the fintech domain is a significant advantage. Strong expertise in cloud platforms such as AWS (S3, Redshift, Glue), GCP (BigQuery, Dataflow), or Azure (Data Factory, Synapse Analytics). ● Proficiency in programming languages like Python, Java, or Scala. ● Solid knowledge of database systems and SQL, including NoSQL technologies (e.g., DynamoDB, MongoDB). ● Experience with big data tools such as Apache Spark, Hadoop, or Databricks. ● Strong familiarity with containerization (Docker) and orchestration (Kubernetes). ● Hands-on experience with Infrastructure-as-Code tools (Terraform, CloudFormation)