Know ATS Score
CV/Résumé Score
  • Expertini Resume Scoring: Our Semantic Matching Algorithm evaluates your CV/Résumé before you apply for this job role: Data Engineer(PySpark).
United Arab Emirates Jobs Expertini

Urgent! Data Engineer(PySpark) Job Opening In Dubai – Now Hiring Synechron

Data Engineer(PySpark)



Job description

Overview

We have an immediate vacancy for a Data Engineer (PySpark) with 6+ years of experience at Synechron, based in Dubai.

Job Location: Dubai

About Company:

At Synechron, we believe in the power of digital to transform businesses for the better.

Our global consulting firm combines creativity and innovative technology to deliver industry-leading digital solutions.

Synechron's technologies and strategies span end-to-end Artificial Intelligence, Consulting, Digital, Cloud & DevOps, Data, and Software Engineering, servicing financial services and technology firms.

We develop solutions for modernization through FinLabs, including AI, Blockchain, Data Science models, Digital Underwriting, and mobile-first applications.

Synechron has 17,000+ employees across 58 offices in 21 countries.

For more information, please visit our website or LinkedIn.

Diversity, Equity, and Inclusion

Synechron's DEI program, Same Difference, promotes a culture of listening, respect, and opportunity.

We value diverse backgrounds and experiences and support our people to excel.

If you have the skills, enthusiasm, and drive, we'll support you like we support each other.

Choose a career with us and pursue innovation together.

Job Descriptions:

About the Role:

We are seeking a highly skilled Data Engineer with deep expertise in PySpark and the Cloudera Data Platform (CDP) to join our data engineering team.

You will design, develop, and maintain scalable data pipelines that ensure high data quality and availability.

This role requires a strong background in big data ecosystems, cloud-native tools, and advanced data processing techniques.

Experience with data ingestion, transformation, and optimization on CDP is essential, along with implementing data engineering best practices.

You will work with other data engineers to build solutions that drive business insights.

Responsibilities

  • Data Pipeline Development: Design, develop, and maintain scalable ETL pipelines using PySpark on the Cloudera Data Platform, ensuring data integrity and accuracy.

  • Data Ingestion: Implement and manage data ingestion from relational databases, APIs, file systems, and other sources to the data lake or data warehouse on CDP.

  • Data Transformation and Processing: Use PySpark to process, cleanse, and transform large datasets to support analytical needs.

  • Performance Optimization: Tune PySpark code and CDP components to optimize resource use and reduce ETL runtimes.

  • Data Quality and Validation: Implement data quality checks and monitoring to ensure data accuracy and reliability.

  • Automation and Orchestration: Automate data workflows using Apache Oozie, Airflow, or similar tools within the Cloudera ecosystem.

  • Monitoring and Maintenance: Monitor pipeline performance, troubleshoot issues, and perform routine maintenance on CDP and related processes.

  • Collaboration: Work with data engineers, analysts, product managers, and stakeholders to understand data requirements and support data-driven initiatives.

  • Documentation: Maintain thorough documentation of data engineering processes and configurations.

Education and Experience

  • Bachelor's or Master's degree in Computer Science, Data Engineering, Information Systems, or a related field.

  • 3+ years of experience as a Data Engineer, with strong focus on PySpark and the Cloudera Data Platform.

Technical Skills

  • PySpark: Advanced proficiency with RDDs, DataFrames, and optimization techniques.

  • Cloudera Data Platform: Experience with CDP components including Cloudera Manager, Hive, Impala, HDFS, and HBase.

  • Data Warehousing: Knowledge of data warehousing concepts, ETL best practices, and SQL-based tools (Hive, Impala).

  • Big Data Technologies: Familiarity with Hadoop, Kafka, and other distributed computing tools.

  • Orchestration and Scheduling: Experience with Apache Oozie, Airflow, or similar tools.

  • Scripting and Automation: Strong Linux scripting skills.

Soft Skills

  • Strong analytical and problem-solving abilities.

  • Excellent verbal and written communication skills.

  • Ability to work independently and as part of a team.

  • Attention to detail and commitment to data quality.

Role Details

  • Seniority level: Mid-Senior level
  • Employment type: Full-time
  • Job function: Information Technology and Finance
  • Industries: Investment Banking, Banking, and Insurance

#J-18808-Ljbffr


Required Skill Profession

Other General



Your Complete Job Search Toolkit

✨ Smart • Intelligent • Private • Secure

Start Using Our Tools

Join thousands of professionals who've advanced their careers with our platform

Rate or Report This Job
If you feel this job is inaccurate or spam kindly report to us using below form.
Please Note: This is NOT a job application form.


    Unlock Your Data Engineer Potential: Insight & Career Growth Guide