company-logo-image

Data Engineer

ashley-avatar-image

AI-generated summary

beta

This job is for a Data Engineer who builds and improves data pipelines to help a company's reporting and analytics. You might like this job because you'll work with advanced tools and methods to ensure data quality and support strategic decisions.

RM 5800 - RM 8000

Putrajaya

Job Description

To support the build and modernization of the Data Hub within the client ecosystem— an enterprise data ecosystem — by developing, automating, and sustaining data pipelines for reporting, performance measurement, and analytics.

🎯 Key Responsibilities

  • Design, build, and maintain automated, scalable data pipelines for performance and reporting datasets.
  • Refactor legacy pipelines to meet enterprise data standards and improve performance. 
  • Implement robust data quality routines (validation, anomaly detection, schema checks). 
  • Enhance orchestration and error handling for workflow reliability.
  • Support infrastructure tasks like environment refreshes, upgrades, and disaster recovery testing.
  • Ensure compliance with metadata, lineage, classification, and documentation standards.
  • Collaborate on datasets for executive reporting and strategic planning across the Personal & Commercial Operations (PCO), ODIA (Operational Data, Information & AI).

Job Requirements

🛠️ Core Technical Skills Required

  • 3+ years in hands-on data engineering
  • Tools & Languages:
    • PySpark, Python, SQL 
    • Hadoop ecosystem: Spark, Hive, HDFS
    • Shell scripting, Linux environments
       
  • Query tuning with Spark SQL, Trino, Hive
  • Deployment experience across dev/test/prod
  • Working with partitioned datasets, Parquet, columnar storage
  • Strong grasp of data governance, metadata, and validation
     

Preferred / Bonus Skills

  • Experience with event-driven or real-time processing (e.g., Kafka)
  • Familiarity with vector databases, LLM-supportive architectures
  • Tools: Apache Airflow, OpenShift (OCP4), AWS (for orchestration/scheduling)

We welcome candidates with the skillset above to immediately drop an email to nisha@talentspothub.com for more information and confidential conversation.


Skills

PySpark
Python (Programming Language)
SQL (Programming Language)
Hadoop Distributed File System (HDFS)
Apache Hive
Shell Script
Linux
Apache Parquet

Company Benefits

Working Arrangement

Flexible / Remote, 18 days annual leave

Fun, Open and Transparent Culture

Believe in Autonomy, Believe in Coaching and Mentoring, Believe that Failure is a Growth Mindset

Health & Mental Health Matters

We value our team members and ensure that health, medical and mental health is taken care off!!


Additional Info

Company Activity

Last active - few minutes ago

Career Level

Junior Executive


Company Profile

TalentSpotHub Management Consulting -logo-image

TalentSpotHub Management Consulting

TalentSpotHub Management Consulting: Your Flexible Executive Search Partner At TalentSpotHub Management Consulting, we’re dedicated to redefining the talent acquisition landscape. As your flexible talent partner, our humble yet passionate team of practice consultants leverages over 16 years of successful placement experience to deliver innovative and ethical talent solutions. We help organizations attract and retain...