Job Summary:
We are seeking a highly skilled Data Engineer to design, build, and maintain the infrastructure required to manage and process large volumes of data. The ideal candidate will work closely with cross-functional teams to ensure the efficient flow, security, and availability of data, enabling analytics, reporting, and AI-driven decision-making.
Key Responsibilities:
- Data Pipeline Development:
- Design and implement robust ETL/ELT pipelines for data ingestion, transformation, and storage.
- Optimize data workflows to ensure scalability and performance.
- Data Integration and Storage:
- Integrate data from multiple sources into centralized data platforms.
- Design and maintain databases, data lakes, and warehouses for structured and unstructured data.
- Data Governance and Quality:
- Ensure data accuracy, consistency, and completeness through monitoring and validation.
- Implement data security and compliance measures in line with industry regulations.
- Collaboration and Support:
- Work with data analysts, AI engineers, and other stakeholders to define data requirements.
- Provide technical support for analytics and reporting tools.
- Continuous Improvement:
- Evaluate and adopt new technologies for improving data architecture and performance.
- Document processes and maintain up-to-date technical guides.
Qualifications and Skills:
Required:
- Bachelor’s or Master’s degree in Computer Science, Data Science, or a related field.
- Proven experience in building and managing data pipelines and architectures.
- Strong knowledge of databases (SQL and NoSQL), data modeling, and schema design.
- Proficiency in data processing tools and frameworks (e.g., Apache Spark, Hadoop).
- Hands-on experience with cloud platforms like AWS, Azure, or Google Cloud.
- Proficiency in programming languages such as Python, Scala, or Java.
Preferred:
Certifications in cloud data services (e.g., AWS Certified Data Analytics, Azure Data Engineer Associate).
Experience with data visualization tools like Power BI or Tableau.
Familiarity with real-time data processing technologies (e.g., Kafka, Flink).
Job Location: This role is Hybrid for Karachi and Islamabad candidates and remote for other cities.