HomeJobSenior Data Engineer – AI/ML & Cloud

Senior Data Engineer – AI/ML & Cloud

About the Role:

exp: 5+ years

We are seeking a Senior Data Engineer with AI/ML expertise to design, build, and optimize scalable data pipelines and intelligent analytics solutions. The ideal candidate will combine strong data engineering capabilities with hands-on experience in AI/ML model development, deployment, and MLOps within a modern cloud ecosystem.

________________________________________

Key Responsibilities:

  • Design, develop, and maintain ETL/ELT pipelines for large-scale structured and unstructured data.
  • Build and optimize data models, data lakes, and warehouses using modern cloud platforms (AWS/Azure/GCP).
  • Collaborate with data scientists to operationalize ML models and automate model training and deployment (MLOps).
  • Implement data quality, validation, and monitoring frameworks to ensure reliable pipelines.
  • Develop feature stores and real-time data streaming solutions for AI/ML use cases.
  • Work with business and product teams to understand data requirements and translate them into scalable engineering solutions.
  • Use AI/ML techniques to enhance data enrichment, predictive analytics, and process automation.
  • Manage and optimize data orchestration workflows using tools like Airflow, Databricks, or Prefect.
  • Ensure compliance with data governance, privacy, and security policies.

________________________________________

Required Skills & Qualifications:

  • Bachelor’s/master’s degree in computer science, Data Engineering, or a related field.
  • 7+ years of experience in data engineering, including data pipeline and data lake design.
  • Strong proficiency in Python, SQL, and Pyspark for data processing and ML integration.
  • Hands-on experience with cloud services – AWS (Glue, Redshift, S3, EMR, Lambda), Azure (Data Factory, Synapse), or GCP (Big Query, Dataflow).
  • Solid understanding of AI/ML lifecycle, including model training, validation, and deployment.
  • Experience with MLOps tools such as MLflow, Sage Maker, Kubeflow, or Vertex AI.
  • Knowledge of data orchestration (Airflow, Databricks Workflows) and CI/CD pipelines for data systems.
  • Familiarity with version control (Git), containerization (Docker), and infrastructure-as-code (Terraform).
  • Strong analytical, problem-solving, and communication skills.

________________________________________

Good to Have:

  • Experience with GenAI, LLM fine-tuning, or vector databases (Pinecone, FAISS, Chroma DB).
  • Exposure to data observability and lineage tools (Monte Carlo, Data band, etc.).
  • Prior experience working in Agile / DevOps environments.
Job Category: Software Engineer
Job Type: Contract
Job Location: Remote

Apply for this position

Allowed Type(s): .pdf, .doc, .docx