About the Role:
exp: 5+ years
We are seeking a Senior Data Engineer with AI/ML expertise to design, build, and optimize scalable data pipelines and intelligent analytics solutions. The ideal candidate will combine strong data engineering capabilities with hands-on experience in AI/ML model development, deployment, and MLOps within a modern cloud ecosystem.
________________________________________
Key Responsibilities:
- Design, develop, and maintain ETL/ELT pipelines for large-scale structured and unstructured data.
 - Build and optimize data models, data lakes, and warehouses using modern cloud platforms (AWS/Azure/GCP).
 - Collaborate with data scientists to operationalize ML models and automate model training and deployment (MLOps).
 - Implement data quality, validation, and monitoring frameworks to ensure reliable pipelines.
 - Develop feature stores and real-time data streaming solutions for AI/ML use cases.
 - Work with business and product teams to understand data requirements and translate them into scalable engineering solutions.
 - Use AI/ML techniques to enhance data enrichment, predictive analytics, and process automation.
 - Manage and optimize data orchestration workflows using tools like Airflow, Databricks, or Prefect.
 - Ensure compliance with data governance, privacy, and security policies.
 
________________________________________
Required Skills & Qualifications:
- Bachelor’s/master’s degree in computer science, Data Engineering, or a related field.
 - 7+ years of experience in data engineering, including data pipeline and data lake design.
 - Strong proficiency in Python, SQL, and Pyspark for data processing and ML integration.
 - Hands-on experience with cloud services – AWS (Glue, Redshift, S3, EMR, Lambda), Azure (Data Factory, Synapse), or GCP (Big Query, Dataflow).
 - Solid understanding of AI/ML lifecycle, including model training, validation, and deployment.
 - Experience with MLOps tools such as MLflow, Sage Maker, Kubeflow, or Vertex AI.
 - Knowledge of data orchestration (Airflow, Databricks Workflows) and CI/CD pipelines for data systems.
 - Familiarity with version control (Git), containerization (Docker), and infrastructure-as-code (Terraform).
 - Strong analytical, problem-solving, and communication skills.
 
________________________________________
Good to Have:
- Experience with GenAI, LLM fine-tuning, or vector databases (Pinecone, FAISS, Chroma DB).
 - Exposure to data observability and lineage tools (Monte Carlo, Data band, etc.).
 - Prior experience working in Agile / DevOps environments.
 
Job Category: Software Engineer 
Job Type: Contract 
Job Location: Remote