top of page

Katalytx

I

C a t a l y z i n g   g r o w t h

AI/ML : Engineers & Architect

Chennai

Immediate

3+ yrs exp

Key Responsibilities

  • Design, develop, train, and deploy machine learning (ML) and deep learning (DL) models for life sciences and manufacturing use cases.

  • Collaborate with cross-functional teams to identify opportunities for AI/ML-driven automation and optimization.

  • Implement end-to-end ML pipelines – from data ingestion and feature engineering to model training, validation, and deployment.

  • Work with big data platforms (Databricks, Snowflake, AWS Sagemaker, Azure ML, etc.) to manage and process large-scale datasets.

  • Develop NLP, computer vision, or predictive analytics models for applications in research, digital labs, and customer analytics.

  • Collaborate with software engineering teams to productionize and integrate AI models into enterprise applications.

  • Conduct performance tuning, model monitoring, and continuous retraining as needed.

  • Stay updated with emerging trends in AI/ML technologies and tools.

 

Required Skills and Qualifications

  • Bachelor’s or Master’s degree in Computer Science, Data Science, Statistics, or related field.

  • Proficiency in Python, SQL, and ML frameworks such as TensorFlow, PyTorch, Scikit-learn, XGBoost, or LightGBM.

  • Strong understanding of data structures, algorithms, and statistics.

  • Experience with cloud-based ML platforms – AWS SageMaker, Azure ML, or GCP Vertex AI.

  • Hands-on experience in MLOps – model versioning, CI/CD, and monitoring using tools like MLflow, Kubeflow, or Airflow.

  • Exposure to NLP, computer vision, or time-series forecasting techniques.

  • Excellent problem-solving, communication, and collaboration skills.

 

Preferred Qualifications

  • Experience in Life Sciences / Healthcare / Pharma / Biotech domains.

  • Familiarity with data engineering tools like Databricks, Apache Spark, or Snowflake.

  • Knowledge of GenAI and LLM fine-tuning (e.g., OpenAI, Hugging Face, LangChain).

  • Experience deploying models as REST APIs or microservices using Flask/FastAPI.

  • Understanding of data governance, model ethics, and regulatory compliance in AI applications.

  • Work on impactful projects driving innovation in healthcare and life sciences.

  • Collaborate with global teams across R&D, engineering, and data science.

  • Access to cutting-edge AI/ML platforms and technology ecosystem.

  • Inclusive, diverse, and innovation-driven culture.

Expert

Skill_edited.jpg

Python, Tensorflow, PyTorch, XGBoost 

Sr. Data Engineer – AWS Databricks / Data Tech Lead – AWS Databricks

Chennai

Immediate

Key Responsibilities

  • Data Pipeline Development: Design and implement robust ETL/ELT pipelines using Databricks, PySpark, and Delta Lake to process structured and unstructured data efficiently.

  • Manage job orchestration, scheduling, and workflow automation through Databricks Workflows or Airflow.

  • Performance Optimization: Tune and optimize Databricks clusters and notebooks for performance, scalability, and cost-efficiency.

  • Data Governance: Implement data governance and lineage using Unity Catalog and other platform-native features

  • Collaboration: Work closely with data scientists, analysts, and business stakeholders to understand data requirements and deliver solutions that meet business needs.

  • Cloud Integration: Leverage cloud platforms (AWS) to build and deploy data solutions, ensuring seamless integration with existing infrastructure.

  • Data Modeling: Develop and maintain data models that support analytics and machine learning workflows.

  • Automation & Monitoring: Implement automated testing, monitoring, and alerting mechanisms to ensure data pipeline reliability and data quality.

  • Documentation & Best Practices: Maintain comprehensive documentation of data workflows and adhere to best practices in coding, version control, and data governance.

 

Required Qualifications

  • Experience: 5+ years in data engineering, with hands-on experience using Databricks and Apache Spark.

  • Programming Skills: Proficiency in Python and SQL; 

  • Cloud Platforms: Strong experience with cloud services such as AWS (e.g., S3, Glue, Redshift

  • Data Engineering Tools: Familiarity with tools like Airflow, Kafka, and dbt.

  • Data Modeling: Experience in designing data models for analytics and machine learning applications.

  • Collaboration: Proven ability to work in cross-functional teams and communicate effectively with non-technical stakeholders.

 

Primary Skill Set

Databricks, Apache Spark/Pyspark, Python, SQL, ETL/ELT development, Delta Lake, Cloud platforms (AWS), Data modeling, Cross-functional collaboration, Communication

 

Secondary Skill Set

Airflow, dbt, Kafka, Hadoop, MLflow, Unity Catalog, Delta Live Tables, Cluster optimization, Data governance, Security and compliance, Databricks certifications

5+ yrs exp

Expert

Skill_edited.jpg

AWS, Databricks

Send your resume to katalytx@gmail.com.

Subject - AI ML Engineer/ Architect

bottom of page