Data Aws Sagemaker

2 days ago


Chennai Tamil Nadu, India Virtusa Full time

Technical Skills & Expertise
SQL: Expert-level proficiency (must have).
AWS: Redshift, S3, ECS, Lambda, Glue, SQS, SNS, CloudWatch, Step Functions, CDK SQS/SNS, Athena (must have).
PySpark: Expert (must have).
Python: Strong experience with API integrations, data handling, and automation.(must have)
LLM Integration: Experience integrating LLMs via APIs (e.g., OpenAI, Claude, Bedrock) into data workflows or analytics pipelines.
LLMOps: Understanding of prompt design, prompt tuning, RAG patterns, and model evaluation in production.
Data Modeling & Query Tuning: Hands-on experience in designing optimized schemas and writing performant queries.
Big Data Ecosystem: Solid understanding of Hadoop, Hive, MapReduce.
Orchestration Tools: Airflow (Open Source), MWAA on AWS (intermediate
- nice to have).
Data Migration: Experience with AWS Data Migration Service (DMS).
Analytical Skills: Strong in Exploratory Data Analysis (EDA).
ETL Design Patterns: Proficiency with window functions, reusable ETL frameworks, and scalable automation.

Preferred Knowledge
Exposure to Data Lake vs. Data Warehouse architecture.
Experience in real-time data ingestion and streaming frameworks.
Hands-on with data quality, compliance, and governance frameworks.
Familiarity with security best practices in AWS environments.
Experience with data enrichment or summarization using LLMs.
Familiarity with RAG pipelines, vector databases (e.g., OpenSearch, Pinecone, FAISS), and metadata extraction using LLMs.

Key Responsibilities
Design & Build Intelligent Data Solutions: Develop and maintain scalable, LLM-enabled data pipelines using AWS services like Glue, Redshift, Lambda, Step Functions, S3, Athena, and ECS.
Integrate LLM APIs or fine-tuned models (e.g., OpenAI, HuggingFace, Amazon Bedrock, SageMaker JumpStart) into existing AWS-based data workflows.
Enable RAG (Retrieval-Augmented Generation) pipelines using AWS services and LLMs for advanced analytics and knowledge management.
ETL Development: Build efficient, reusable ETL frameworks that can incorporate LLMs for data enrichment, summarization, classification, or metadata generation.
Use orchestration tools like Airflow/MWAA or Step Functions to manage both data and LLM workflows.
LLM & AI Services in AWS: Work with Amazon Bedrock, SageMaker, or custom containers to deploy, monitor, and scale LLM-based solutions.
Optimize LLM usage for performance and cost within the AWS ecosystem (e.g., caching responses, throttling, model selection).
Data Security & Integrity: Implement security best practices and compliance standards for handling sensitive data within LLM pipelines.
Ensure proper prompt auditing, logging, and governance for LLM interactions.
Monitoring & Troubleshooting: Continuously monitor pipelines (including LLM calls), troubleshoot failures, and optimize performance and cost.
Documentation: Maintain detailed documentation of data architectures, frameworks, and engineering processes.
Architecture & Reviews: Participate in solution architecture, code reviews, and sign-offs to ensure quality and scalability.

**About Virtusa**

Teamwork, quality of life, professional and personal development: values that Virtusa is proud to embody. When you join us, you join a team of 27,000 people globally that cares about your growth — one that seeks to provide you with exciting projects, opportunities and work with state of the art technologies throughout your career with us.

Great minds, great potential: it all comes together at Virtusa. We value collaboration and the team environment of our company, and seek to provide great minds with a dynamic place to nurture new ideas and foster excellence.

Virtusa was founded on principles of equal opportunity for all, and so does not discriminate on the basis of race, religion, color, sex, gender identity, sexual orientation, age, non-disqualifying physical or mental disability, national origin, veteran status or any other basis covered by appropriate law. All employment is decided on the basis of qualifications, merit, and business need.


  • Aws

    3 days ago


    Chennai, Tamil Nadu, India WARPE JOBS Full time

    **Key Responsibilities**: - Design, implement, and manage scalable AI/ML solutions on AWS using services like SageMaker, Lambda, EC2, S3, and DynamoDB. - Collaborate with data scientists and AI engineers to build and deploy machine learning models in a cloud environment. - Develop and manage infrastructure as code (IaC) using tools such as AWS...

  • AWS Data Engineer

    7 days ago


    Chennai, Tamil Nadu, India Tata Consultancy Services Full time ₹ 15,00,000 - ₹ 25,00,000 per year

    Experience: 5-10 YrsLocation- Bangalore,Chennai,Hyderabad,Pune,Kochi,Bhubaneshawar,KolkataKey SkillsAWS Lambda, Python, Boto3 ,Pyspark, GlueMust have SkillsStrong experience in Python to package, deploy and monitor data science appsKnowledge in Python based automationKnowledge of Boto3 and related Python packagesWorking experience in AWS and AWS LambdaGood...

  • AWS Data Engineer

    10 hours ago


    Chennai, India Tata Consultancy Services Full time

    Experience: 5-10 Yrs Location - Bangalore,Chennai,Hyderabad,Pune,Kochi,Bhubaneshawar,Kolkata Key Skills AWS Lambda, Python, Boto3 ,Pyspark, Glue Must have Skills Strong experience in Python to package, deploy and monitor data science apps Knowledge in Python based automation Knowledge of Boto3 and related Python packages Working experience in AWS and AWS...

  • AWS Data Engineer

    4 days ago


    chennai, India Tata Consultancy Services Full time

    Experience: 5-10 YrsLocation - Bangalore,Chennai,Hyderabad,Pune,Kochi,Bhubaneshawar,KolkataKey SkillsAWS Lambda, Python, Boto3 ,Pyspark, GlueMust have SkillsStrong experience in Python to package, deploy and monitor data science appsKnowledge in Python based automationKnowledge of Boto3 and related Python packagesWorking experience in AWS and AWS LambdaGood...

  • Aws Data

    3 days ago


    Chennai, Tamil Nadu, India Virtusa Full time

    Job Summary: Design and implement ETL workflows using AWS Glue, Python, and PySpark. Develop and optimize queries using Amazon Athena and Redshift. Build scalable data pipelines to ingest, transform, and load data from various sources. Ensure data quality, integrity, and security across AWS services. Collaborate with data analysts, data scientists, and...

  • Aws Data

    7 days ago


    Chennai, Tamil Nadu, India Virtusa Full time

    P3-C3-TSTS AWS Data Engineer Design and build scalable data pipelines using AWS services like AWS Glue, Amazon Redshift, and QS/SNS/Cloudwatch/Step function/CDK(or Terrafoam.Develop efficient ETL processes for data extraction, transformation, and loading into data warehouses and lakes. Automate repetitive tasks and build reusable frameworks to improve...

  • Python/aws/databricks

    2 weeks ago


    Chennai, India Resource Tree Full time

    Job Name : Data Engineers with AWS and Databricks experience Role : Data Engineer Min years of experience : 3+ Location : Chennai/WFH, Open Positions : 2 Technology Requirement **:Databricks, SQL, Python, AWS **(Glue, S3, Lambda, EC2, SageMaker) Responsibilities: - Develop and implement data lake solutions and analytics solution using AWS and...

  • Aws Data Pm

    3 days ago


    Chennai, Tamil Nadu, India Virtusa Full time

    P3-C3-TSTS AWS DATA PM We are seeking an experienced AWS Data Manager to lead and manage the design, implementation, and optimization of our cloud-based data infrastructure. This role involves overseeing the architecture, governance, and operations of data platforms hosted on Amazon Web Services (AWS), ensuring scalability, security, and performance to...

  • Architect - AWS

    2 days ago


    tamil nadu, India Tiger Analytics Full time

    AWS Architect Location : Chennai/ Bangalore / Hyderabad Who we areTiger Analytics is a global leader in Data, AI, and Analytics, helping Fortune 500 companies solve their most complex business challenges. We offer full-stack AI and analytics services & solutions to empower businesses to achieve real outcomes and value at scale. We are on a mission to push...

  • Data Analytics

    7 days ago


    Chennai, Tamil Nadu, India NTT DATA Full time

    **Must Have**: - Scripting :Python, Pyspark, Unix scripting (Basic) - Cloud : AWS -Lambda, Glue, Athena,RDS,S3, DynamoDB,SNS, SQS, IAM, Step function, Pipeline monitoring, Security, Access and Role permissions, Logging and monitoring Services. - Database : SQL/Oracle - Tools : Pycham,VishualStudio,AWS CLI - Devops Tools : Terraform ,GIT,GIT actions, CICD...