Python Pyspark with Aws

1 week ago


Andhra Pradesh, India Virtusa Full time

**Responsibilities**:
1.Analyze various data sources present with the client, including structured and unstructured data formats

2.Map entities across different data sources to identify relationships and data inconsistencies

3.Analyze existing reports created with Tableau and other tools to understand data usage and identify gaps

4.Colloborate with data engineers to define data extraction, transformation, and Loading (ETL) processes

5.Develop a target data model with tables and relationships representing the business domain

6.Document data model and its components for future reference

7.support data viz efforts providing insights and analysis

Required skills:
1.Strong analytical and problem-solving skills

2.Experience with data analysis tools and techniques (SQL, Python, Excel)

3.Proficient in data visualization tools (Tableau, Power BI)

4. Understanding of data modeling concepts

5. Excellent communication and collaboration skills

Good to have:
Familiarity with querying data from salesforce.

understanding of data integration tools like Informatica and their use in data analysis projects.

**About Virtusa**

Teamwork, quality of life, professional and personal development: values that Virtusa is proud to embody. When you join us, you join a team of 30,000 people globally that cares about your growth — one that seeks to provide you with exciting projects, opportunities and work with state of the art technologies throughout your career with us.

Great minds, great potential: it all comes together at Virtusa. We value collaboration and the team environment of our company, and seek to provide great minds with a dynamic place to nurture new ideas and foster excellence.

Virtusa was founded on principles of equal opportunity for all, and so does not discriminate on the basis of race, religion, color, sex, gender identity, sexual orientation, age, non-disqualifying physical or mental disability, national origin, veteran status or any other basis covered by appropriate law. All employment is decided on the basis of qualifications, merit, and business need.


  • Pyspark, Hadoop

    1 week ago


    Andhra Pradesh, India Virtusa Full time

    Job Summary: Design and implement ETL workflows using AWS Glue, Python, and PySpark. Develop and optimize queries using Amazon Athena and Redshift. Build scalable data pipelines to ingest, transform, and load data from various sources. Ensure data quality, integrity, and security across AWS services. Collaborate with data analysts, data scientists, and...

  • Pyspark QA

    1 week ago


    Andhra Pradesh, India Virtusa Full time

    **JOB DESCRIPTION** **Skill: PySpark QA** **Role / Tier: Lead Software Engineer/Tier 2** **Experience: 6 - 9 years** Primary Skills BIG Data technology mentioned below Hadoop / Big Data (HDFS, PYTHON, SPARK-SQL, MapReduce) with PYSpark. build CI/CD pipelines Spark APIs to cleanse, explore, aggregate, transform, store & analyse installing, configuring,...

  • Bigdata Pyspark

    1 week ago


    Andhra Pradesh, India Virtusa Full time

    Overall 10+ years of experience in Datawarehouse and Hadoop platform. MUST have experience with Python/PySpark and Hive in Big Data environments Should have strong skills in writing complex SQL Queries and good understanding of Data warehouse concepts. exposure to migration of legacy Data warehouse platform to Hadoop platform experience will be a big...

  • Aws Python

    4 days ago


    Andhra Pradesh, India Virtusa Full time

    JD for Data Engineer Python At least 5 to 8 years of experience in AWS Python programming and who can design, build, test & deploy the code. Should have experience in using following AWS services: AWS SQS, AWS MSK, AWS RDS Aurora DB, BOTO 3. Very strong SQL knowledge is a must, should be able to understand build complex queries. He/she should be closely...

  • Databricks + Pyspark

    2 weeks ago


    Andhra Pradesh, India Virtusa Full time

    **Detailed Job Description for**Databricks + PySpark Developer**: - Data Pipeline Development: Design, implement, and maintain scalable and efficient data pipelines using PySpark and Databricks for ETL processing of large volumes of data. - Cloud Integration: Develop solutions leveraging Databricks on cloud platforms (AWS/Azure/GCP) to process and analyze...

  • Aws Databricks

    3 days ago


    Andhra Pradesh, India Virtusa Full time

    Must have 7plus years of experience total IT experience in Data engineering DW area Must have 3plus yrs of exposure on AWS ecosystem specifically DataBricks, Redshift Lambda Glue Starbust Minimum 5 yrs of technical expertise with data models data mining and segmentation techniques Expert on programming languages SQL Python Pyspark Must be good at...

  • Aws Architect

    12 hours ago


    Andhra Pradesh, India Virtusa Full time

    Good proficiency in Java. Preferable to have Python skills. Expert level skills in SQL queries in databases like Oracle, SQL Server Expertise in AWS Compute (Fargate), AWS Storage (S3), AWS Data Integration (preferably Glue), Lambda, Secrets Manager and AWS Core services. Excellent knowledge in Spark, Databricks Workflows and Pipelines. Good understanding of...

  • Gcp Python

    4 days ago


    Andhra Pradesh, India Virtusa Full time

    Role: Lead Application Developer GCP JD: **Skills**:Knowledge of Google Cloud Platform Google Cloud : GCP cloud storage, Data proc, Big query sQL - Strong SQL & Advanced SQL Spark - writing skills on Pyspark DWH - Data warehousing concepts & dimension modeling (Good to have)Python GIT Any GCP Certification Roles & Responsibilities: Perform data analytics,...

  • Data Engineer

    2 weeks ago


    Andhra Pradesh, India Virtusa Full time US$ 90,000 - US$ 1,20,000 per year

    RequirementsExperience At least 5 years of experience in AWS based projects.Technical skills Proficiency in Python and PySpark for data engineering tasks.Big Data Strong knowledge of Big Data technologies and data warehousing concepts.AWS services Experience with AWS Data Engineering stack, including S3, RDS, Athena, Glue, Lambda, and Step Functions.SQL...

  • Python

    1 week ago


    Andhra Pradesh, India Virtusa Full time

    Solid understanding and working experience with AWS cloud platform - fundamentals: AWS (e.g., S3, Lambda, SageMaker AI, EC2, Bedrock Agents, CodePipeline, EKS) Python environment setup, dependency management (Ex. pip, conda ) and API Integrations(API Keys, OAuth) **About Virtusa** Teamwork, quality of life, professional and personal development: values...