Data Engineer - Pune, India - Publicis Sapient

    Publicis Sapient
    Publicis Sapient Pune, India

    Found in: Talent IN 2A C2 - 1 week ago

    Publicis Sapient background
    Description

    Publicis Sapient Overview :


    Publicis Sapient is a digital transformation partner helping established organizations get to their future, digitally enabled state, both in the way they work and the way they serve their customers.

    We help unlock value through a start-up mindset and modern methods, fusing strategy, consulting and customer experience with agile engineering and problem-solving creativity.

    United by our core values and our purpose of helping people thrive in the brave pursuit of next, our 20,000+ people in 53 offices around the world combine experience across technology, data sciences, consulting and customer obsession to accelerate our clients' businesses through designing the products and services their customers truly value.

    Job Summary :

    As Senior Associate L1 in Data Engineering, you will do technical design, and implement components for data engineering solution. Utilize deep understanding of data integration and big data design principles in creating custom solutions or implementing package solutions.

    • You will independently drive design discussions to insure the necessary health of the overall solution
    • The role requires a handson technologist who has strong programming background in spark/Pyspark and Java / Scala / Python, should have experience in Data Ingestion, Integration and data Wrangling, Computation, Analytics pipelines and exposure to Hadoop ecosystem components.
    - Hands-on knowledge on GCP.

    Role & Responsibilities :

    Your role is focused on Design, Development and delivery of solutions involving:

    • Data Ingestion, Integration and Transformation
    • Data Storage and Computation Frameworks, Performance Optimizations
    • Analytics & Visualizations
    • Infrastructure & Cloud Computing
    • Data Management Platforms
    • Build functionality for data ingestion from multiple heterogeneous sources in batch & realtime
    • Build functionality for data analytics, search and aggregation
    Experience Guidelines :

    Mandatory Experience and Competencies :


    • Overall 4+ years of IT experience with 2+ years in Data related technologies
    • Minimum 2+ years of experience in Big Data technologies
    • Hands-on experience with the Hadoop stack - HDFS, sqoop, kafka, Pulsar, NiFi, Spark, Spark Streaming, Flink, Storm, hive, oozie, airflow and other components required in building end to end data pipeline. Working knowledge on real-time data pipelines is added advantage.
    • Strong experience in at least of the programming language Java, Scala, Python. Java preferable
    • Hands-on working knowledge of NoSQL and MPP data platforms like Hbase, MongoDb, Cassandra, AWS Redshift, Azure SQLDW, GCP BigQuery etc
    Preferred Experience and Knowledge (Good to Have) knowledge of traditional ETL tools (Informatica, Talend, etc) and database technologies (Oracle, MySQL, SQL Server, Postgres) with hands on experience


    • Knowledge on data governance processes (security, lineage, catalog) and tools like Collibra, Alation etc
    • Knowledge on distributed messaging frameworks like ActiveMQ / RabbiMQ / Solace, search & indexing and Micro services architectures
    • Performance tuning and optimization of data pipelines
    • CI/CD - Infra provisioning on cloud, auto build & deployment pipelines, code quality
    • Working knowledge with data platform related services on at least 1 cloud platform, IAM and data security
    • Cloud data specialty and other related Big data technology certifications
    Personal Attributes :


    • Strong written and verbal communication skills
    • Articulation skills
    • Good team player
    • Selfstarter who requires minimal oversight
    • Ability to prioritize and manage multiple tasks
    • Process orientation and the ability to define and set up processes
    )