Lead II - Software Engineering

UST Global


Date: 1 day ago
City: Hyderabad
Contract type: Full time
    7 - 9 Years 3 Openings Hyderabad


Role description

Job Summary: As a Product Engineer - Big Data, you will be responsible
for designing, building, and optimizing large-scale data processing pipelines
using the latest Big Data technologies. You will collaborate with
cross-functional teams, including data scientists, analysts, and product
managers, to ensure data is easily accessible, secure, and reliable. Your focus
will be on delivering high-quality, scalable solutions for data storage,
ingestion, and analysis, while also driving continuous improvements across the
data lifecycle. Key Responsibilities:

  • ETL Pipeline Development &
    Optimization Design and implement complex end-to-end ETL pipelines to handle
    large-scale data ingestion and processing.
  • Big Data Processing Develop and
    optimize real-time and batch data processing systems using Apache Spark, Scala
    Spark, and Apache Kafka. Ensure the data is processed in a fault-tolerant
    manner, with strong focus on scalability and performance. Knowledge on Java
    & NoSQL will be good to have.
  • Cloud Infrastructure Development Build
    scalable cloud-based data infrastructure leveraging AWS tools. Ensure data
    pipelines are resilient and adaptable to changes in data volume and variety,
    with a focus on minimizing costs and maximizing efficiency.
  • Data Analysis
    & Insights Work closely with business teams and data scientists to
    understand data needs and deliver high-quality datasets. Conduct in-depth
    analysis to derive insights from the data, identifying key trends, patterns,
    and anomalies that can drive business decisions. Present findings in a clear
    and actionable format.
  • Real-time & Batch Data Integration Enable seamless
    integration of both real-time streaming & batch data from systems like AWS
    MSK. Ensure consistency in data ingestion and processing across different
    formats and sources, providing a unified view of the data ecosystem.
  • CI/CD
    & Automation Utilize Jenkins to establish and maintain continuous
    integration and delivery pipelines. Implement automated testing and deployment
    workflows, ensuring that new features and updates are seamlessly integrated
    into production environments without disruptions.
  • Data Security &
    Compliance Collaborate with security teams to ensure that data pipelines comply
    with organizational and regulatory standards, including GDPR, HIPAA, or any
    other relevant compliance frameworks. Implement data governance frameworks to
    ensure data integrity, security, and traceability throughout the data
    lifecycle.
  • Collaboration & Cross-Functional Work Partner with other
    engineers, data scientists, product managers, and business stakeholders to
    understand data requirements and deliver scalable solutions. Collaborate in
    agile teams, participate in sprint planning, and contribute to architectural
    discussions.
  • Troubleshooting & Performance Tuning Identify and resolve
    performance bottlenecks in data pipelines. Ensure optimal performance through
    proactive monitoring, tuning, and applying best practices for data ingestion
    and storage. Skills & Qualifications: Must-Have Skills: 1. AWS Expertise: Hands-on
    experience with core AWS services related to Big Data, including but not
    limited to EMR, Managed Apache Airflow, Glue, S3, DMS, MSK, and EC2. Deep
    understanding of cloud-native data architecture. 2. Big Data Technologies:
    Proficiency in PySpark/Scala Spark and SQL for data transformations and
    analysis. Experience working with large-scale data processing frameworks such
    as Apache Spark and Kafka. 3. Data Frameworks: Strong knowledge of Spark
    Dataframe & datasets. 4. Database Modeling & Data Warehousing: Expertise
    in designing and implementing scalable data models for OLAP and OLTP systems.
    5. ETL Pipeline Development: Proven experience in building robust, scalable ETL
    pipelines for processing both real-time and batch data across various
    platforms. 6. Data Analysis & Insights: Ability to conduct complex data
    analysis to extract valuable business insights. Strong problem-solving skills
    with a data-driven approach to decision-making. 7. CI/CD & Automation:
    Basic to intermediate knowledge of CI/CD pipelines using Jenkins or similar
    tools to automate deployment and monitoring of data pipelines. Preferred
    Skills:
  • Familiarity with data governance frameworks and tools to ensure
    compliance and security.
  • Knowledge of monitoring tools such as AWS
    CloudWatch, Splunk or Dynatrace to keep track of the health and performance of
    data systems.

Skills

big data,scala spark,apache spark,etl pipeline development,


About UST

UST is a global digital transformation solutions provider. For more than 20 years, UST has worked side by side with the world’s best companies to make a real impact through transformation. Powered by technology, inspired by people and led by purpose, UST partners with their clients from design to operation. With deep domain expertise and a future-proof philosophy, UST embeds innovation and agility into their clients’ organizations. With over 30,000 employees in 30 countries, UST builds for boundless impact—touching billions of lives in the process.

How to apply

To apply for this job you need to authorize on our website. If you don't have an account yet, please register.

Post a resume

Similar jobs

Product Manager

Ryan, LLC, Hyderabad
16 hours ago
Why Ryan? Global Award-Winning Culture Flexible Work Environment Generous Paid Time Off World-Class Benefits and Compensation Rapid Growth Opportunities Company Sponsored Two-Way Transportation Exponential Career Growth The Manager will solicit, understand, and document the customer's and product team’s business requirements, processes, and workflows developing written requirements and process flows. The Manager will also work with developers to create the functional...

Al Architect

Tata Consultancy Services, Hyderabad
1 day ago
JD Ø 3+ years on experience in Java Spring boot / .Net technologies Ø 1+ year hand-on experience in Python and building AI/ML applications using pre-built AI/ML models Ø Language & Frameworks - Python, First API, Flask, NLP, Semantic Kernel/Google Agentic Development Kit, LangGraph etc · Ø Development Platform – Azure AI Factory, Google AgentSpace, AWS Bedrock etc · Ø...

Software Engineer

Solera, Hyderabad
2 days ago
Who We Are Solera is a global leader in data and software services that strives to transform every touchpoint of the vehicle lifecycle into a connected digital experience. In addition, we provide products and services to protect life’s other most important assets: our homes and digital identities. Today, Solera processes over 300 million digital transactions annually for approximately 235,000 partners...