OneMain Financial Jobs

Job Information

UnitedHealth Group Software Engineer Lead - Apache Spark, Scala,Python in Gurgaon, India

Optum is a global organization that delivers care, aided by technology to help millions of people live healthier lives. The work you do with our team will directly improve health outcomes by connecting people with the care, pharmacy benefits, data and resources they need to feel their best. Here, you will find a culture guided by inclusion, talented peers, comprehensive benefits and career development opportunities. Come make an impact on the communities we serve as you help us advance health optimization on a global scale. Join us to start Caring. Connecting. Growing together.

Primary Responsibilities:

  • Design, develop, and support scalable data engineering and AI-enabled application solutions using Apache Spark, Scala, and Python

  • Build, optimize, and maintain high-performance batch and near real-time data pipelines for ingestion, transformation, and analytics across large-scale datasets

  • Develop and manage cloud-based data workflows on AWS, with hands-on use of services such as Amazon EMR, AWS Step Functions, AWS Lambda and Amazon OpenSearch

  • Contribute to the design and deployment of AI agents and Generative AI solutions using platforms such as Google Gemini and AWS Bedrock

  • Implement Retrieval-Augmented Generation (RAG) architectures by integrating large language models with enterprise data sources, search capabilities, and knowledge repositories

  • Work with modern AI integration standards and technologies, including MCP (Model Context Protocol) and related agentic AI frameworks

  • Collaborate closely with cross-functional teams including product management, data science, platform engineering, and business stakeholders to deliver scalable and reliable solutions

  • Ensure application performance, reliability, maintainability, and operational excellence across data and AI systems

  • Participate in technical design discussions, architecture reviews, code reviews, and engineering best practice initiatives

  • Troubleshoot production issues, perform root cause analysis, and implement corrective and preventive measures

  • Support automation, monitoring, deployment, and continuous improvement initiatives across the engineering lifecycle

  • Comply with the terms and conditions of the employment contract, company policies and procedures, and any and all directives (such as, but not limited to, transfer and/or re-assignment to different work locations, change in teams and/or work shifts, policies in regards to flexibility of work benefits and/or work environment, alternative work arrangements, and other decisions that may arise due to the changing business environment). The Company may adopt, vary or rescind these policies and directives in its absolute discretion and without any limitation (implied or otherwise) on its ability to do so

Required Qualifications:

  • Bachelor's degree in Computer Science, Engineering, or a related technical discipline

  • 3+ years of professional software engineering experience, with a solid focus on data engineering, big data systems, AI platforms, or backend application development

  • Solid hands-on experience with Scala and Apache Spark in developing distributed data processing applications

  • Hands-on experience working with AWS cloud services, particularly:

  • Amazon EMR

  • AWS Step Functions

  • Amazon OpenSearch

  • Experience or working knowledge in building and deploying AI agents or Generative AI applications using Google Gemini, AWS Bedrock, or similar LLM platforms

  • Good understanding of RAG concepts, vector-based retrieval, prompt orchestration, and enterprise knowledge integration

  • Familiarity with MCP (Model Context Protocol) and awareness of emerging AI and agentic application technologies

  • Solid understanding of distributed systems, data pipeline architecture, ETL/ELT design, and performance optimization techniques

  • Good knowledge of SQL, data modeling, and processing of structured and unstructured data

  • Familiarity with software engineering best practices including version control, automated testing, code reviews, documentation, and secure development principles

  • Proficiency in Python for data engineering, workflow automation, and AI/ML-related development

  • Proven solid analytical, problem-solving, and debugging capabilities

  • Proven effective communication and collaboration skills, with the ability to work successfully in a team-oriented environment

Preferred Qualifications:

  • Experience with Apache Kafka for streaming, event-driven, or asynchronous data architectures

  • Experience with Docker, Kubernetes, or similar containerization and orchestration technologies

  • Exposure to DevOps practices and related tools for CI/CD, deployment automation, infrastructure management, and monitoring

  • Good working knowledge of shell scripting for operational automation and workflow support

  • Solid Unix/Linux command-line knowledge and system troubleshooting skills

  • Familiarity with workflow orchestration and scheduling platforms used in modern data ecosystems

At UnitedHealth Group, our mission is to help people live healthier lives and make the health system work better for everyone. We believe everyone-of every race, gender, sexuality, age, location and income-deserves the opportunity to live their healthiest life. Today, however, there are still far too many barriers to good health which are disproportionately experienced by people of color, historically marginalized groups and those with lower incomes. We are committed to mitigating our impact on the environment and enabling and delivering equitable care that addresses health disparities and improves health outcomes - an enterprise priority reflected in our mission.

DirectEmployers