Skip navigation EPAM

Lead Software Engineer (Big Data) Remote

  • hot

Lead Software Engineer (Big Data) Description

Job #: 74256
EPAM is a leading global provider of digital platform engineering and development services. We are committed to having a positive impact on our customers, our employees, and our communities. We embrace a dynamic and inclusive culture. Here you will collaborate with multi-national teams, contribute to a myriad of innovative projects that deliver the most creative and cutting-edge solutions, and have an opportunity to continuously learn and grow. No matter where you are located, you will join a dedicated, creative, and diverse community that will help you discover your fullest potential.

Description


You are curious, persistent, logical and clever – a true techie at heart. You enjoy living by the code of your craft and developing elegant solutions for complex problems. If this sounds like you, this could be the perfect opportunity to join EPAM as a Lead Software Engineer. Scroll down to learn more about the position’s responsibilities and requirements.

REQ #: 276957983

What You’ll Do

  • Lead, design and implement innovative analytical solution using Hadoop, NoSQL and other Big Data related technologies, evaluating new features and architecture in Cloud/ on premise/ Hybrid solutions
  • Work with product and engineering teams to understand requirements, evaluate new features and architecture to help drive decisions
  • Build collaborative partnerships with architects and key individuals within other functional groups
  • Perform detailed analysis of business problems and technical environments and use this in designing quality technical solution
  • Actively participate in code review and test solutions to ensure it meets best practice specifications
  • Build and foster a high-performance engineering culture, mentor team members and provide team with the tools and motivation
  • Write project documentation

Requirements

  • Advanced experience in software development with Big Data technologies (e.g., administration, configuration management, monitoring, debugging and performance tuning)
  • Engineering experience and practice in Data Management, Data Storage, Data Visualization, Disaster Recovery, Integration, Operation, Security
  • Strong experience building data ingestion pipelines (simulating Extract, Transform, Load workload), Data Warehouse or Database architecture
  • Strong experience with data modeling; hands-on development experience with modern Big Data components
  • Cloud: experience in designing, automation, provisioning, deploying and administering scalable, available and fault tolerant systems
  • Good understanding of CI/CD principles and best practices
  • Analytical approach to problem-solving with an ability to work at an abstract level and gain consensus; excellent interpersonal, leadership and communication skills
  • Data-oriented personality and possessing compliance awareness, such as PI, GDPR, HIPAA
  • Motivated, independent, efficient and able to handle several projects: work under pressure with a solid sense for setting priorities
  • Ability to work in a fast-paced (startup like) agile development environment
  • Strong experience in high load and IoT Data Platform architectures and infrastructures
  • Vast experience with Containers and Resource Management systems: Docker, Kubernetes, Yarn
  • Experience in direct customer communications
  • Experience in technology/team leading of data-oriented projects
  • Solid skills in infrastructure troubleshooting, support and practical experience in performance tuning and optimization, bottleneck problem analysis
  • Experienced in different business domains
  • English proficiency
  • Advanced understanding of distributed computing principles
  • Technology stack:
    • Programming Languages: Java/ Scala; Python; SQL; Bash
    • Big Data stack: Hadoop, Yarn, HDFS, MapReduce, Hive, Spark, Kafka, Flume, Sqoop, Zookeper
    • NoSQL: Cassandra/ Hbase; MongoDB
    • Queues and Stream processing: Kafka Streams; Flink; Spark Streaming; Storm; Event Hub; IOT Hub; MQTT; Storage Queues; Service Bus; Stream Analytics
    • Data Visualization: Tableau, QlikView
    • ETL & Streaming Pipelines: Pentaho; Talend; Apache Oozie, Airflow, NiFi; Streamsets
    • Operation: Cluster operation, Cluster planning
    • Search: Solr, Elasticsearch/ELK
    • InMemory: Ignite, Redis
    • Solid Cloud experience with 2 or more leading cloud providers (AWS/Azure/GCP): Storage; Compute; Networking; Identity and Security; NoSQL; RDBMS and Cubes; Big Data Processing; Queues and Stream Processing; Serverless; Data Analysis and Visualization; ML as a service (SageMaker; Tensorflow)
    • Enterprise Design Patterns (ORM, Inversion of Control etc.)
    • Version Control Systems (Git, SVN)
    • Testing: Component/ Integration Testing, Unit testing (JUnit)
    • Deep understanding of SQL queries, joins, stored procedures, relational schemas; SQL optimization
    • Experience in various messaging systems, such as Kafka, ZeroMQ/ RabbitMQ
    • Rest, Thrift, GRPC, SOAP
    • Build Systems: Maven, SBT, Ant, Gradle
    • Docker, Kubernetes, Yarn, Mesos

What We Offer

  • Extended Healthcare with Prescription Drugs, Dental and Vision Insurance (Company Paid)
  • Life and AD&D Insurance (Company Paid)
  • Employee Assistance Program (Company Paid)
  • Long-Term Disability
  • Registered Retirement Savings Plan (RRSP) with company match
  • Paid Time Off
  • Critical Illness Insurance
  • Employee Discounts
  • Unlimited access to LinkedIn learning solutions

在亿磐成长

周剑
解决方案架构师
苏州

朱晓华
首席软件测试工程师
苏州

金秋
首席软件工程师
苏州

我们在世界其他地方。。。