Remote Senior Big Data Engineer | EPAM Anywhere

This website uses cookies for analytics, personalization and advertising. Click here to learn more or change your cookie settings. By continuing to browse, you agree to our use of cookies.

Back icon

Senior Big Data Engineer

Senior Big Data Engineer 40 hrs/week, 12+ months

We are currently looking for a remote Senior Big Data Engineer with 3+ years of experience in enterprise software development to join our team.

Please note that even though you are applying for this position, you may be offered other projects to join within EPAM Anywhere.

We accept CVs only in English.

Join EPAM Anywhere to quickly and easily find projects that match your knowledge and experience, while working with Forbes Global 2000 clients, building a successful IT career, and earning competitive rewards. The platform provides additional perks, including a flexible schedule, professional development opportunities, and access to a community of experts.

Responsibilities

  • Design and implement innovative analytical solution using Hadoop, NoSQL and other Big Data related technologies
    • Work with product and engineering teams to understand requirements, evaluate new features and architecture to help drive decisions
      • Perform detailed analysis of business problems and technical environments
        • Participate in code review and test solutions to ensure it meets best practice specifications
          • Build and foster a high-performance engineering culture, mentor team members and provide team with the tools and motivation
            • Write project documentation

              Requirements

              • Over 3 years of experience in enterprise software development
                • Solid background in BigData and distributed computing for 3+ years
                  • Experienced and highly self-motivated professional with outstanding analytical and problem solving skills
                    • Able to work closely with customers and other stakeholders
                      • Good experience in developing BigData, highly available, largely scalable applications and systems
                        • Able to play mentoring role on a project and ensure that solutions meet business requirements and expectations
                          • Experienced in working with modern Agile developing methodologies and tools
                            • Advanced experience in software development with Big Data technologies (e.g. administration, configuration management, monitoring, debugging and performance tuning)
                              • Engineering experience and practice in Data Management, Data Storage, Data Visualization, Disaster Recovery, Integration, Operation, Security
                                • Experience building data ingestion pipelines, Data Warehouse or Database architecture.
                                  • Experience with data modeling; hands-on development experience with modern Big Data components
                                    • Cloud: experience in designing, deploying and administering scalable, available and fault tolerant systems
                                      • Good understanding of CI/CD principles and best practices
                                        • Analytical approach to problem; excellent interpersonal, mentoring and communication skills
                                          • Data-oriented personality and possessing compliance awareness, such as PI, GDPR, HIPAA
                                            • Motivated, independent, efficient and able work under pressure with a solid sense for setting priorities
                                              • Ability to work in a fast-paced (startup like) agile development environment
                                                • Experience in high load and IoT Data Platform architectures and infrastructures
                                                  • Experience with Containers and Resource Management systems: Docker, Kubernetes, Yarn
                                                    • Experience in direct customer communications
                                                      • Solid skills in infrastructure troubleshooting, support and practical experience in performance tuning and optimization, bottleneck problem analysis
                                                        • Experienced in different business domains
                                                          • English proficiency
                                                            • Advanced understanding of distributed computing principles

                                                              Technologies

                                                              • Programming Languages: Java, Scala, Python, SQL, Bash
                                                                • Big Data stack: Hadoop, Yarn, HDFS, MapReduce, Hive, Spark, Kafka, Flume, Sqoop, ZooKeeper
                                                                  • NoSQL: Cassandra/Hbase, MongoDB
                                                                    • Queues and Stream processing: Kafka Streams, Spark Streaming, Event Hub, IoT Hub, Storage Queues, Service Bus, Stream Analytics
                                                                      • Data Visualization: Tableau, QlikView
                                                                        • ETL & Streaming Pipelines: Pentaho, Talend, Apache Oozie, Airflow, NiFi, Streamsets
                                                                          • Operation: Cluster operation, Cluster planning
                                                                            • Search: Solr, Elasticsearch/ELK
                                                                              • InMemory: Ignite, Redis
                                                                                • Cloud (AWS/Azure/GCP): Storage, Compute, Networking, Identity and Security, NoSQL, RDBMS and Cubes, Big Data Processing, Queues and Stream Processing, Serverless. Data Analysis and Visualization, ML as a service (SageMaker, Tensorflow)
                                                                                  • Enterprise Design Patterns (ORM, Inversion of Control etc.)
                                                                                    • Development Methods (TDD, BDD, DDD)
                                                                                      • Version Control Systems (Git, SVN)
                                                                                        • Testing: Component/ Integration Testing, Unit testing (JUnit)
                                                                                          • Deep understanding of SQL queries, joins, stored procedures, relational schemas, SQL optimization
                                                                                            • Experience in various messaging systems, such as Kafka, RabbitMQ
                                                                                              • Rest, Thrift, GRPC, SOAP
                                                                                                • Build Systems: Maven, SBT, Ant, Gradle
                                                                                                  • Docker, Kubernetes, Yarn

                                                                                                    We offer

                                                                                                    • Competitive compensation depending on experience and skills
                                                                                                      • Work on enterprise-level projects on a long-term basis
                                                                                                        • You will have 100% remote full-time job
                                                                                                          • Unlimited access to learning resources (EPAM training courses, English classes, Internal Library)
                                                                                                            • Community of 38,000+ industry's top professionals
                                                                                                              Big Data
                                                                                                              Management
                                                                                                              Docker
                                                                                                              Kubernetes

                                                                                                              40 hrs/week

                                                                                                              Hours per week

                                                                                                              12+ months

                                                                                                              Project length

                                                                                                              Belarus, Russia, Ukraine

                                                                                                              Locations eligible for the position