Requisition Id : BCEJP00020267
Advanced software development skills using Scala, Java, Spark, Python and Linux scripting
Description: Suitable candidates will have advanced software development skills using Scala, Java, Spark, Python and Linux scripting including prior in-depth experience in: 1. Functional and object-oriented programming, including experience in Scala or Java development using Spark for batch and streaming ingestion data flows and transformation workflows for data within HDFS, HBase and Solr. 2. Development of Sqoop jobs with experience integration to RDBMS such as Oracle, Teradata, SQL Server, DB2, Sybase, etc. 3. Development and implementation of workflow/ schedulers like Oozie 4. Scripting in Linux using bash 5. Development and implementation of Kafka topics and integration of publishers and consumers, including sources such as system or web logs, databases, web services and file sources using StreamSets or Rsyslog 6. Development and implementation of streaming data workflows, including integration of collectors such as system or web logs, databases, web services and file sources and destinations to Kafka, HDFS, Spark Streaming, etc. Familiarity with streaming technologies such as Kafka and Spark streaming is required 7. Developing solutions consistent with security best practices to ensure that sensitive data is secured properly 8. Knowledge of security protocols and methodologies including: SFTP, SSH, SSL Other attributes that are valuable for the role include: - Proven skills in developing high-quality, highly optimized, high performance and maintainable software for big data solutions specifically in the Hadoop ecosystem - Experience in architecture, design, software development, testing, deployment, maintenance, production and operation of data solutions - Experience building and testing code in non-production environments. This includes unit, regression, performance and end-to-end testing - Working experience developing projects in IntelliJ or Eclipse with Maven and integrating to GitHub - Able to follow software development life cycle (SDLC), development and security standards - Ability to measure software performance in non-production and production environments and improve its efficiency - The ability to support customer issues and incidents regarding the big data platforms through to resolution - Building automation for repetitive yet complex tasks through the use of automation technologies to streamline operations - Exposure to Continuous Improvement methods - Proficient understanding of distributed computing principles The candidate should possess a degree in Engineering, Mathematics, Science or Computer Science or alternatively a diploma in software development with a focus on Big Data languages/tools
CDG ISO 14001:2004
Callas/Fort with Minority Supplier Development Council
NCTRCA Minority Buiseness Enterprise Certification
CDG OHSAS 18001:2007
Tech Titans
Deloitte
Best Employer
Graham Bell Innovation
Huawei Gold Partner
Fastest Growing Fastest