This job has expired

Big Data / Kafka / Confluent Architect

Closing date
23 Feb 2021

View more

Technology & New Media
Contract Type
You need to sign in or create an account to save a job.

Job Details

We welcome applications from Candidates seeking both Permanent employment or Contract roles.WHISHWORKS is a global IT services and consulting company, specialising in Big Data / Real-time event / Data Streaming design, solutioning and systems integration. Over the years we have helped hundreds of businesses access, control and monetise their data, becoming the partner of choice for data-driven organisations.

Why WHISHWORKSAt WHISHWORKS, we offer our employees a world of potential. We invest in our people to help them develop and achieve their goals. We encourage them to roll up their sleeves, dive in and make the most out of the opportunities bound to come their way.

Role SummaryAs a Big Data Architect, you will be implementing several projects for our clients and become a go to person. You would ideally be a specialist in Data and Cloud technologies with excellent consultative skills. You would be working with various data technologies from companies like Azure, AWS, Kafka, Spark, Hadoop platforms etc.

Primary Responsibility
  • Full lifecycle implementation from requirements analysis, platform selection and setup, technical architecture design, application design and development, testing, and deployment.
  • Provide consultative recommendations to clients to solve their Big Data challenges.
  • Work as an individual contributor and or team player based on the projects
Desired Skills, Attributes & Experience
  • A minimum of 10+ years' experience in Data systems and solution consulting.
  • A minimum of 5 years' experience in Cloud, Data Management Platforms and Streaming related technologies
  • A minimum of 2 years' experience as Data Architect.
The ideal candidate should have 'Platform Consulting & Implementation Experience' and 'Solution Consulting and Implementation Experience'.

Skills RequiredMust HaveExperience in Kafka / Confluent installation and administration and deep knowledge of Kafka internals

Experience of Kubernetes and Confluent Operator

Experience of physical data modelling and serialisation formats such as Apache Avro

Strong experience in implementing software solutions in enterprise Linux or Unix environments

Experience in Data Ingestion tools, Streaming, Spark, Hadoop etc.

Hands-on experience in setting up and running Azure and AWS Data platforms, Hadoop, Streaming clusters etc.

Should HaveThe ideal candidate should have 'Platform Consulting & Implementation Experience' and 'Solution Consulting and Implementation Experience'.

Platform Consulting & Implementation Experience

Extensive background in Systems Architecture, Clustering & Distributed Systems, Programming, Security, Networking & Load Balancing, Monitoring, Scripting and Automation gained in large mission critical infrastructures.

Experience on cloud platforms like AWS and Azure, hands-on working knowledge on cloud native data components.

Experience with integrating various security solutions such as LDAP, Kerberos, SPNego or system / installation management tools into the overall solution.

Strong understanding of OS, network configuration, devices, protocols, and performance optimisation.

Understanding of configuration management systems, devops and automation (e.g. Ansible, Puppet, Chef).

Solution Consulting & Implementation Experience

Demonstrable experience in gathering and understanding customer's business requirements, solutioning and implementing them.

Strong programming experience in any programming languages with design patterns, algorithms, ETL/ELT pipeline design, and data analytics.

In depth work experience on Messaging, Data Ingestion, Stream and Batch Processing etc.

Experience in one or more visualisation tools like Qlikview, Tableau etc.

Nice to HaveAzure, AWS, Hortonworks, MapR or Cloudera certifications.

Experience in installation / upgrade / management of multi node Hadoop clusters using distributions such as Hortonworks, Cloudera, or MapR.

Experience in securing Hadoop clusters using technologies such as Ranger, Sentry, data encryption over wire or rest, Kerberos, etc.

Experience in managing Hadoop clusters and performing daily activities to have a cluster with minimal downtime.

Experience in Scala, Spark implementation projects.

Position Requirements
  • Ability to travel up to 50% of the time
  • Bachelors or master's degree in computer technology
Our Commitment to YouYou'll benefit from our network of global communities and collaborative culture and can develop valuable industry-specific expertise. The scale of our capabilities, client engagements and the unique way we innovate, operate and deliver value, will provide you the support to deepen your dynamic leadership skills.
You need to sign in or create an account to save a job.

Get job alerts

Create a job alert and receive personalised job recommendations straight to your inbox.

Create alert