Big Data / Kafka / Confluent Architect
- Employer
- WHISHWORKS
- Location
- UK
- Salary
- Competitive
- Closing date
- 23 Feb 2021
View more
- Sector
- Technology & New Media
- Contract Type
- Permanent
You need to sign in or create an account to save a job.
Job Details
We welcome applications from Candidates seeking both Permanent employment or Contract roles.WHISHWORKS is a global IT services and consulting company, specialising in Big Data / Real-time event / Data Streaming design, solutioning and systems integration. Over the years we have helped hundreds of businesses access, control and monetise their data, becoming the partner of choice for data-driven organisations.
Why WHISHWORKSAt WHISHWORKS, we offer our employees a world of potential. We invest in our people to help them develop and achieve their goals. We encourage them to roll up their sleeves, dive in and make the most out of the opportunities bound to come their way.
Role SummaryAs a Big Data Architect, you will be implementing several projects for our clients and become a go to person. You would ideally be a specialist in Data and Cloud technologies with excellent consultative skills. You would be working with various data technologies from companies like Azure, AWS, Kafka, Spark, Hadoop platforms etc.
Primary Responsibility
Skills RequiredMust HaveExperience in Kafka / Confluent installation and administration and deep knowledge of Kafka internals
Experience of Kubernetes and Confluent Operator
Experience of physical data modelling and serialisation formats such as Apache Avro
Strong experience in implementing software solutions in enterprise Linux or Unix environments
Experience in Data Ingestion tools, Streaming, Spark, Hadoop etc.
Hands-on experience in setting up and running Azure and AWS Data platforms, Hadoop, Streaming clusters etc.
Should HaveThe ideal candidate should have 'Platform Consulting & Implementation Experience' and 'Solution Consulting and Implementation Experience'.
Platform Consulting & Implementation Experience
Extensive background in Systems Architecture, Clustering & Distributed Systems, Programming, Security, Networking & Load Balancing, Monitoring, Scripting and Automation gained in large mission critical infrastructures.
Experience on cloud platforms like AWS and Azure, hands-on working knowledge on cloud native data components.
Experience with integrating various security solutions such as LDAP, Kerberos, SPNego or system / installation management tools into the overall solution.
Strong understanding of OS, network configuration, devices, protocols, and performance optimisation.
Understanding of configuration management systems, devops and automation (e.g. Ansible, Puppet, Chef).
Solution Consulting & Implementation Experience
Demonstrable experience in gathering and understanding customer's business requirements, solutioning and implementing them.
Strong programming experience in any programming languages with design patterns, algorithms, ETL/ELT pipeline design, and data analytics.
In depth work experience on Messaging, Data Ingestion, Stream and Batch Processing etc.
Experience in one or more visualisation tools like Qlikview, Tableau etc.
Nice to HaveAzure, AWS, Hortonworks, MapR or Cloudera certifications.
Experience in installation / upgrade / management of multi node Hadoop clusters using distributions such as Hortonworks, Cloudera, or MapR.
Experience in securing Hadoop clusters using technologies such as Ranger, Sentry, data encryption over wire or rest, Kerberos, etc.
Experience in managing Hadoop clusters and performing daily activities to have a cluster with minimal downtime.
Experience in Scala, Spark implementation projects.
Position Requirements
Why WHISHWORKSAt WHISHWORKS, we offer our employees a world of potential. We invest in our people to help them develop and achieve their goals. We encourage them to roll up their sleeves, dive in and make the most out of the opportunities bound to come their way.
Role SummaryAs a Big Data Architect, you will be implementing several projects for our clients and become a go to person. You would ideally be a specialist in Data and Cloud technologies with excellent consultative skills. You would be working with various data technologies from companies like Azure, AWS, Kafka, Spark, Hadoop platforms etc.
Primary Responsibility
- Full lifecycle implementation from requirements analysis, platform selection and setup, technical architecture design, application design and development, testing, and deployment.
- Provide consultative recommendations to clients to solve their Big Data challenges.
- Work as an individual contributor and or team player based on the projects
- A minimum of 10+ years' experience in Data systems and solution consulting.
- A minimum of 5 years' experience in Cloud, Data Management Platforms and Streaming related technologies
- A minimum of 2 years' experience as Data Architect.
Skills RequiredMust HaveExperience in Kafka / Confluent installation and administration and deep knowledge of Kafka internals
Experience of Kubernetes and Confluent Operator
Experience of physical data modelling and serialisation formats such as Apache Avro
Strong experience in implementing software solutions in enterprise Linux or Unix environments
Experience in Data Ingestion tools, Streaming, Spark, Hadoop etc.
Hands-on experience in setting up and running Azure and AWS Data platforms, Hadoop, Streaming clusters etc.
Should HaveThe ideal candidate should have 'Platform Consulting & Implementation Experience' and 'Solution Consulting and Implementation Experience'.
Platform Consulting & Implementation Experience
Extensive background in Systems Architecture, Clustering & Distributed Systems, Programming, Security, Networking & Load Balancing, Monitoring, Scripting and Automation gained in large mission critical infrastructures.
Experience on cloud platforms like AWS and Azure, hands-on working knowledge on cloud native data components.
Experience with integrating various security solutions such as LDAP, Kerberos, SPNego or system / installation management tools into the overall solution.
Strong understanding of OS, network configuration, devices, protocols, and performance optimisation.
Understanding of configuration management systems, devops and automation (e.g. Ansible, Puppet, Chef).
Solution Consulting & Implementation Experience
Demonstrable experience in gathering and understanding customer's business requirements, solutioning and implementing them.
Strong programming experience in any programming languages with design patterns, algorithms, ETL/ELT pipeline design, and data analytics.
In depth work experience on Messaging, Data Ingestion, Stream and Batch Processing etc.
Experience in one or more visualisation tools like Qlikview, Tableau etc.
Nice to HaveAzure, AWS, Hortonworks, MapR or Cloudera certifications.
Experience in installation / upgrade / management of multi node Hadoop clusters using distributions such as Hortonworks, Cloudera, or MapR.
Experience in securing Hadoop clusters using technologies such as Ranger, Sentry, data encryption over wire or rest, Kerberos, etc.
Experience in managing Hadoop clusters and performing daily activities to have a cluster with minimal downtime.
Experience in Scala, Spark implementation projects.
Position Requirements
- Ability to travel up to 50% of the time
- Bachelors or master's degree in computer technology
You need to sign in or create an account to save a job.
Get job alerts
Create a job alert and receive personalised job recommendations straight to your inbox.
Create alert