2,222 Kafka Developer jobs in South Africa
Big Data Developer
Posted today
Job Viewed
Job Description
Data Engineer
The Data Engineer's role entails building and supporting data pipelines that must be scalable, repeatable, and secure. This role functions as a core member of an agile team, whereby these professionals are responsible for the infrastructure that provides insights from raw data, handling and integrating diverse sources of data seamlessly. They enable solutions by handling large volumes of data in batch and real-time by leveraging emerging technologies from both the big data and cloud spaces. Additional responsibilities include developing proof of concepts and implementing complex big data solutions with a focus on collecting, parsing, managing, analysing, and visualising large datasets. They know how to apply technologies to solve the problems of working with large volumes of data in diverse formats to deliver innovative solutions. Data Engineering is a technical job that requires substantial expertise in a broad range of software development and programming fields. These professionals have a knowledge of data analysis, end-user requirements, and business requirements analysis to develop a clear understanding of the business need and to incorporate these needs into a technical solution. They have a solid understanding of physical database design and the systems development lifecycle.
Responsibilities
- Architects' Data Analytics Framework
- Translates complex functional and technical requirements into detailed architecture, design, and high-performing software
- Leads Data and batch/real-time analytical solutions leveraging transformational technologies
- Works on multiple projects as a technical lead, driving user story analysis and elaboration, design and development of software applications, testing, and building automation tools
- Development and Operations
- Database Development and Operations
- Policies, Standards, and Procedures
- Business Continuity & Disaster Recovery
- Research and Evaluation
- Creating data feeds from on-premise to AWS Cloud
- Support data feeds in production on a break-fix basis
- Creating data marts using Talend or a similar ETL development tool
- Manipulating data using Python
- Processing data using the Hadoop paradigm, particularly using EMR, AWS's distribution of Hadoop
- Develop for Big Data and Business Intelligence, including automated testing and deployment
Requisite Experience, Education, Knowledge, and/ or Skills
- Bachelor's Degree in Computer Science, Computer Engineering, or equivalent
- AWS Certification
- Extensive knowledge in different programming or scripting languages
- Expert knowledge of data modelling and understanding of different data structures and their benefits and limitations under particular use cases
- Capability to architect highly scalable distributed systems, using different open source tools
- 5+ years of Data engineering or software engineering experience
- 2+ years of Big Data experience
- 2+ years' experience with Extract, Transform, and Load (ETL) processes
- 2+ years of AWS experience
- 5 years of demonstrated experience with object-oriented design, coding, and testing patterns, as well as experience in engineering (commercial or open source) software platforms and large-scale data infrastructures
- Big Data batch and streaming tools
- Talend
- AWS: EMR, EC2, S3
- Python
- PySpark or Spark
Big Data Developer
Posted today
Job Viewed
Job Description
We are looking for an experienced Big Data Developer to join an international banking technology team in Málaga, Spain. In this role, you will contribute to the development of business applications within the Regulatory & Compliance domain, covering the full software lifecycle from problem analysis to deployment.
You'll work with modern big data technologies, collaborate with users to understand business needs, and provide innovative solutions that meet regulatory and compliance requirements. This is a fantastic opportunity to advance your career in a global environment while enjoying the lifestyle benefits of living in Spain.
Key Responsibilities- Participate in the end-to-end software lifecycle, including analysis, design, development, testing, and deployment.
- Collaborate with business users to identify requirements and deliver strategic technology solutions.
- Optimise and analyse code, applying best practices such as threat modelling and SAST.
- Manage tools and processes for documentation and Application Lifecycle Management (ALM).
- Plan and deliver projects using Agile methodology.
- Support incident resolution, including planned interventions.
- Execute unit, integration, and regression testing.
- Manage release processes and deployment tools.
Qualifications and Experience
Required:
- 3+ years of experience as a Big Data Developer.
- Bachelor's degree in Computer Science, Telecommunications, Mathematics, or a related field.
- Proficiency with GitHub.
- Strong knowledge of databases (Oracle PL/SQL, PostgreSQL).
- Experience with Java and JavaScript.
- Hands-on ETL experience.
- Fluency in English (Spanish is advantageous).
Preferred:
- Familiarity with microservices frameworks (Spring Boot), OpenShift.
- Knowledge of Flink, Drools, Kafka, DevOps tools.
- Agile methodology experience with tools such as Jira and Confluence.
- Exposure to S3, Elastic, and Angular.
- Experience in Transactional Regulatory Reporting.
- Innovative mindset and ability to generate strategic ideas.
Other Requirements:
- Availability to travel.
- Willingness to relocate to Málaga, Spain.
Big Data Data Engineer
Posted 1 day ago
Job Viewed
Job Description
Big Data Data Engineer job vacancy in Johannesburg.
We are seeking a skilled Data Engineer to design and develop scalable data pipelines that ingest raw, unstructured JSON data from source systems and transform it into clean, structured datasets within our Hadoop-based data platform.
The ideal candidate will play a critical role in enabling data availability, quality, and usability by engineering the movement of data from the Raw Layer to the Published and Functional Layers.
OverviewBig Data Data Engineer job vacancy in Johannesburg.
Key Responsibilities:
- Design, build, and maintain robust data pipelines to ingest raw JSON data from source systems into the Hadoop Distributed File System (HDFS).
- Transform and enrich unstructured data into structured formats (e.g., Parquet, ORC) for the Published Layer using tools like PySpark, Hive, or Spark SQL.
- Develop workflows to further process and organize data into Functional Layers optimized for business reporting and analytics.
- Implement data validation, cleansing, schema enforcement, and deduplication as part of the transformation process.
- Collaborate with Data Analysts, BI Developers, and Business Users to understand data requirements and ensure datasets are production-ready.
- Optimize ETL/ELT processes for performance and reliability in a large-scale distributed environment.
- Maintain metadata, lineage, and documentation for transparency and governance.
- Monitor pipeline performance and implement error handling and alerting mechanisms.
- 3+ years of experience in data engineering or ETL development within a big data environment.
- Strong experience with Hadoop ecosystem tools: HDFS, Hive, Spark, YARN, and Sqoop.
- Proficiency in PySpark, Spark SQL, and HQL (Hive Query Language).
- Experience working with unstructured JSON data and transforming it into structured formats.
- Solid understanding of data lake architectures: Raw, Published, and Functional layers.
- Familiarity with workflow orchestration tools like Airflow, Oozie, or NiFi.
- Experience with schema design, data modeling, and partitioning strategies.
- Comfortable with version control tools (e.g., Git) and CI/CD processes.
- Experience with data cataloging and governance tools (e.g., Apache Atlas, Alation).
- Exposure to cloud-based Hadoop platforms like AWS EMR, Azure HDInsight, or GCP Dataproc.
- Experience with containerization (e.g., Docker) and/or Kubernetes for pipeline deployment.
- Familiarity with data quality frameworks (e.g., Deequ, Great Expectations).
- Bachelor’s degree in Computer Science, Information Systems, Engineering, or a related field.
- Relevant certifications (e.g., Cloudera, Databricks, AWS Big Data) are a plus.
Big Data Data Engineer
Posted today
Job Viewed
Job Description
Contract
Experience4 to 25 years
SalaryNegotiable
Job Published03 September 2025
Job Reference No.Job Description
We are seeking a skilled Data Engineer to design and develop scalable data pipelines that ingest raw, unstructured JSON data from source systems and transform it into clean, structured datasets within our Hadoop-based data platform. The ideal candidate will play a critical role in enabling data availability, quality, and usability by engineering the movement of data from the Raw Layer to the Published and Functional Layers.
Key Responsibilities:
- Design, build, and maintain robust data pipelines to ingest raw JSON data from source systems into the Hadoop Distributed File System (HDFS).
- Transform and enrich unstructured data into structured formats (e.g., Parquet, ORC) for the Published Layer using tools like PySpark, Hive, or Spark SQL.
- Develop workflows to further process and organize data into Functional Layers optimized for business reporting and analytics.
- Implement data validation, cleansing, schema enforcement, and deduplication as part of the transformation process.
- Collaborate with Data Analysts, BI Developers, and Business Users to understand data requirements and ensure datasets are production-ready.
- Optimize ETL/ELT processes for performance and reliability in a large-scale distributed environment.
- Maintain metadata, lineage, and documentation for transparency and governance.
- Monitor pipeline performance and implement error handling and alerting mechanisms.
Technical Skills & Experience:
- 3+ years of experience in data engineering or ETL development within a big data environment.
- Strong experience with Hadoop ecosystem tools: HDFS, Hive, Spark, YARN, and Sqoop.
- Proficiency in PySpark, Spark SQL, and HQL (Hive Query Language).
- Experience working with unstructured JSON data and transforming it into structured formats.
- Solid understanding of data lake architectures: Raw, Published, and Functional layers.
- Familiarity with workflow orchestration tools like Airflow, Oozie, or NiFi.
- Experience with schema design, data modeling, and partitioning strategies.
- Comfortable with version control tools (e.g., Git) and CI/CD processes.
Nice to Have:
- Experience with data cataloging and governance tools (e.g., Apache Atlas, Alation).
- Exposure to cloud-based Hadoop platforms like AWS EMR, Azure HDInsight, or GCP Dataproc.
- Experience with containerization (e.g., Docker) and/or Kubernetes for pipeline deployment.
- Familiarity with data quality frameworks (e.g., Deequ, Great Expectations).
Qualifications:
- Bachelor's degree in Computer Science, Information Systems, Engineering, or a related field.
Relevant certifications (e.g., Cloudera, Databricks, AWS Big Data) are a plus.
In order to comply with the POPI Act, for future career opportunities, we require your permission to maintain your personal details on our database. By completing and returning this form you give PBT your consent
If you have not received any feedback after 2 weeks, please consider you application as unsuccessful.
Big DataApache HadoopApache HivePySparkSQLJSONData Engineering
IndustriesBankingFinancial Services
Cloudera Big Data Administrator/Engineer
Posted 10 days ago
Job Viewed
Job Description
What you'll do:
- Platform Administration: Install, configure, upgrade, and monitor Cloudera/CDP clusters, manage HDFS/Ozone storage, and ensure security (Kerberos, Ranger, Sentry).
- Data Pipelines: Build and optimize ingestion and processing pipelines using NiFi and Informatica BDM/DEI, supporting both real-time and batch flows.
- ETL Integration: Develop Informatica mappings and workflows, leveraging pushdown execution to Hive/Impala/Spark; integrate diverse on-prem and cloud data sources.
- Performance Governance: Optimize queries, orchestrate jobs (Airflow, Oozie, Control-M), and ensure compliance with governance/security standards.
Your Expertise:
- Strong hands-on expertise in Cloudera tools: Hive, Impala, HDFS, Ozone, Hue, NiFi.
- Proficiency with Informatica BDM/DEI (ETL/ELT, pushdown optimization, data quality).
- Solid SQL, Linux administration, and scripting (Bash, Python).
- Familiarity with cloud data platforms (AWS, Azure, GCP) and orchestration tools.
- 4+ years in big data administration/engineering, including 2+ years in Informatica BDM/DEI.
Qualifications:
- Bachelorâ€s degree in Computer Science, Engineering, or related field.
- Experience in hybrid or cloud-based big data environments.
Soft Skills:
- Strong troubleshooting and problem-solving mindset.
- Ability to work independently and within cross-functional teams.
- Clear communication and documentation skills.
Other information applicable to the opportunity:
- Contract position
- Location: Johannesburg
Why work for us?
Want to work for an organization that solves complex real-world problems with innovative software solutions? At iOCO, we believe anything is possible with modern technology, software, and development expertise. We are continuously pushing the boundaries of innovative solutions across multiple industries using an array of technologies.†/p>
You will be part of a consultancy, working with some of the most knowledgeable minds in the industry on interesting solutions across different business domains.†/p>
Our culture of continuous learning will ensure that you will have all the opportunities, tools, and support to hone and grow your craft.†/p>
By joining IOCO you will have an open invitation to developer inspiring forums. A place where you will be able to connect and learn from and with your peers by sharing ideas, experiences, practices, and solutions.†/p>
iOCO is an equal opportunity employer with an obligation to achieve its own unique EE objectives in the context of Employment Equity targets. Therefore, our employment strategy gives primary preference to previously disadvantaged individuals or groups.
Cloudera Big Data Administrator/Engineer
Posted today
Job Viewed
Job Description
Research Assistant (Administrative tax data | Big Data)
Posted 26 days ago
Job Viewed
Job Description
UNU-WIDER is seeking exceptional candidates for the position of Research Assistant, based in Pretoria, South Africa, to support the SA-TIED programme. This role involves managing and enhancing tax datasets, assisting researchers, and ensuring high standards of data confidentiality.
For the full job description and application details, please click here.
UNU offers three types of contracts: fixed-term staff positions (General Service, National Officer and Professional), Personnel Service Agreement positions (PSA), and consultant positions (CTC). For more information, see the Contract Types page.
1 articles, publications, projects, experts. #J-18808-LjbffrBe The First To Know
About the latest Kafka developer Jobs in South Africa !
Big Data Developer - Regulatory & Compliance (Relocation to Spain)
Posted 22 days ago
Job Viewed
Job Description
Big Data Developer - Regulatory & Compliance (Relocation to Spain)
We are looking for an experienced Big Data Developer to join an international banking technology team in Málaga, Spain . In this role, you will contribute to the development of business applications within the Regulatory & Compliance domain, covering the full software lifecycle from problem analysis to deployment.
You’ll work with modern big data technologies, collaborate with users to understand business needs, and provide innovative solutions that meet regulatory and compliance requirements.
Responsibilities- Participate in the end-to-end software lifecycle , including analysis, design, development, testing, and deployment.
- Collaborate with business users to identify requirements and deliver strategic technology solutions.
- Optimise and analyse code, applying best practices such as threat modelling and SAST.
- Manage tools and processes for documentation and Application Lifecycle Management (ALM).
- Plan and deliver projects using Agile methodology.
- Support incident resolution, including planned interventions.
- Execute unit, integration, and regression testing.
- Manage release processes and deployment tools.
Required:
- 3+ years of experience as a Big Data Developer .
- Bachelor’s degree in Computer Science, Telecommunications, Mathematics, or a related field.
- Proficiency with GitHub .
- Strong knowledge of databases (Oracle PL/SQL, PostgreSQL).
- Hands-on ETL experience.
- Fluency in English (Spanish is advantageous).
Preferred:
- Familiarity with microservices frameworks (Spring Boot), OpenShift.
- Knowledge of Flink, Drools, Kafka, DevOps tools .
- Agile methodology experience with tools such as Jira and Confluence .
- Exposure to S3, Elastic, and Angular .
- Experience in Transactional Regulatory Reporting .
- Innovative mindset and ability to generate strategic ideas.
- Availability to travel.
- Mid-Senior level
- Full-time
- Information Technology
- Staffing and Recruiting
Intermediate C# Developer with KAFKA
Posted 4 days ago
Job Viewed
Job Description
The role involves building and maintaining scalable, resilient services that integrate into a distributed financial-technology ecosystem.
Required Skills & Experience
- 35 years C#/.NET Core development experience in production systems.
- Strong understanding of microservices principles (domain-driven design, bounded contexts, service-to-service communication).
- Proficiency in containerisation (Best practices, image optimisation, debugging containerised apps).
- Hands-on deployment experience with AKS or Kubernetes (RBAC, ConfigMaps, Secrets, Ingress, scaling strategies).
- Apache Kafka (production experience):
- Administering Kafka clusters
- Designing event-driven applications and event schemas
- Monitoring (Prometheus/Grafana, Confluent Control Center, or similar)
- Handling data consistency and exactly-once/at-least-once semantics
- Experience with Azure cloud services :
- Azure DevOps (pipelines, repos, artifacts)
- Azure Monitor / Application Insights
- Networking basics (VNETs, load balancers, firewalls)
- Azure Storage and Messaging (Event Hubs, Service Bus a plus)
- PostgreSQL experience (schema design, queries, performance tuning).
Intermediate C# Developer with KAFKA
Posted today
Job Viewed