Job Description
Our client is an information technology company with corporate headquarters in Limassol, Cyprus. They are a world leader in the research and development world of state-of-the-art methods for ingesting data from heterogenous sources and adapting it to bespoke and intuitive solutions for civilian protection and they are currently looking to hire a skilled Big Data Team Lead to join their team.
The ideal candidate will be instrumental in accelerating and scaling their data pipelines and data lakes.
Reference #: GC004373
Job Responsibilities
- Establish, lead, manage and mentor the big data team
- Own the development of an inhouse Data Lake for storing structured and unstructured data
- Research, design, and develop appropriate algorithms for Big Data collection, processing, and analysis
- Define how data will be streamed, stored, consumed, integrated by different data systems
- Identify relevant Big Data tools required to support new and existing product capabilities
- Collaborate closely with the product team to define the requirements and milestones that relate to Big Data features
- Closely interact with the Data Scientists in providing feature-ed datasets
- Design, create, deploy, manage data pipelines within the organization
- Create data architecture documents, standards, and principles and maintain knowledge on the data models
- Collaborate and coordinate with multiple teams/departments to identify the data domains and data gaps between current state systems and future goals
- Communicate clearly and effectively the data entities and their relationship within a business model
- Audit performance and advise any necessary infrastructure changes
- Develop key metrics for tests on data end create data quality rules
- Focus on scalability, availability, and data governance
Requirements
- Previous experience as a Big Data Team Lead or relevant is mandatory
- Proficiency in one of the following: Scala, Java or Python
- Proficiency with Hadoop ecosystem services such as MapReduce v2, HDFS, YARN, Hive, HBase
- Experience with any data lake table formats (e.q. Apache Hudi, Apache Iceberg, Delta Lake)
- Experience with building stream-processing systems using solutions such as Apache Kafka and Apache Spark streaming
- Experience in orchestration tools (e.q. Apache Airflow)
- Experience with integrating data from multiple heterogeneous sources and various formats (Parquet, CSV, XML, JSON, Avro)
- Experience with SQL databases and NoSQL databases, such as Elasticsearch and MongoDB
- Nice to have hands-on experience with Kubernetes
- Strong communication, leadership and teamwork skills
What You’ll Enjoy:
- Competitive salary based on skills and experience
- Medical insurance allowance
- Yearly performance bonus
- Working from home policy & Flexible hours
- Daily lunch allowance
- Udemy unlimited membership
- Sport/Gym monthly allowance
- Happy hours/events and many more team bonding activities
- Engage with new technologies and innovative products
- Internal hackathons and technical seminars