Job Description

Our client is an information technology company with corporate headquarters in Limassol, Cyprus. They are a world leader in the research and development world of state-of-the-art methods for ingesting data from heterogenous sources and adapting it to bespoke and intuitive solutions for civilian protection and they are currently looking to hire a skilled Big Data Team Lead to join their team.

The ideal candidate will be instrumental in accelerating and scaling their data pipelines and data lakes.

Reference #: GC004373

Job Responsibilities

  • Establish, lead, manage and mentor the big data team
  • Own the development of an inhouse Data Lake for storing structured and unstructured data
  • Research, design, and develop appropriate algorithms for Big Data collection, processing, and analysis
  • Define how data will be streamed, stored, consumed, integrated by different data systems
  • Identify relevant Big Data tools required to support new and existing product capabilities
  • Collaborate closely with the product team to define the requirements and milestones that relate to Big Data features
  • Closely interact with the Data Scientists in providing feature-ed datasets
  • Design, create, deploy, manage data pipelines within the organization
  • Create data architecture documents, standards, and principles and maintain knowledge on the data models
  • Collaborate and coordinate with multiple teams/departments to identify the data domains and data gaps between current state systems and future goals
  • Communicate clearly and effectively the data entities and their relationship within a business model
  • Audit performance and advise any necessary infrastructure changes
  • Develop key metrics for tests on data end create data quality rules
  • Focus on  scalability, availability, and data governance


  • Previous experience as a Big Data Team Lead or relevant is mandatory
  • Proficiency in one of the following: Scala, Java or Python
  • Proficiency with Hadoop ecosystem services such as MapReduce v2, HDFS, YARN, Hive, HBase
  • Experience with any data lake table formats (e.q. Apache Hudi, Apache Iceberg, Delta Lake)
  • Experience with building stream-processing systems using solutions such as Apache Kafka and Apache Spark streaming
  • Experience in orchestration tools (e.q. Apache Airflow)
  • Experience with integrating data from multiple heterogeneous sources and various formats (Parquet, CSV, XML, JSON, Avro)
  • Experience with SQL databases and NoSQL databases, such as Elasticsearch and MongoDB
  • Nice to have hands-on experience with Kubernetes
  • Strong communication, leadership and teamwork skills

What You’ll Enjoy:

  • Competitive salary based on skills and experience
  • Medical insurance allowance
  • Yearly performance bonus
  • Working from home policy & Flexible hours
  • Daily lunch allowance
  • Udemy unlimited membership
  • Sport/Gym monthly allowance
  • Happy hours/events and many more team bonding activities
  • Engage with new technologies and innovative products
  • Internal hackathons and technical seminars