Our client, a technology solutions company passionate about Customer tailored product development, is currently hiring a Data Tech Lead to join their team in Estonia!
From requirements definition and specification, software coding and development, to application support and maintenance; they understand and assist with the entire product lifecycle.
If you are passionate about building and managing data pipelines, you are always keeping in mind that the data flow is efficient and reliable and you are keen to lead a talented Data team – then this role might be for you!
Role:
Build and manage efficient and reliable (batch and real-time) data pipelines from disparate data sources (Kafka, and 3rd party tools)
Design, develop and launch data ingestion and storage systems with high availability and reliability that can scale (don’t worry here we like scalable and high available cloud solutions)
Drive the advancement of data infrastructure by developing and implementing underlying logic and structure for how data is set up, cleaned, and stored
Participate in & facilitate interviews of future Data Engineer teammates
Participate in the E2E lifecycle of a product delivery and guide the teams as required
Architect, launch and manage automated extraction and transformation processes
Build scalable data aggregation layer from streams and batches of data for data visualisation
Collaborate with development teams on design, architecture, and expansion of infrastructure
Collaborate on building the future development plans
Collaborate with other Tech Leads and Product Owners on prioritisation of technical and architectural features
Work as an SME Operational Data Stores, Data Warehouse, and Data Marts development
Guide the development design activities with input and data dependencies
Skills & Qualifications:
Strong communication skills
Experience leading a technical team
Ability to mentor other DE team members
Ability to clarify/translate customer requirements into Epics/Stories7+ years Data Warehouse/Data Lake Architecture and Development
Knowledge of programming languages like Scala, Java and Python
Experience in building architectures based on streaming data technologies for low-latency data processing (Apache Spark/Flink, Apache Kafka, Hadoop ecosystem)
Experience in data pipeline orchestration (Apache Airflow)
Confidence in using Git, CI\CD, and containerisation
Experience with Data Quality Tools, Monitoring and Alerting
Familiarity with core Kubernetes concepts
Experience with working with data coming from various sources (RDMS, APIs, files) in various formats (JSON, Avro, Parquet, Delta)
Experience participating in an Agile software development team, e.g. SCRUM
Knowledge in software development lifecycles/methodologies i.e. agile as strong presentation and collaboration skills and can communicate all aspects of the job requirements including the creation of formal documentation
Our working environment is fully English speaking, ability to communicate in English is required
Next steps: Apply with your CV and we will contact you soon!
Posted: 6th Sep 2022
Job contact:
Anthony Watters
Pentasia
Spain
InterGame is the only monthly magazine covering the international coin-op amusements and gaming industry.
Learn moreInterGaming is the leading magazine covering the international land-based casino and gaming industry.
Learn morePublished bimonthly, iNTERGAMINGi is the leading publication for the growing international online gaming industry.
Learn moreAn extensive annual buyers’ directory with more than 1,500 suppliers listed in the coin-op, casino and iGaming sectors.
Learn moreAll the latest innovations for children in the indoor play, edutainment, outdoor adventure and simulation sectors.
Learn moreCoin-op Community covers the latest news for the UK’s amusements, attractions and leisure sectors.
Learn more