Lead Data Engineer
Ciklum
Remote, Spain
About the role
As a Lead Data Engineer, you'll become a part of a cross-functional development team engineering experiences of tomorrow.
Responsibilities
- Responsible for the building, deployment, and maintenance of mission critical analytics solutions that process data quickly at big data scales;
- Contributes design, code, configurations, and documentation for components that manage data ingestion, real time streaming, batch processing, data extraction, transformation, and loading across multiple data storages;
- Owns one or more key components of the infrastructure and works to continually improve it, identifying gaps and improving the platform’s quality, robustness, maintainability, and speed;
- Cross-trains other team members on technologies being developed, while also continuously learning new technologies from other team members;
- Interacts with engineering teams and ensures that solutions meet customer requirements in terms of functionality, performance, availability, scalability, and reliability
- Performs development, QA, and dev-ops roles as needed to ensure total end to end responsibility of solutions;
- Works directly with business analysts and data scientists to understand and support their use-cases;
- Contribute in the Unit’s activities and community building, participate in conferences, provide excellence in exercise and best practices;
- Help in sales activities, customer meetings and digital services.
Requirements
- 5+ years of experience coding in SQL, Java, Python, Scala, with solid CS fundamentals including data structure and algorithm design;
- 3+ years contributing to production deployments of large backend data processing and analysis systems as a team lead;
- 2+ years of hands-on implementation experience working with a combination of the following technologies: Hadoop, Map Reduce, Pig, Hive, Impala, Spark, Kafka, Storm, SQL and NoSQL data warehouses such as Hbase and Cassandra;
- 3+ years of experience in cloud data platforms (AWS, Azure, GCP);
- Experience with Kubernetes;
- Knowledge of SQL and MPP databases (e.g. Vertica, Netezza, Greenplum, Aster Data);
- Knowledge of professional software engineering best practices for the full software;
- Knowledge of Data Warehousing, design, implementation and optimization;
- Knowledge of Data Quality testing, automation and results visualization;
- Knowledge of BI reports and dashboards design and implementation (PowerBI, Tableau);
- Knowledge of development life cycle, including coding standards, code reviews, source control management, build processes, testing, and operations;
- Experience participating in an Agile software development team, e.g. SCRUM;
- Experience designing, documenting, and defending designs for key components in large distributed computing systems;
- A consistent track record of delivering exceptionally high-quality software on large, complex, cross-functional projects;
- Demonstrated ability to learn new technologies quickly and independently;
- Ability to handle multiple competing priorities in a fast-paced environment;
- Undergraduate degree in Computer Science or Engineering from a top CS program required. Masters preferred;
- Experience with supporting data scientists and complex statistical usecases highly desirable.
Desirable
- Understanding of cloud infrastructure design and implementation;
- Experience in data science and machine learning;
- Experience in backend development and deployment;
- Experience in CI/CD configuration;
- Good knowledge of data analysis in enterprises;
- Experience with Databricks, Snowflake.
Apply Now
Don't forget to mention EuroTechJobs when applying.