Location: Wroclaw, Poland(Travel 3 time in a week)
Experience: 5 to 12 years
Responsibility
-
Create and implement highly scalable and reliable data distribution solution using VQL, Python, Spark & open-source technologies, to deliver data to business components.
-
Work with Denodo, ADLS, Databricks, Kafka, data modelling, data replication, clustering, SQL Query patterns and indexing for handling for large data sets.
-
Demonstrate experience with Python and data access (Numpy, Scipy, panda etc.), machine learning (Tensorflow etc.), and AI libraries (Chat GTP etc.)
Requirements
-
4-5 years of hands-on experience in developing large scale applications using data virtualization and/or data streaming technologies.
-
Software engineer/developer focused on cloud-based data virtualization and data delivery technologies.
-
Denodo platform familiarity and SQL experience highly desirable.
-
Know-how to apply standards, methods, techniques and templates as defined by our SDLC including code control, code inspection, code deployment.
Additional Responsibilities
-
Design, plan and deliver solutions in a large scale enterprise environment. Working with solution architect & business analysts to define implementation design & coding of the assigned modules/responsibilities with highest quality (bug-free).
-
Determining technical approaches to be used, and defining the appropriate methodologies.
Must-Have Skills
-
Must be capable of working in a collaborative, multi-site environment to support rapid development and delivery of results and capabilities (i.e. AGILE SDLC).
-
Effectively communicating technical analyses, recommendations, status, and results to project management team. Produce secure and clean code that is stable, operational, consistent and well-performing.