Senior Data Engineer (Databricks Expert) (Remote)
28600-31900 PLN miesięcznie (B2B)
DCG sp. z o.o.
Czym będziesz się zajmować?
RESPONSIBILITIES:
- Designing new solutions and coming up with initiatives for improvements to existing solutions within data platforms - both as part of orders coming from the business (functional changes) and from technology (architectural changes)
- Development of data platforms and ETL/ELT processes: technical support and active participation in the development of data platforms
- Work on building and optimizing ETL/ELT processes that are responsible for processing large data sets
- Implement processes to ensure optimal data processing, using data engineering best practices
- Standardize and streamline technical processes: implementing and optimizing code, test and documentation management standards
- Selecting and configuring tools and development environments that support data engineering processes to maintain code quality and facilitate code scaling
- Ensure standards compliance and code review: responsible for applying existing platform development standards, initiating new guidelines where improvements are needed, and monitoring the quality of delivered solutions by conducting regular code-reviews
- Work directly with technology as a Data Engineer and Data Analyst to maintain a high level of technology sophistication
- Act as a mentor to the team
Kogo poszukujemy?
Senior Data Engineer (Databricks Expert)
REQUIREMENTS:
- Minimum 5 years of experience in designing and building Business Intelligence, ETL/ELT, Data Warehouse, Data Lake, Data Lakehouse, Big Data and OLAP-class solutions, with Databricks used as the core platform in production environments
- Practical knowledge of various relational and non-relational database engines, applied in the context of building and operating Databricks-based data platforms
- Strong proficiency in SQL and Python
- Familiarity with data engineering and orchestration tools, particularly Spark/Databricks (including structured streaming mechanisms, DLT, etc.), Hadoop/CDP, Azure/Fabric Data Factory, Apache Flink, Apache Kafka, Apache Airflow, dbt, Debezium, and more
- Understanding of data governance, data quality, and batch/streaming data processing challenges
- Knowledge of architectural patterns in data, including Data Mesh, Data Vault, Dimensional Modeling, Medallion Architecture, and Lambda/Kappa Architectures
- Proficiency in using git repositories (Bitbucket, GitHub, GitLab)
- Experience with cloud data services on Azure and/or AWS, with Databricks used as a key component of the data platform
- Flexibility, self-reliance, and efficiency, with a strong sense of responsibility for assigned tasks
Czego wymagamy?
Znajomości:
Mile widziane:
Języki:
- Polski
- Angielski
Jakie warunki i benefity otrzymasz?
- 170-190 PLN godzinowo (B2B)
- B2B - Elastyczne godziny pracy (100%)
- Praca zdalna: W całości
- Budżet szkoleniowy
- Pakiet medyczny, Pakiet sportowy
Gdzie będziesz pracował?
Zdalnie
Kim jesteśmy?
DCG to przestrzeń, w której spotykają się potrzeby biznesu i ambicje ludzi. Znamy wartość dobrze dopasowanej współpracy, dlatego pomagamy kandydatom znaleźć środowisko, w którym będą mogli rozwinąć skrzydła, a firmom - zbudować zespoły, które naprawdę działają. Pracujemy blisko ludzi i organizacji, uważnie słuchając i reagując na to, co dla nich ważne. Dzięki temu wspólnie tworzymy trwałe i wartościowe relacje, które procentują na lata.