Why join GFT? You will work with and learn from top IT experts. You will join a crew of experienced engineers: 60% of our employees are senior level. Interested in the cloud? You will enjoy our full support in developing your skills: training programs, certifications and our internal community of experts. We have strong partnerships with top cloud providers: Google, Amazon and Microsoft - we are number one in Poland in numbers of GCP certificates. Apart from GCP, you can also develop in AWS or
Why join GFT?
You will work with and learn from top IT experts. You will join a crew of experienced engineers: 60% of our employees are senior level.
Interested in the cloud? You will enjoy our full support in developing your skills: training programs, certifications and our internal community of experts. We have strong partnerships with top cloud providers: Google, Amazon and Microsoft - we are number one in Poland in numbers of GCP certificates. Apart from GCP, you can also develop in AWS or Azure.
We are focused on development and knowledge sharing. Internal expert communities provide a comfortable environment where you can develop your skillset in areas such as blockchain, Big Data, cloud computing or artificial intelligence.
You will work in a stable company (32 years on the market) in demanding and challenging projects for the biggest financial institutions in the world.
We offer you:
- Working in a highly experienced and dedicated team
- Competitive salary and extra benefit package that can be tailored to your personal needs (private medical coverage, sport & recreation package, lunch subsidy, life insurance, etc.)
- Permanent contract or B2B contract
- On-line training and certifications fit for career path
- Free on-line foreign languages lessons
- Regular social events
- Access to e-learning platform
- Ergonomic and functional working space with 2 monitors
Your profile:
- Mid or senior level of your previous experience as a Data Engineer (2-4 years)
- Openness to work in a hybrid model (2 days from the office per week)
- Strong SQL and Python and Spark/PySpark
- Experience with Data Modelling
- Understanding of key concepts around Data Warehousing, Data Lakes and Data Lakehouses
- Experience with at least one of commercial clouds (GCP, AWS, Azure)
- Experience with Cloud Data engineering toolset
Nice to have:
- Experience with Business Intelligence tools
- Snowflake
- DataBricks
Why join GFT?
You will work with and learn from top IT experts. You will join a crew of experienced engineers: 60% of our employees are senior level.
Interested in the cloud? You will enjoy our full support in developing your skills: training programs, certifications and our internal community of experts. We have strong partnerships with top cloud providers: Google, Amazon and Microsoft - we are number one in Poland in numbers of GCP certificates. Apart from GCP, you can also develop in AWS or Azure.
We are focused on development and knowledge sharing. Internal expert communities provide a comfortable environment where you can develop your skillset in areas such as blockchain, Big Data, cloud computing or artificial intelligence.
You will work in a stable company (32 years on the market) in demanding and challenging projects for the biggest financial institutions in the world.
We offer you:
- Working in a highly experiencedand dedicated team
- Competitive salary and extra benefit package that can be tailored to your personal needs (private medical coverage, sport & recreation package, lunch subsidy, life insurance, etc.)
- Permanent contract or B2B contract
- On-line training and certifications fit for career path
- Free on-line foreign languages lessons
- Regular social events
- Access to e-learning platform
- Ergonomic and functional working space with 2 monitors
,[Design, build, test and deploy Cloud and on-premise data models and transformations in Cloud Native or dedicated toolset, ETL development, Optimize data views for specific visualization use cases making use of schema design partitions, indexes, down-sampling, archiving, etc. to manage trade-offs such as performance and flexibility, Review and refine, interpret and implement business and technical requirements, Ensure you are part of the on-going productivity and priorities by refining User Stories, Epics and Backlogs in Jira, Onboarding new data sources, design, build, test and deploy Cloud data ingest, pipelines, warehouses and data models/products]
Requirements: Dataproc for Spark, Dataflow for Apache Beam, Composer/Airflow, GPC, Python, Spark, PySpark, SQL, Data modelling, Data Lake, data lakehouses, CloudSQL for Postgres, GCS, ETL, GKE, AWS, Azure, Snowflake, Databricks
Additionally: Home office, Knowledge sharing, Life insurance, Sport subscription, Training budget, Private healthcare, International projects, Free coffee, Playroom, Free snacks, Free beverages, In-house trainings, In-house hack days, Modern office.