Hiring bonus: 10% of monthly salary

Next job

Senior Data Engineer in Alcor

Posted more than 30 days ago

549 views

3 applications   0 responses

Alcor
Alcor
IT Services and Consulting
0
0 reviews
Miroslava
Miroslava
recruiter
More than 5 years
Upper Intermediate
Remote work
We are looking for a Senior Data Engineer (GCP) to join the Ukrainian team of a Canadian product company on a full-time remote basis. The company Questrade is one of Canada's leading online brokerages. Its desktop version of Questrade Edge is our favorite trading platform out of the Canadian brokers we tested. The broker offers a great balance of user-friendly technology, markets to trade, and account services.

Questrade opened its doors in 1999 and is still one of Canada's fastest-growing online brokerages. It offers a great balance of platforms, a range of markets to trade, and services including multiple account types.

The first product was a platform that allows people to do trading, then the company added a couple of new business lines, and the company is on the way to the full digital banking solution.

The main office is in Toronto.


Right now we are looking for Senior Data Engineer (GCP) on a full-time remote basis.


Strong Must-Have Skills:

● At least 5 or more years of experience working in the data engineering field;

● Very Strong GCP data engineering experience working with BigQuery, Dataflow, Cloud Data Fusion, Dataprep, Data Catalog, Cloud Composer and

CloudSQL Functions;

● Experience in architecting, developing software, or Internet-scale production-grade big data solutions in virtualized environments;

● Experience in working with/on data warehouses, including data warehouse technical architectures, infrastructure components, ETL/ELT and

reporting/analytic tools, environments, and data structures;

● Design, document, and develop complex MS Power Platform (PowerBI, Power Apps, etc.) solutions;

● Work with UX/UI designers to construct wireframes, mock-ups, and in an agile development process creates Power BI Embedded reports;

● Work closely with our application solution architects, data scientists, and data engineering teams to develop BI solutions, and enhance and/or define new data integrations and/or modeling efforts. Monitor, troubleshoot, and resolve issues for the deployed reporting solutions;

● Ability to work, communicate effectively and influence stakeholders on external engineering teams, product development teams, Business stakeholders and external partners;

● Experience working on multiple projects simultaneously while troubleshooting technical issues and working with cross-functional stakeholders;

● Creating and maintaining optimal data architecture pipelines;

● Good Knowledge in SQL language;

● Good Knowledge in Message Broker systems (e.g., Kafka, PubSub);

● Good Knowledge of Python language;

● Develop and maintain code and documentation for ETL and other data

integration projects and procedures;

● Collaborate with the team to decide on which tools and strategies to use

within specific data integration scenarios;

● Demonstrable track record of dealing well with ambiguity, prioritizing needs,

and delivering results in a dynamic environment.


Good to have skills:

● SSIS, SSRS;

● Knowledge of NoSQL DB.


Nice to have skills:

● Knowledge of the Financial industry (Banking/Insurance/Mortgages);

● Experience building applications using SparkML and TensorFlow;

● GCP Data Engineering Certification preferred.


Let’s discuss more details 😊

Specializations: Data Science
Keywords: Kafka, gcp, Big Data & BI, pubsub
More than 5 years
Upper Intermediate
Remote work
Want to get related jobs?
New job openings in your Telegram
Subscribe
We use cookies
accept