This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
Collaborate with Product teams ensuring that raw data is cleansed and transformed and useable by downstream consumers (ML Engineers, BI analytics). Assist and advise on the re-development and modernisation of end-to-end ETL pipelines and introduce new technologies where appropriate in a real-time streaming environment dealing with very large data volumes. Working with AWS cloud infrastructure (specifically SQS, SNS, Redshift, OpenSearch, Athena, Kinesis, AWS code pipeline) to develop, innovate and maintain data flowing through various queues and data warehouses. Working with a variety of data repository platforms (including SQL stores such as Oracle), as well as implementing data visualisation and network analysis (e.g GraphdB). Maintain and ‘productionise’ machine learning and AI models. Assist in the creation of next generation data ingestion platforms – sourcing data using Webscrapes, APIs, Email, and flat file (FTP) methods. Understanding conflict resolution methods and assist subject matter experts in the debugging on data ingestion and managing overall feed uptimes across a large set of data collectors. Create and maintain detailed documentation and functional design specifications including data flows and data conversion. Provide technical information to assist in the development of client facing product documentation. Adhere to change management protocols and version control. Present advanced technical designs with non-technical stakeholders.
Job Responsibility:
Collaborate with Product teams ensuring that raw data is cleansed and transformed and useable by downstream consumers (ML Engineers, BI analytics)
Assist and advise on the re-development and modernisation of end-to-end ETL pipelines and introduce new technologies where appropriate in a real-time streaming environment dealing with very large data volumes
Requirements:
Expect level PL-SQL
Working with AWS cloud infrastructure (specifically SQS, SNS, Redshift, OpenSearch, Athena, Kinesis, AWS code pipeline)
Working with a variety of data repository platforms (including SQL stores such as Oracle)
Implementing data visualisation and network analysis (e.g GraphdB)
Maintain and ‘productionise’ machine learning and AI models
Assist in the creation of next generation data ingestion platforms – sourcing data using Webscrapes, APIs, Email, and flat file (FTP) methods
Understanding conflict resolution methods
Assist subject matter experts in the debugging on data ingestion and managing overall feed uptimes across a large set of data collectors
Create and maintain detailed documentation and functional design specifications including data flows and data conversion
Provide technical information to assist in the development of client facing product documentation
Adhere to change management protocols and version control
Present advanced technical designs with non-technical stakeholders
Welcome to CrawlJobs.com – Your Global Job Discovery Platform
At CrawlJobs.com, we simplify finding your next career opportunity by bringing job listings directly to you from all corners of the web. Using cutting-edge AI and web-crawling technologies, we gather and curate job offers from various sources across the globe, ensuring you have access to the most up-to-date job listings in one place.
We use cookies to enhance your experience, analyze traffic, and serve personalized content. By clicking “Accept”, you agree to the use of cookies.