Machine Learning Engineer (The AI Innovator)
Tech Stack
Job Description
Are you passionate about designing, building, and maintaining data pipelines that support robust data architectures and facilitate seamless data flow?
Do you excel in creating scalable solutions that empower data-driven decision-making?
If you’re ready to develop and optimize data systems that drive impactful analytics, our client has the perfect role for you.
We’re seeking a Data Engineer (aka The Data Pipeline Architect) to build and manage cloud-based data infrastructures that support analytical needs and operational efficiencies.As a Data Engineer at our client, you’ll collaborate with data scientists, analysts, and software engineers to construct data pipelines and storage solutions that are both efficient and secure.
Your role will be critical in ensuring data systems are optimized for performance, reliability, and scalability.Key Responsibilities: Design and Implement Scalable Data Pipelines: Develop and maintain data pipelines that support data ingestion, transformation, and integration using cloud technologies.
You’ll automate data workflows and ensure the seamless movement of data between various systems.
Manage and Optimize Data Storage Solutions: Architect and maintain data lakes and data warehouses using platforms like BigQuery, Redshift, Snowflake, or similar cloud-based solutions.
You’ll ensure data structures are built for performance and scalability.
Collaborate with Data Teams for Strategy Development: Work closely with data scientists, analysts, and business stakeholders to understand data requirements and align data solutions with business goals.
You’ll provide input on data models and storage strategies.
Ensure Data Quality and Reliability: Implement and manage processes for data validation, error handling, and consistency checks.
You’ll ensure the quality of data is maintained through robust testing and monitoring practices.
Develop and Automate ETL Processes: Build ETL (Extract, Transform, Load) workflows to handle complex data transformations.
You’ll automate data extraction and transformation to support efficient data integration and reporting.
Monitor and Maintain Data Infrastructure: Use monitoring tools to track the performance and reliability of data systems.
You’ll proactively identify and resolve potential issues to maintain system health and performance.
Optimize Data Processing and Resource Management: Implement strategies for efficient resource allocation and cost-effective data processing.
You’ll leverage parallel processing and cloud capabilities to enhance performance.