Data Engineer – AI & Data Products
Generali Visualizza tutti gli annunci
- Milano
- Tempo indeterminato
- Full time
- Design, build, and evolve scalable cloud-native data pipelines supporting analytics, AI use cases, data products, and digital applications.
- Develop and optimize batch and streaming data ingestion frameworks, integrating structured, semi-structured, and unstructured data into modern data lakehouse architectures.
- Implement data quality, observability, and data reliability frameworks, ensuring trust, lineage, and compliance by design.
- Enable AI and ML workflows by:
- Designing and maintaining feature stores
- Automating data preprocessing pipelines
- Managing high-quality training and inference datasets
- Supporting GenAI and advanced AI use cases with optimized data access patterns
- Build backend data services and APIs to support AI-powered and digital solutions in production environments.
- Collaborate with AI Engineers and Data Scientists to industrialize models through MLOps best practices, CI/CD pipelines, and scalable deployment patterns.
- Work closely with Platform Engineering and SRE teams to ensure reliable deployment, monitoring, scalability, and cost optimization of data and AI workloads.
- Contribute to the evolution of the Cloud Data Platform architecture, optimizing storage, compute engines, and distributed processing frameworks.
- Ensure security, privacy, and regulatory compliance, especially in handling sensitive financial data.
- Provide technical guidance on emerging trends such as GenAI integration, real time analytics, data mesh architectures, and scalable AI platforms.
- Coordinate with internal and external partners in delivering complex data, AI, and digital initiatives.
- 3+ years of experience in Data Engineering or Machine Learning Engineering.
- Strong proficiency in Python (Pandas or similar, PySpark), SQL, or Scala for data manipulation and transformation.
- Experience with cloud platforms such as AWS, Azure, or Google Cloud and data storage technologies (Redshift, DynamoDB, BigQuery, etc.).
- Hands-on experience with distributed computing frameworks (Spark, Ray, etc.).
- Solid understanding of data modeling, data contracts, and governance frameworks.
- Experience with MLOps and AI lifecycle tools (e.g., MLflow, Kubeflow, Vertex AI, SageMaker).
- Experience with MLOps and AI lifecycle tools (e.g., MLflow, Kubeflow, Vertex AI, SageMaker).
- Bachelor’s degree in Computer Science, Engineering, or a related field.
- Master’s degree in Engineering, or a related field, or PhD in Computer Science.
- Experience in the financial sector.
- Familiarity with BI tools such as Power BI, Tableau, or Looker.
- Understanding of regulatory requirements in financial data governance.