Capco
Data & AI Warsaw Tech Summit 2026: Machine Learning Engineer – From Models to Production
Capco at Data & AI Warsaw Tech Summit 2026About Capco Poland
At Capco Poland, we’re not just another consultancy - we’re the spark behind digital transformation in the financial world.
As a global technology and management consultancy focused on financial services, we partner with leading banks, fintechs, and financial institutions to design and deliver next-generation data platforms, AI solutions, and digital ecosystems.
From data strategy and modern data platforms to AI-driven decision systems and GenAI innovation, our teams help clients unlock the true value of their data.
Our secret?
A culture that’s fast, flexible, and fiercely entrepreneurial. We move quickly, think creatively, and empower our people to push the boundaries of what technology can achieve.
At Capco Poland, we are proud to be:
• Technology partners for leading banks, payments providers, and financial institutions
• Builders of modern data platforms and AI-powered systems
• Champions of innovation across cloud, data engineering, machine learning, and GenAI
• A community of engineers, architects, and consultants passionate about solving complex problems
At this year’s Data & AI Warsaw Tech Summit, Capco will share how financial institutions can move from experimentation to production-grade AI and scalable data ecosystems.
Our experts will explore how organizations can:
• Build AI-native architectures on modern cloud platforms
• Scale machine learning and generative AI solutions across enterprise environments
• Transform fragmented data into high-value data products
• Embed AI into real business workflows and decision-making systems
Topic: From Data to Meaning: Educating AI in Banking with Ontologies: Lessons from FIBO and Conversational Banking
Time: 11:50-12:10 CET
Intro:
Many AI solutions still fall short when it comes to understanding and reasoning about complex financial concepts. The real challenge is about how financial knowledge is represented and shared with machines. Why does AI still misunderstand basic banking terms despite having access to vast amounts of data?
How can AI truly understand financial concepts? Using the Financial Industry Business Ontology (FIBO) as an example of structured domain knowledge, we will discuss how formal, machine-readable definitions can provide the contextual foundation AI needs. By analysing selected conversational banking scenarios and example solutions, we will invite participants to reflect together on what the right semantic layer for AI in banking should look like.
Join us to discover why the next leap in AI for banking isn’t just about more data or better models, but about building a structured understanding of financial meaning.
We are looking for a Machine Learning Engineer to design, build, and deploy scalable machine learning solutions. In this role, you will work closely with data scientists, data engineers, and product teams to bring ML models into production and ensure their performance, reliability, and scalability.
Key ResponsibilitiesDesign, develop, and deploy machine learning models into production
Build and maintain scalable ML pipelines and workflows
Collaborate with data scientists to operationalize models (MLOps)
Optimize model performance, scalability, and latency
Monitor, evaluate, and retrain models in production
Work with large datasets and feature engineering processes
Implement best practices for versioning, testing, and deployment of ML models
Integrate ML solutions into existing systems and applications
Document models, pipelines, and processes
Proven experience as a Machine Learning Engineer or similar role (X+ years)
Strong programming skills in Python (or similar language)
Experience with ML frameworks (e.g., TensorFlow, PyTorch, scikit-learn)
Solid understanding of machine learning algorithms and statistics
Experience with data processing tools (e.g., Pandas, Spark)
Familiarity with MLOps practices and tools (e.g., MLflow, Kubeflow, Airflow)
Experience with cloud platforms (AWS, Azure, or GCP)
Knowledge of APIs and microservices architecture
Strong problem-solving and communication skills
Experience with deep learning and NLP or computer vision
Familiarity with Docker and Kubernetes
Experience with CI/CD pipelines for ML
Knowledge of data engineering concepts and tools
Experience with real-time or streaming data systems
- Screening call with the Recruiter
- Hiring Manager Technical Interview
- Feedback
- Offer
We offer a flexible collaboration model based on a B2B contract with the opportunity to work on diverse projects.

