EU Horizon 2020
Horizon 2020
HomeNewsCase StudiesPeopleKey Prior PublicationsPublications

FUN2MODEL - News

September 2021 - *Current Vacancy* We are looking for an exceptional and motivated Senior Research Associate to play a key role in this ERC funded FUN2MODEL project. You would be responsible for carrying out research with an emphasis on probabilistic reasoning and inference, including cognitive aspects and automated decision making.

If this interests you, and you hold a PhD in computer science, mathematics or related discipline and post-qualification research experience, we'd love to hear from you. Please contact Marta (marta.kwiatkowska@cs.ox.ac.uk) to learn more or apply here.

September 2021 - Award: Marta who has been awarded the Van Wijngaarden Award 2021 for Computer Science in recognition of her numerous and highly significant contributions to preventing software faults. The five-yearly award also recognised mathematician, Susan Murphy, for her work in improving decision making in health. The Van Wijngaarden Award was established by CWI, and is named after former CWI director, Aad van Wijngaarden. Congratulations Marta!
September 2021 - Welcome: We are delighted that Senior Research Fellow, Gethin Norman, has joined the FUN2MODEL project. Gethin specialises in modelling the formalisms, theories, logics and algorithms that underpin the PRISM model checker.
June 2021 - Research Opportunities: We are always seeking exceptional researchers interested in formulating theories, models and algorithms for probabilistic verification and synthesis to enable robust AI.

If this sounds like you, and you hold a PhD in computer science, mathematics or related discipline and post-qualification research experience, please contact Marta (marta.kwiatkowska@cs.ox.ac.uk) or Anita (anita.hancox@ox.ac.ox.uk) to learn more about opportunities on the FUN2MODEL project.

May 2021 - Paper: accepted at UAI 2021! Here we propose a framework to provide safety certification for given control policies, and synthesize control policies that improve the certification bounds.
May 2021 - Event: Marta was awarded the prestigious 2019 BCS Lovelace Medal for her research in probabilistic and quantitative verification and so led the 2020 Lovelace Lecture “Probabilistic model checking for the data-rich world.” A recording can be found here.

The BCS Lovelace Lecture was sponsored by The Ada Lovelace Institute

April 2021 - Papers: three papers accepted at IJCAI 2021!

Here we build on abduction-based explanations for machine learning and develop a method for computing local explanations for neural network models in natural language processing (NLP).

Here we demonstrate provable guarantees on the robustness of decision rules, paving the way towards provably causally robust decision-making systems.

Here we introduce the first method for verifying the time-unbounded safety of neural networks controlling dynamical systems.

April 2021 - Welcome: We warmly welcome DPhil Student Emanuele La Malfa, to the FUN2MODEL project. Emanuele's work will focus on robustness and explainability for natural language processing (NLP).
March 2021 - Software Release: PRISM 4.7 is now available, including support for POMDPs, improved accuracy reporting and more.
March 2021 - Welcome: We warmly welcome researcher, Rui Yan, to the FUN2MODEL project. Rui's work will focus on probabilistic verification and synthesis, including Bayesian and psychological games.
December 2020 - News: Marta becomes a Fellow of the European Laboratory for Learning and Intelligent Systems (ELLIS). ELLIS Fellows advance science, provide strategic advice and leadership, and act as ambassadors of ELLIS.
November 2020 - Event: Professor David Parker, co-PI on FUN2MODEL, gives a keynote talk at iFM 2020 titled "Verification with Stochastic Games: Advances and Challenges".
October 2020 - Event: Marta is a panelist at the Royal Society's Briefing for Making Europe a Leader in AI: in conversation with Venki Ramakrishnan, Antoine Petit and Martin Stratmann.
October 2020 - Welcome: We are delighted that graduate student, Elias Benussi, has joined the FUN2MODEL project. Elias will focus on fairness in AI.
October 2020 - Paper: accepted at FORMATS 2020! Here we propose MOSAIC, an algorithm for measuring the safety of deep reinforcement learning controllers in stochastic settings.
October 2020 - Paper: accepted at EMNLP 2020! Here we focus on robustness of text classification against word substitutions.
September 2020 - Event: Marta gives a webinar on 'Safety and robustness for deep learning with provable guarantees' in the ICE-TCS Reykjavik University series.
September 2020 - Event: Marta gives a keynote speech at ASE 2020 on 'Safety and robustness for deep learning with provable guarantees'.
September 2020 - Event: Marta gives a plenary talk at DNA26 on 'Probabilistic verification and synthesis for reliable molecular circuit designs'.
September 2020 - Event: Marta gives a keynote speech at KR 2020 on 'Probabilistic model checking for strategic equilibria-based decision making'.
August 2020 - Welcome: We are delighted to have researcher, Gabriel Santos, join the FUN2MODEL project. Gabriel will focus on strategic reasoning and game-theoretic techniques in AI.
August 2020 - Paper: accepted at QEST 2020! Here we propose multi-coalitional verification techniques for concurrent stochastic games.
July 2020 - Event: Marta gives a Webinar on "When to trust a self-driving car" at The National Academy of Sciences, India (NASI) - Delhi Chapter attended by over 440+ participants. Please see here to watch a recording.
June 2020 - Event: Marta has been invited to sit on the Global Partnership on Artificial Intelligence (GPAI) Working Group on Responsible AI, nominated by the European Commission.

GPAI is an international and multi-stakeholder initiative to guide the responsible development and use of artificial intelligence consistent with human rights, fundamental freedoms, and shared democratic values, as reflected in the OECD Recommendation on AI.

Please see here for more details.

May 2020 - Paper: accepted at UAI 2020! Here we show how to compute worst-case adversarial guarantees for Bayesian Neural Networks (BNNs).
May 2020 - Paper: by Clare Lyle accepted at ICML 2020! Here we consider the problem of learning abstractions that generalize in block MDPs, families of environments with a shared latent state space, and dynamics structure over that latent space, but varying observations.
April 2020 - Welcome: We are delighted to welcome researcher, Andrea Patane, to the project. Andrea brings with him expertise in safety verification of Bayesian models and the role played by uncertainty in adversarial prediction settings. He will focus on data-driven perception modelling and cognitive reasoning, for integration within agent-based models.
April 2020 - Paper: accepted at CAV 2020!

Here we present a major new release of the PRISM-games model checker, featuring multiple significant advances in its support for verification and strategy synthesis of stochastic games.

Download the tool and case studies here.

March 2020 - Event: Marta becomes a member of the Royal Society’s Working Group on Digital Technology and the Planet which aims to help the UK play a leading role in data-enabled innovation and the adoption of digital technologies to tackle climate change.
February 2020 - Paper: by Min Wu accepted at CVPR 2020! Here we consider the robustness of deep neural networks on videos.
January 2020 - Paper: accepted at AISTATS 2020! Here we show how to compute worst-case adversarial guarantees for classification with Gaussian processes.
January 2020 - Event: Marta gives a keynote speech at ERTS 2020 titled 'Safety verification for deep neural networks with provable guarantees'.
January 2020 - Event: Marta is an invited speaker at VMCAI 2020.
January 2020 - Software Release: PRISM-games 3.0 is now available, providing concurrent stochastic games, equilibria, real-time models and many new examples. More information here.
October 2019 - Welcome: We warmly welcome graduate student, Emanuele La Malfa, who joins the project as an Associate Member and will formally join in April 2021. Emanuele is working on robustness and explainability of AI, with specific focus on natural language processing (NLP) models.
October 2019 - Welcome: We are delighted that graduate student, Benjie Wang, has joined the FUN2MODEL project. Benjie is particularly interested in causal modelling as a means to enhance the robustness and explainability of deep learning.
October 2019 - Welcome: We are excited to appoint postdoctoral researcher, Luca Laurenti, to the project. Luca will work on on developing probabilistic verification and synthesis methods for deep learning, with a particular focus on Baysian neural networks.
1st October 2019 - The Fun Begins! FUN2MODEL kicks off and work begins as we aim to make major advances in the quest towards provably robust and beneficial AI.
March 2019 - Announcement: We are delighted to announce that Professor Marta Kwiatkowska has been awarded a highly competitive European Research Council Advanced Investigators Grant for a new five-year project FUN2MODEL.

Our objectives are to develop novel probabilistic verification and synthesis techniques to guarantee safety, robustness and fairness for complex decisions based on machine learning, formulate a comprehensive, compositional game-based modelling framework for reasoning about systems of autonomous agents and their interactions, and evaluate the techniques on a variety of case studies.

This is the second ERC Advanced Grant awarded to Marta Kwiatkowska: from 2010 until 2016 she held the grant VERIWARE. Please see the press release here and here for more details.