---
_id: '14830'
abstract:
- lang: eng
  text: We study the problem of learning controllers for discrete-time non-linear
    stochastic dynamical systems with formal reach-avoid guarantees. This work presents
    the first method for providing formal reach-avoid guarantees, which combine and
    generalize stability and safety guarantees, with a tolerable probability threshold
    p in [0,1] over the infinite time horizon. Our method leverages advances in machine
    learning literature and it represents formal certificates as neural networks.
    In particular, we learn a certificate in the form of a reach-avoid supermartingale
    (RASM), a novel notion that we introduce in this work. Our RASMs provide reachability
    and avoidance guarantees by imposing constraints on what can be viewed as a stochastic
    extension of level sets of Lyapunov functions for deterministic systems. Our approach
    solves several important problems -- it can be used to learn a control policy
    from scratch, to verify a reach-avoid specification for a fixed control policy,
    or to fine-tune a pre-trained policy if it does not satisfy the reach-avoid specification.
    We validate our approach on 3 stochastic non-linear reinforcement learning tasks.
acknowledgement: This work was supported in part by the ERC-2020-AdG 101020093, ERC
  CoG 863818 (FoRM-SMArt) and the European Union’s Horizon 2020 research and innovation
  programme under the Marie Skłodowska-Curie Grant Agreement No. 665385.
article_processing_charge: No
arxiv: 1
author:
- first_name: Dorde
  full_name: Zikelic, Dorde
  id: 294AA7A6-F248-11E8-B48F-1D18A9856A87
  last_name: Zikelic
  orcid: 0000-0002-4681-1699
- first_name: Mathias
  full_name: Lechner, Mathias
  id: 3DC22916-F248-11E8-B48F-1D18A9856A87
  last_name: Lechner
- first_name: Thomas A
  full_name: Henzinger, Thomas A
  id: 40876CD8-F248-11E8-B48F-1D18A9856A87
  last_name: Henzinger
  orcid: 0000-0002-2985-7724
- first_name: Krishnendu
  full_name: Chatterjee, Krishnendu
  id: 2E5DCA20-F248-11E8-B48F-1D18A9856A87
  last_name: Chatterjee
  orcid: 0000-0002-4561-241X
citation:
  ama: 'Zikelic D, Lechner M, Henzinger TA, Chatterjee K. Learning control policies
    for stochastic systems with reach-avoid guarantees. In: <i>Proceedings of the
    37th AAAI Conference on Artificial Intelligence</i>. Vol 37. Association for the
    Advancement of Artificial Intelligence; 2023:11926-11935. doi:<a href="https://doi.org/10.1609/aaai.v37i10.26407">10.1609/aaai.v37i10.26407</a>'
  apa: 'Zikelic, D., Lechner, M., Henzinger, T. A., &#38; Chatterjee, K. (2023). Learning
    control policies for stochastic systems with reach-avoid guarantees. In <i>Proceedings
    of the 37th AAAI Conference on Artificial Intelligence</i> (Vol. 37, pp. 11926–11935).
    Washington, DC, United States: Association for the Advancement of Artificial Intelligence.
    <a href="https://doi.org/10.1609/aaai.v37i10.26407">https://doi.org/10.1609/aaai.v37i10.26407</a>'
  chicago: Zikelic, Dorde, Mathias Lechner, Thomas A Henzinger, and Krishnendu Chatterjee.
    “Learning Control Policies for Stochastic Systems with Reach-Avoid Guarantees.”
    In <i>Proceedings of the 37th AAAI Conference on Artificial Intelligence</i>,
    37:11926–35. Association for the Advancement of Artificial Intelligence, 2023.
    <a href="https://doi.org/10.1609/aaai.v37i10.26407">https://doi.org/10.1609/aaai.v37i10.26407</a>.
  ieee: D. Zikelic, M. Lechner, T. A. Henzinger, and K. Chatterjee, “Learning control
    policies for stochastic systems with reach-avoid guarantees,” in <i>Proceedings
    of the 37th AAAI Conference on Artificial Intelligence</i>, Washington, DC, United
    States, 2023, vol. 37, no. 10, pp. 11926–11935.
  ista: 'Zikelic D, Lechner M, Henzinger TA, Chatterjee K. 2023. Learning control
    policies for stochastic systems with reach-avoid guarantees. Proceedings of the
    37th AAAI Conference on Artificial Intelligence. AAAI: Conference on Artificial
    Intelligence vol. 37, 11926–11935.'
  mla: Zikelic, Dorde, et al. “Learning Control Policies for Stochastic Systems with
    Reach-Avoid Guarantees.” <i>Proceedings of the 37th AAAI Conference on Artificial
    Intelligence</i>, vol. 37, no. 10, Association for the Advancement of Artificial
    Intelligence, 2023, pp. 11926–35, doi:<a href="https://doi.org/10.1609/aaai.v37i10.26407">10.1609/aaai.v37i10.26407</a>.
  short: D. Zikelic, M. Lechner, T.A. Henzinger, K. Chatterjee, in:, Proceedings of
    the 37th AAAI Conference on Artificial Intelligence, Association for the Advancement
    of Artificial Intelligence, 2023, pp. 11926–11935.
conference:
  end_date: 2023-02-14
  location: Washington, DC, United States
  name: 'AAAI: Conference on Artificial Intelligence'
  start_date: 2023-02-07
date_created: 2024-01-18T07:44:31Z
date_published: 2023-06-26T00:00:00Z
date_updated: 2025-07-14T09:10:02Z
day: '26'
department:
- _id: ToHe
- _id: KrCh
doi: 10.1609/aaai.v37i10.26407
ec_funded: 1
external_id:
  arxiv:
  - '2210.05308'
intvolume: '        37'
issue: '10'
keyword:
- General Medicine
language:
- iso: eng
month: '06'
oa_version: Preprint
page: 11926-11935
project:
- _id: 62781420-2b32-11ec-9570-8d9b63373d4d
  call_identifier: H2020
  grant_number: '101020093'
  name: Vigilant Algorithmic Monitoring of Software
- _id: 0599E47C-7A3F-11EA-A408-12923DDC885E
  call_identifier: H2020
  grant_number: '863818'
  name: 'Formal Methods for Stochastic Models: Algorithms and Applications'
- _id: 2564DBCA-B435-11E9-9278-68D0E5697425
  call_identifier: H2020
  grant_number: '665385'
  name: International IST Doctoral Program
publication: Proceedings of the 37th AAAI Conference on Artificial Intelligence
publication_identifier:
  eissn:
  - 2374-3468
  issn:
  - 2159-5399
publication_status: published
publisher: Association for the Advancement of Artificial Intelligence
quality_controlled: '1'
related_material:
  record:
  - id: '14600'
    relation: earlier_version
    status: public
status: public
title: Learning control policies for stochastic systems with reach-avoid guarantees
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 37
year: '2023'
...
---
_id: '12510'
abstract:
- lang: eng
  text: "We introduce a new statistical verification algorithm that formally quantifies
    the behavioral robustness of any time-continuous process formulated as a continuous-depth
    model. Our algorithm solves a set of global optimization (Go) problems over a
    given time horizon to construct a tight enclosure (Tube) of the set of all process
    executions starting from a ball of initial states. We call our algorithm GoTube.
    Through its construction, GoTube ensures that the bounding tube is conservative
    up to a desired probability and up to a desired tightness.\r\n GoTube is implemented
    in JAX and optimized to scale to complex continuous-depth neural network models.
    Compared to advanced reachability analysis tools for time-continuous neural networks,
    GoTube does not accumulate overapproximation errors between time steps and avoids
    the infamous wrapping effect inherent in symbolic techniques. We show that GoTube
    substantially outperforms state-of-the-art verification tools in terms of the
    size of the initial ball, speed, time-horizon, task completion, and scalability
    on a large set of experiments.\r\n GoTube is stable and sets the state-of-the-art
    in terms of its ability to scale to time horizons well beyond what has been previously
    possible."
acknowledgement: SG is funded by the Austrian Science Fund (FWF) project number W1255-N23.
  ML and TH are supported in part by FWF under grant Z211-N23 (Wittgenstein Award)
  and the ERC-2020-AdG 101020093. SS is supported by NSF awards DCL-2040599, CCF-1918225,
  and CPS-1446832. RH and DR are partially supported by Boeing. RG is partially supported
  by Horizon-2020 ECSEL Project grant No. 783163 (iDev40).
article_processing_charge: No
article_type: original
arxiv: 1
author:
- first_name: Sophie A.
  full_name: Gruenbacher, Sophie A.
  last_name: Gruenbacher
- first_name: Mathias
  full_name: Lechner, Mathias
  id: 3DC22916-F248-11E8-B48F-1D18A9856A87
  last_name: Lechner
- first_name: Ramin
  full_name: Hasani, Ramin
  last_name: Hasani
- first_name: Daniela
  full_name: Rus, Daniela
  last_name: Rus
- first_name: Thomas A
  full_name: Henzinger, Thomas A
  id: 40876CD8-F248-11E8-B48F-1D18A9856A87
  last_name: Henzinger
  orcid: 0000-0002-2985-7724
- first_name: Scott A.
  full_name: Smolka, Scott A.
  last_name: Smolka
- first_name: Radu
  full_name: Grosu, Radu
  last_name: Grosu
citation:
  ama: 'Gruenbacher SA, Lechner M, Hasani R, et al. GoTube: Scalable statistical verification
    of continuous-depth models. <i>Proceedings of the AAAI Conference on Artificial
    Intelligence</i>. 2022;36(6):6755-6764. doi:<a href="https://doi.org/10.1609/aaai.v36i6.20631">10.1609/aaai.v36i6.20631</a>'
  apa: 'Gruenbacher, S. A., Lechner, M., Hasani, R., Rus, D., Henzinger, T. A., Smolka,
    S. A., &#38; Grosu, R. (2022). GoTube: Scalable statistical verification of continuous-depth
    models. <i>Proceedings of the AAAI Conference on Artificial Intelligence</i>.
    Association for the Advancement of Artificial Intelligence. <a href="https://doi.org/10.1609/aaai.v36i6.20631">https://doi.org/10.1609/aaai.v36i6.20631</a>'
  chicago: 'Gruenbacher, Sophie A., Mathias Lechner, Ramin Hasani, Daniela Rus, Thomas
    A Henzinger, Scott A. Smolka, and Radu Grosu. “GoTube: Scalable Statistical Verification
    of Continuous-Depth Models.” <i>Proceedings of the AAAI Conference on Artificial
    Intelligence</i>. Association for the Advancement of Artificial Intelligence,
    2022. <a href="https://doi.org/10.1609/aaai.v36i6.20631">https://doi.org/10.1609/aaai.v36i6.20631</a>.'
  ieee: 'S. A. Gruenbacher <i>et al.</i>, “GoTube: Scalable statistical verification
    of continuous-depth models,” <i>Proceedings of the AAAI Conference on Artificial
    Intelligence</i>, vol. 36, no. 6. Association for the Advancement of Artificial
    Intelligence, pp. 6755–6764, 2022.'
  ista: 'Gruenbacher SA, Lechner M, Hasani R, Rus D, Henzinger TA, Smolka SA, Grosu
    R. 2022. GoTube: Scalable statistical verification of continuous-depth models.
    Proceedings of the AAAI Conference on Artificial Intelligence. 36(6), 6755–6764.'
  mla: 'Gruenbacher, Sophie A., et al. “GoTube: Scalable Statistical Verification
    of Continuous-Depth Models.” <i>Proceedings of the AAAI Conference on Artificial
    Intelligence</i>, vol. 36, no. 6, Association for the Advancement of Artificial
    Intelligence, 2022, pp. 6755–64, doi:<a href="https://doi.org/10.1609/aaai.v36i6.20631">10.1609/aaai.v36i6.20631</a>.'
  short: S.A. Gruenbacher, M. Lechner, R. Hasani, D. Rus, T.A. Henzinger, S.A. Smolka,
    R. Grosu, Proceedings of the AAAI Conference on Artificial Intelligence 36 (2022)
    6755–6764.
date_created: 2023-02-05T17:27:42Z
date_published: 2022-06-28T00:00:00Z
date_updated: 2023-09-26T10:46:59Z
day: '28'
department:
- _id: ToHe
doi: 10.1609/aaai.v36i6.20631
ec_funded: 1
external_id:
  arxiv:
  - '2107.08467'
intvolume: '        36'
issue: '6'
keyword:
- General Medicine
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: https://arxiv.org/abs/2107.08467
month: '06'
oa: 1
oa_version: Preprint
page: 6755-6764
project:
- _id: 25F42A32-B435-11E9-9278-68D0E5697425
  call_identifier: FWF
  grant_number: Z211
  name: The Wittgenstein Prize
- _id: 62781420-2b32-11ec-9570-8d9b63373d4d
  call_identifier: H2020
  grant_number: '101020093'
  name: Vigilant Algorithmic Monitoring of Software
publication: Proceedings of the AAAI Conference on Artificial Intelligence
publication_identifier:
  eissn:
  - 2374-3468
  isbn:
  - '978577358350'
  issn:
  - 2159-5399
publication_status: published
publisher: Association for the Advancement of Artificial Intelligence
quality_controlled: '1'
scopus_import: '1'
status: public
title: 'GoTube: Scalable statistical verification of continuous-depth models'
type: journal_article
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 36
year: '2022'
...
---
_id: '12511'
abstract:
- lang: eng
  text: "We consider the problem of formally verifying almost-sure (a.s.) asymptotic
    stability in discrete-time nonlinear stochastic control systems. While verifying
    stability in deterministic control systems is extensively studied in the literature,
    verifying stability in stochastic control systems is an open problem. The few
    existing works on this topic either consider only specialized forms of stochasticity
    or make restrictive assumptions on the system, rendering them inapplicable to
    learning algorithms with neural network policies. \r\n In this work, we present
    an approach for general nonlinear stochastic control problems with two novel aspects:
    (a) instead of classical stochastic extensions of Lyapunov functions, we use ranking
    supermartingales (RSMs) to certify a.s. asymptotic stability, and (b) we present
    a method for learning neural network RSMs. \r\n We prove that our approach guarantees
    a.s. asymptotic stability of the system and\r\n provides the first method to obtain
    bounds on the stabilization time, which stochastic Lyapunov functions do not.\r\n
    Finally, we validate our approach experimentally on a set of nonlinear stochastic
    reinforcement learning environments with neural network policies."
acknowledgement: "This work was supported in part by the ERC-2020-AdG 101020093, ERC
  CoG 863818 (FoRM-SMArt) and the European Union’s Horizon 2020 research and innovation
  programme\r\nunder the Marie Skłodowska-Curie Grant Agreement No. 665385."
article_processing_charge: No
article_type: original
arxiv: 1
author:
- first_name: Mathias
  full_name: Lechner, Mathias
  id: 3DC22916-F248-11E8-B48F-1D18A9856A87
  last_name: Lechner
- first_name: Dorde
  full_name: Zikelic, Dorde
  id: 294AA7A6-F248-11E8-B48F-1D18A9856A87
  last_name: Zikelic
  orcid: 0000-0002-4681-1699
- first_name: Krishnendu
  full_name: Chatterjee, Krishnendu
  id: 2E5DCA20-F248-11E8-B48F-1D18A9856A87
  last_name: Chatterjee
  orcid: 0000-0002-4561-241X
- first_name: Thomas A
  full_name: Henzinger, Thomas A
  id: 40876CD8-F248-11E8-B48F-1D18A9856A87
  last_name: Henzinger
  orcid: 0000-0002-2985-7724
citation:
  ama: Lechner M, Zikelic D, Chatterjee K, Henzinger TA. Stability verification in
    stochastic control systems via neural network supermartingales. <i>Proceedings
    of the AAAI Conference on Artificial Intelligence</i>. 2022;36(7):7326-7336. doi:<a
    href="https://doi.org/10.1609/aaai.v36i7.20695">10.1609/aaai.v36i7.20695</a>
  apa: Lechner, M., Zikelic, D., Chatterjee, K., &#38; Henzinger, T. A. (2022). Stability
    verification in stochastic control systems via neural network supermartingales.
    <i>Proceedings of the AAAI Conference on Artificial Intelligence</i>. Association
    for the Advancement of Artificial Intelligence. <a href="https://doi.org/10.1609/aaai.v36i7.20695">https://doi.org/10.1609/aaai.v36i7.20695</a>
  chicago: Lechner, Mathias, Dorde Zikelic, Krishnendu Chatterjee, and Thomas A Henzinger.
    “Stability Verification in Stochastic Control Systems via Neural Network Supermartingales.”
    <i>Proceedings of the AAAI Conference on Artificial Intelligence</i>. Association
    for the Advancement of Artificial Intelligence, 2022. <a href="https://doi.org/10.1609/aaai.v36i7.20695">https://doi.org/10.1609/aaai.v36i7.20695</a>.
  ieee: M. Lechner, D. Zikelic, K. Chatterjee, and T. A. Henzinger, “Stability verification
    in stochastic control systems via neural network supermartingales,” <i>Proceedings
    of the AAAI Conference on Artificial Intelligence</i>, vol. 36, no. 7. Association
    for the Advancement of Artificial Intelligence, pp. 7326–7336, 2022.
  ista: Lechner M, Zikelic D, Chatterjee K, Henzinger TA. 2022. Stability verification
    in stochastic control systems via neural network supermartingales. Proceedings
    of the AAAI Conference on Artificial Intelligence. 36(7), 7326–7336.
  mla: Lechner, Mathias, et al. “Stability Verification in Stochastic Control Systems
    via Neural Network Supermartingales.” <i>Proceedings of the AAAI Conference on
    Artificial Intelligence</i>, vol. 36, no. 7, Association for the Advancement of
    Artificial Intelligence, 2022, pp. 7326–36, doi:<a href="https://doi.org/10.1609/aaai.v36i7.20695">10.1609/aaai.v36i7.20695</a>.
  short: M. Lechner, D. Zikelic, K. Chatterjee, T.A. Henzinger, Proceedings of the
    AAAI Conference on Artificial Intelligence 36 (2022) 7326–7336.
date_created: 2023-02-05T17:29:50Z
date_published: 2022-06-28T00:00:00Z
date_updated: 2025-07-14T09:09:58Z
day: '28'
department:
- _id: ToHe
- _id: KrCh
doi: 10.1609/aaai.v36i7.20695
ec_funded: 1
external_id:
  arxiv:
  - '2112.09495'
intvolume: '        36'
issue: '7'
keyword:
- General Medicine
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: https://arxiv.org/abs/2112.09495
month: '06'
oa: 1
oa_version: Preprint
page: 7326-7336
project:
- _id: 62781420-2b32-11ec-9570-8d9b63373d4d
  call_identifier: H2020
  grant_number: '101020093'
  name: Vigilant Algorithmic Monitoring of Software
- _id: 0599E47C-7A3F-11EA-A408-12923DDC885E
  call_identifier: H2020
  grant_number: '863818'
  name: 'Formal Methods for Stochastic Models: Algorithms and Applications'
- _id: 2564DBCA-B435-11E9-9278-68D0E5697425
  call_identifier: H2020
  grant_number: '665385'
  name: International IST Doctoral Program
publication: Proceedings of the AAAI Conference on Artificial Intelligence
publication_identifier:
  eissn:
  - 2374-3468
  isbn:
  - '9781577358350'
  issn:
  - 2159-5399
publication_status: published
publisher: Association for the Advancement of Artificial Intelligence
quality_controlled: '1'
related_material:
  record:
  - id: '14539'
    relation: dissertation_contains
    status: public
scopus_import: '1'
status: public
title: Stability verification in stochastic control systems via neural network supermartingales
type: journal_article
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 36
year: '2022'
...
---
_id: '12568'
abstract:
- lang: eng
  text: We treat the problem of risk-aware control for stochastic shortest path (SSP)
    on Markov decision processes (MDP). Typically, expectation is considered for SSP,
    which however is oblivious to the incurred risk. We present an alternative view,
    instead optimizing conditional value-at-risk (CVaR), an established risk measure.
    We treat both Markov chains as well as MDP and introduce, through novel insights,
    two algorithms, based on linear programming and value iteration, respectively.
    Both algorithms offer precise and provably correct solutions. Evaluation of our
    prototype implementation shows that risk-aware control is feasible on several
    moderately sized models.
article_processing_charge: No
arxiv: 1
author:
- first_name: Tobias
  full_name: Meggendorfer, Tobias
  id: b21b0c15-30a2-11eb-80dc-f13ca25802e1
  last_name: Meggendorfer
  orcid: 0000-0002-1712-2165
citation:
  ama: 'Meggendorfer T. Risk-aware stochastic shortest path. In: <i>Proceedings of
    the 36th AAAI Conference on Artificial Intelligence, AAAI 2022</i>. Vol 36. Association
    for the Advancement of Artificial Intelligence; 2022:9858-9867. doi:<a href="https://doi.org/10.1609/aaai.v36i9.21222">10.1609/aaai.v36i9.21222</a>'
  apa: 'Meggendorfer, T. (2022). Risk-aware stochastic shortest path. In <i>Proceedings
    of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022</i> (Vol. 36,
    pp. 9858–9867). Virtual: Association for the Advancement of Artificial Intelligence.
    <a href="https://doi.org/10.1609/aaai.v36i9.21222">https://doi.org/10.1609/aaai.v36i9.21222</a>'
  chicago: Meggendorfer, Tobias. “Risk-Aware Stochastic Shortest Path.” In <i>Proceedings
    of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022</i>, 36:9858–67.
    Association for the Advancement of Artificial Intelligence, 2022. <a href="https://doi.org/10.1609/aaai.v36i9.21222">https://doi.org/10.1609/aaai.v36i9.21222</a>.
  ieee: T. Meggendorfer, “Risk-aware stochastic shortest path,” in <i>Proceedings
    of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022</i>, Virtual,
    2022, vol. 36, no. 9, pp. 9858–9867.
  ista: Meggendorfer T. 2022. Risk-aware stochastic shortest path. Proceedings of
    the 36th AAAI Conference on Artificial Intelligence, AAAI 2022. Conference on
    Artificial Intelligence vol. 36, 9858–9867.
  mla: Meggendorfer, Tobias. “Risk-Aware Stochastic Shortest Path.” <i>Proceedings
    of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022</i>, vol. 36,
    no. 9, Association for the Advancement of Artificial Intelligence, 2022, pp. 9858–67,
    doi:<a href="https://doi.org/10.1609/aaai.v36i9.21222">10.1609/aaai.v36i9.21222</a>.
  short: T. Meggendorfer, in:, Proceedings of the 36th AAAI Conference on Artificial
    Intelligence, AAAI 2022, Association for the Advancement of Artificial Intelligence,
    2022, pp. 9858–9867.
conference:
  end_date: 2022-03-01
  location: Virtual
  name: Conference on Artificial Intelligence
  start_date: 2022-02-22
date_created: 2023-02-19T23:00:56Z
date_published: 2022-06-28T00:00:00Z
date_updated: 2023-02-20T07:19:12Z
day: '28'
department:
- _id: KrCh
doi: 10.1609/aaai.v36i9.21222
external_id:
  arxiv:
  - '2203.01640'
intvolume: '        36'
issue: '9'
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: ' https://doi.org/10.48550/arXiv.2203.01640'
month: '06'
oa: 1
oa_version: Preprint
page: 9858-9867
publication: Proceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI
  2022
publication_identifier:
  eissn:
  - 2374-3468
  isbn:
  - '1577358767'
publication_status: published
publisher: Association for the Advancement of Artificial Intelligence
quality_controlled: '1'
scopus_import: '1'
status: public
title: Risk-aware stochastic shortest path
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 36
year: '2022'
...
---
_id: '10665'
abstract:
- lang: eng
  text: "Formal verification of neural networks is an active topic of research, and
    recent advances have significantly increased the size of the networks that verification
    tools can handle. However, most methods are designed for verification of an idealized
    model of the actual network which works over real arithmetic and ignores rounding
    imprecisions. This idealization is in stark contrast to network quantization,
    which is a technique that trades numerical precision for computational efficiency
    and is, therefore, often applied in practice. Neglecting rounding errors of such
    low-bit quantized neural networks has been shown to lead to wrong conclusions
    about the network’s correctness. Thus, the desired approach for verifying quantized
    neural networks would be one that takes these rounding errors\r\ninto account.
    In this paper, we show that verifying the bitexact implementation of quantized
    neural networks with bitvector specifications is PSPACE-hard, even though verifying
    idealized real-valued networks and satisfiability of bit-vector specifications
    alone are each in NP. Furthermore, we explore several practical heuristics toward
    closing the complexity gap between idealized and bit-exact verification. In particular,
    we propose three techniques for making SMT-based verification of quantized neural
    networks more scalable. Our experiments demonstrate that our proposed methods
    allow a speedup of up to three orders of magnitude over existing approaches."
acknowledgement: "This research was supported in part by the Austrian Science Fund
  (FWF) under grant Z211-N23 (Wittgenstein\r\nAward), ERC CoG 863818 (FoRM-SMArt),
  and the European Union’s Horizon 2020 research and innovation programme under the
  Marie Skłodowska-Curie Grant Agreement No. 665385.\r\n"
alternative_title:
- Technical Tracks
article_processing_charge: No
arxiv: 1
author:
- first_name: Thomas A
  full_name: Henzinger, Thomas A
  id: 40876CD8-F248-11E8-B48F-1D18A9856A87
  last_name: Henzinger
  orcid: 0000-0002-2985-7724
- first_name: Mathias
  full_name: Lechner, Mathias
  id: 3DC22916-F248-11E8-B48F-1D18A9856A87
  last_name: Lechner
- first_name: Dorde
  full_name: Zikelic, Dorde
  id: 294AA7A6-F248-11E8-B48F-1D18A9856A87
  last_name: Zikelic
  orcid: 0000-0002-4681-1699
citation:
  ama: 'Henzinger TA, Lechner M, Zikelic D. Scalable verification of quantized neural
    networks. In: <i>Proceedings of the AAAI Conference on Artificial Intelligence</i>.
    Vol 35. AAAI Press; 2021:3787-3795.'
  apa: 'Henzinger, T. A., Lechner, M., &#38; Zikelic, D. (2021). Scalable verification
    of quantized neural networks. In <i>Proceedings of the AAAI Conference on Artificial
    Intelligence</i> (Vol. 35, pp. 3787–3795). Virtual: AAAI Press.'
  chicago: Henzinger, Thomas A, Mathias Lechner, and Dorde Zikelic. “Scalable Verification
    of Quantized Neural Networks.” In <i>Proceedings of the AAAI Conference on Artificial
    Intelligence</i>, 35:3787–95. AAAI Press, 2021.
  ieee: T. A. Henzinger, M. Lechner, and D. Zikelic, “Scalable verification of quantized
    neural networks,” in <i>Proceedings of the AAAI Conference on Artificial Intelligence</i>,
    Virtual, 2021, vol. 35, no. 5A, pp. 3787–3795.
  ista: 'Henzinger TA, Lechner M, Zikelic D. 2021. Scalable verification of quantized
    neural networks. Proceedings of the AAAI Conference on Artificial Intelligence.
    AAAI: Association for the Advancement of Artificial Intelligence, Technical Tracks,
    vol. 35, 3787–3795.'
  mla: Henzinger, Thomas A., et al. “Scalable Verification of Quantized Neural Networks.”
    <i>Proceedings of the AAAI Conference on Artificial Intelligence</i>, vol. 35,
    no. 5A, AAAI Press, 2021, pp. 3787–95.
  short: T.A. Henzinger, M. Lechner, D. Zikelic, in:, Proceedings of the AAAI Conference
    on Artificial Intelligence, AAAI Press, 2021, pp. 3787–3795.
conference:
  end_date: 2021-02-09
  location: Virtual
  name: 'AAAI: Association for the Advancement of Artificial Intelligence'
  start_date: 2021-02-02
date_created: 2022-01-25T15:15:02Z
date_published: 2021-05-28T00:00:00Z
date_updated: 2025-07-14T09:10:11Z
day: '28'
ddc:
- '000'
department:
- _id: GradSch
- _id: ToHe
ec_funded: 1
external_id:
  arxiv:
  - '2012.08185'
file:
- access_level: open_access
  checksum: 2bc8155b2526a70fba5b7301bc89dbd1
  content_type: application/pdf
  creator: mlechner
  date_created: 2022-01-26T07:41:16Z
  date_updated: 2022-01-26T07:41:16Z
  file_id: '10684'
  file_name: 16496-Article Text-19990-1-2-20210518 (1).pdf
  file_size: 137235
  relation: main_file
  success: 1
file_date_updated: 2022-01-26T07:41:16Z
has_accepted_license: '1'
intvolume: '        35'
issue: 5A
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: https://ojs.aaai.org/index.php/AAAI/article/view/16496
month: '05'
oa: 1
oa_version: Published Version
page: 3787-3795
project:
- _id: 2564DBCA-B435-11E9-9278-68D0E5697425
  call_identifier: H2020
  grant_number: '665385'
  name: International IST Doctoral Program
- _id: 25F42A32-B435-11E9-9278-68D0E5697425
  call_identifier: FWF
  grant_number: Z211
  name: The Wittgenstein Prize
- _id: 0599E47C-7A3F-11EA-A408-12923DDC885E
  call_identifier: H2020
  grant_number: '863818'
  name: 'Formal Methods for Stochastic Models: Algorithms and Applications'
publication: Proceedings of the AAAI Conference on Artificial Intelligence
publication_identifier:
  eissn:
  - 2374-3468
  isbn:
  - 978-1-57735-866-4
  issn:
  - 2159-5399
publication_status: published
publisher: AAAI Press
quality_controlled: '1'
related_material:
  record:
  - id: '11362'
    relation: dissertation_contains
    status: public
scopus_import: '1'
status: public
title: Scalable verification of quantized neural networks
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 35
year: '2021'
...
---
_id: '10669'
abstract:
- lang: eng
  text: "We show that Neural ODEs, an emerging class of timecontinuous neural networks,
    can be verified by solving a set of global-optimization problems. For this purpose,
    we introduce Stochastic Lagrangian Reachability (SLR), an\r\nabstraction-based
    technique for constructing a tight Reachtube (an over-approximation of the set
    of reachable states\r\nover a given time-horizon), and provide stochastic guarantees
    in the form of confidence intervals for the Reachtube bounds. SLR inherently avoids
    the infamous wrapping effect (accumulation of over-approximation errors) by performing
    local optimization steps to expand safe regions instead of repeatedly forward-propagating
    them as is done by deterministic reachability methods. To enable fast local optimizations,
    we introduce a novel forward-mode adjoint sensitivity method to compute gradients
    without the need for backpropagation. Finally, we establish asymptotic and non-asymptotic
    convergence rates for SLR."
acknowledgement: "The authors would like to thank the reviewers for their insightful
  comments. RH and RG were partially supported by\r\nHorizon-2020 ECSEL Project grant
  No. 783163 (iDev40). RH was partially supported by Boeing. ML was supported\r\nin
  part by the Austrian Science Fund (FWF) under grant Z211-N23 (Wittgenstein Award).
  SG was funded by FWF\r\nproject W1255-N23. JC was partially supported by NAWA Polish
  Returns grant PPN/PPO/2018/1/00029. SS was supported by NSF awards DCL-2040599,
  CCF-1918225, and CPS-1446832.\r\n"
alternative_title:
- Technical Tracks
article_processing_charge: No
arxiv: 1
author:
- first_name: Sophie
  full_name: Grunbacher, Sophie
  last_name: Grunbacher
- first_name: Ramin
  full_name: Hasani, Ramin
  last_name: Hasani
- first_name: Mathias
  full_name: Lechner, Mathias
  id: 3DC22916-F248-11E8-B48F-1D18A9856A87
  last_name: Lechner
- first_name: Jacek
  full_name: Cyranka, Jacek
  last_name: Cyranka
- first_name: Scott A
  full_name: Smolka, Scott A
  last_name: Smolka
- first_name: Radu
  full_name: Grosu, Radu
  last_name: Grosu
citation:
  ama: 'Grunbacher S, Hasani R, Lechner M, Cyranka J, Smolka SA, Grosu R. On the verification
    of neural ODEs with stochastic guarantees. In: <i>Proceedings of the AAAI Conference
    on Artificial Intelligence</i>. Vol 35. AAAI Press; 2021:11525-11535.'
  apa: 'Grunbacher, S., Hasani, R., Lechner, M., Cyranka, J., Smolka, S. A., &#38;
    Grosu, R. (2021). On the verification of neural ODEs with stochastic guarantees.
    In <i>Proceedings of the AAAI Conference on Artificial Intelligence</i> (Vol.
    35, pp. 11525–11535). Virtual: AAAI Press.'
  chicago: Grunbacher, Sophie, Ramin Hasani, Mathias Lechner, Jacek Cyranka, Scott
    A Smolka, and Radu Grosu. “On the Verification of Neural ODEs with Stochastic
    Guarantees.” In <i>Proceedings of the AAAI Conference on Artificial Intelligence</i>,
    35:11525–35. AAAI Press, 2021.
  ieee: S. Grunbacher, R. Hasani, M. Lechner, J. Cyranka, S. A. Smolka, and R. Grosu,
    “On the verification of neural ODEs with stochastic guarantees,” in <i>Proceedings
    of the AAAI Conference on Artificial Intelligence</i>, Virtual, 2021, vol. 35,
    no. 13, pp. 11525–11535.
  ista: 'Grunbacher S, Hasani R, Lechner M, Cyranka J, Smolka SA, Grosu R. 2021. On
    the verification of neural ODEs with stochastic guarantees. Proceedings of the
    AAAI Conference on Artificial Intelligence. AAAI: Association for the Advancement
    of Artificial Intelligence, Technical Tracks, vol. 35, 11525–11535.'
  mla: Grunbacher, Sophie, et al. “On the Verification of Neural ODEs with Stochastic
    Guarantees.” <i>Proceedings of the AAAI Conference on Artificial Intelligence</i>,
    vol. 35, no. 13, AAAI Press, 2021, pp. 11525–35.
  short: S. Grunbacher, R. Hasani, M. Lechner, J. Cyranka, S.A. Smolka, R. Grosu,
    in:, Proceedings of the AAAI Conference on Artificial Intelligence, AAAI Press,
    2021, pp. 11525–11535.
conference:
  end_date: 2021-02-09
  location: Virtual
  name: 'AAAI: Association for the Advancement of Artificial Intelligence'
  start_date: 2021-02-02
date_created: 2022-01-25T15:47:20Z
date_published: 2021-05-28T00:00:00Z
date_updated: 2022-05-24T06:33:14Z
day: '28'
ddc:
- '000'
department:
- _id: GradSch
- _id: ToHe
external_id:
  arxiv:
  - '2012.08863'
file:
- access_level: open_access
  checksum: 468d07041e282a1d46ffdae92f709630
  content_type: application/pdf
  creator: mlechner
  date_created: 2022-01-26T07:38:08Z
  date_updated: 2022-01-26T07:38:08Z
  file_id: '10680'
  file_name: 17372-Article Text-20866-1-2-20210518.pdf
  file_size: 286906
  relation: main_file
  success: 1
file_date_updated: 2022-01-26T07:38:08Z
has_accepted_license: '1'
intvolume: '        35'
issue: '13'
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: https://ojs.aaai.org/index.php/AAAI/article/view/17372
month: '05'
oa: 1
oa_version: Published Version
page: 11525-11535
project:
- _id: 25F42A32-B435-11E9-9278-68D0E5697425
  call_identifier: FWF
  grant_number: Z211
  name: The Wittgenstein Prize
publication: Proceedings of the AAAI Conference on Artificial Intelligence
publication_identifier:
  eissn:
  - 2374-3468
  isbn:
  - 978-1-57735-866-4
  issn:
  - 2159-5399
publication_status: published
publisher: AAAI Press
quality_controlled: '1'
status: public
title: On the verification of neural ODEs with stochastic guarantees
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 35
year: '2021'
...
---
_id: '10671'
abstract:
- lang: eng
  text: We introduce a new class of time-continuous recurrent neural network models.
    Instead of declaring a learning system’s dynamics by implicit nonlinearities,
    we construct networks of linear first-order dynamical systems modulated via nonlinear
    interlinked gates. The resulting models represent dynamical systems with varying
    (i.e., liquid) time-constants coupled to their hidden state, with outputs being
    computed by numerical differential equation solvers. These neural networks exhibit
    stable and bounded behavior, yield superior expressivity within the family of
    neural ordinary differential equations, and give rise to improved performance
    on time-series prediction tasks. To demonstrate these properties, we first take
    a theoretical approach to find bounds over their dynamics, and compute their expressive
    power by the trajectory length measure in a latent trajectory space. We then conduct
    a series of time-series prediction experiments to manifest the approximation capability
    of Liquid Time-Constant Networks (LTCs) compared to classical and modern RNNs.
acknowledgement: "R.H. and D.R. are partially supported by Boeing. R.H. and R.G. were
  partially supported by the Horizon-2020 ECSEL\r\nProject grant No. 783163 (iDev40).
  M.L. was supported in part by the Austrian Science Fund (FWF) under grant Z211-N23
  (Wittgenstein Award). A.A. is supported by the National Science Foundation (NSF)
  Graduate Research Fellowship Program. This research work is partially drawn from
  the PhD dissertation of R.H."
alternative_title:
- Technical Tracks
article_processing_charge: No
arxiv: 1
author:
- first_name: Ramin
  full_name: Hasani, Ramin
  last_name: Hasani
- first_name: Mathias
  full_name: Lechner, Mathias
  id: 3DC22916-F248-11E8-B48F-1D18A9856A87
  last_name: Lechner
- first_name: Alexander
  full_name: Amini, Alexander
  last_name: Amini
- first_name: Daniela
  full_name: Rus, Daniela
  last_name: Rus
- first_name: Radu
  full_name: Grosu, Radu
  last_name: Grosu
citation:
  ama: 'Hasani R, Lechner M, Amini A, Rus D, Grosu R. Liquid time-constant networks.
    In: <i>Proceedings of the AAAI Conference on Artificial Intelligence</i>. Vol
    35. AAAI Press; 2021:7657-7666.'
  apa: 'Hasani, R., Lechner, M., Amini, A., Rus, D., &#38; Grosu, R. (2021). Liquid
    time-constant networks. In <i>Proceedings of the AAAI Conference on Artificial
    Intelligence</i> (Vol. 35, pp. 7657–7666). Virtual: AAAI Press.'
  chicago: Hasani, Ramin, Mathias Lechner, Alexander Amini, Daniela Rus, and Radu
    Grosu. “Liquid Time-Constant Networks.” In <i>Proceedings of the AAAI Conference
    on Artificial Intelligence</i>, 35:7657–66. AAAI Press, 2021.
  ieee: R. Hasani, M. Lechner, A. Amini, D. Rus, and R. Grosu, “Liquid time-constant
    networks,” in <i>Proceedings of the AAAI Conference on Artificial Intelligence</i>,
    Virtual, 2021, vol. 35, no. 9, pp. 7657–7666.
  ista: 'Hasani R, Lechner M, Amini A, Rus D, Grosu R. 2021. Liquid time-constant
    networks. Proceedings of the AAAI Conference on Artificial Intelligence. AAAI:
    Association for the Advancement of Artificial Intelligence, Technical Tracks,
    vol. 35, 7657–7666.'
  mla: Hasani, Ramin, et al. “Liquid Time-Constant Networks.” <i>Proceedings of the
    AAAI Conference on Artificial Intelligence</i>, vol. 35, no. 9, AAAI Press, 2021,
    pp. 7657–66.
  short: R. Hasani, M. Lechner, A. Amini, D. Rus, R. Grosu, in:, Proceedings of the
    AAAI Conference on Artificial Intelligence, AAAI Press, 2021, pp. 7657–7666.
conference:
  end_date: 2021-02-09
  location: Virtual
  name: 'AAAI: Association for the Advancement of Artificial Intelligence'
  start_date: 2021-02-02
date_created: 2022-01-25T15:48:36Z
date_published: 2021-05-28T00:00:00Z
date_updated: 2022-05-24T06:36:54Z
day: '28'
ddc:
- '000'
department:
- _id: GradSch
- _id: ToHe
external_id:
  arxiv:
  - '2006.04439'
file:
- access_level: open_access
  checksum: 0f06995fba06dbcfa7ed965fc66027ff
  content_type: application/pdf
  creator: mlechner
  date_created: 2022-01-26T07:36:03Z
  date_updated: 2022-01-26T07:36:03Z
  file_id: '10678'
  file_name: 16936-Article Text-20430-1-2-20210518 (1).pdf
  file_size: 4302669
  relation: main_file
  success: 1
file_date_updated: 2022-01-26T07:36:03Z
has_accepted_license: '1'
intvolume: '        35'
issue: '9'
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: https://ojs.aaai.org/index.php/AAAI/article/view/16936
month: '05'
oa: 1
oa_version: Published Version
page: 7657-7666
project:
- _id: 25F42A32-B435-11E9-9278-68D0E5697425
  call_identifier: FWF
  grant_number: Z211
  name: The Wittgenstein Prize
publication: Proceedings of the AAAI Conference on Artificial Intelligence
publication_identifier:
  eissn:
  - 2374-3468
  isbn:
  - 978-1-57735-866-4
  issn:
  - 2159-5399
publication_status: published
publisher: AAAI Press
quality_controlled: '1'
status: public
title: Liquid time-constant networks
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 35
year: '2021'
...
---
_id: '11436'
abstract:
- lang: eng
  text: Asynchronous distributed algorithms are a popular way to reduce synchronization
    costs in large-scale optimization, and in particular for neural network training.
    However, for nonsmooth and nonconvex objectives, few convergence guarantees exist
    beyond cases where closed-form proximal operator solutions are available. As training
    most popular deep neural networks corresponds to optimizing nonsmooth and nonconvex
    objectives, there is a pressing need for such convergence guarantees. In this
    paper, we analyze for the first time the convergence of stochastic asynchronous
    optimization for this general class of objectives. In particular, we focus on
    stochastic subgradient methods allowing for block variable partitioning, where
    the shared model is asynchronously updated by concurrent processes. To this end,
    we use a probabilistic model which captures key features of real asynchronous
    scheduling between concurrent processes. Under this model, we establish convergence
    with probability one to an invariant set for stochastic subgradient methods with
    momentum. From a practical perspective, one issue with the family of algorithms
    that we consider is that they are not efficiently supported by machine learning
    frameworks, which mostly focus on distributed data-parallel strategies. To address
    this, we propose a new implementation strategy for shared-memory based training
    of deep neural networks for a partitioned but shared model in single- and multi-GPU
    settings. Based on this implementation, we achieve on average1.2x speed-up in
    comparison to state-of-the-art training methods for popular image classification
    tasks, without compromising accuracy.
acknowledgement: Vyacheslav Kungurtsev was supported by the OP VVV project CZ.02.1.01/0.0/0.0/16
  019/0000765 “Research Center for Informatics. Bapi Chatterjee was supported by the
  European Union’s Horizon 2020 research and innovation programme under the Marie
  Sklodowska-Curie grant agreement No. 754411 (ISTPlus). Dan Alistarh has received
  funding from the European Research Council (ERC) under the European Union’s Horizon
  2020 research and innovation programme (grant agreement No 805223 ScaleML).
article_processing_charge: No
arxiv: 1
author:
- first_name: Vyacheslav
  full_name: Kungurtsev, Vyacheslav
  last_name: Kungurtsev
- first_name: Malcolm
  full_name: Egan, Malcolm
  last_name: Egan
- first_name: Bapi
  full_name: Chatterjee, Bapi
  id: 3C41A08A-F248-11E8-B48F-1D18A9856A87
  last_name: Chatterjee
- first_name: Dan-Adrian
  full_name: Alistarh, Dan-Adrian
  id: 4A899BFC-F248-11E8-B48F-1D18A9856A87
  last_name: Alistarh
  orcid: 0000-0003-3650-940X
citation:
  ama: 'Kungurtsev V, Egan M, Chatterjee B, Alistarh D-A. Asynchronous optimization
    methods for efficient training of deep neural networks with guarantees. In: <i>35th
    AAAI Conference on Artificial Intelligence, AAAI 2021</i>. Vol 35. AAAI Press;
    2021:8209-8216.'
  apa: 'Kungurtsev, V., Egan, M., Chatterjee, B., &#38; Alistarh, D.-A. (2021). Asynchronous
    optimization methods for efficient training of deep neural networks with guarantees.
    In <i>35th AAAI Conference on Artificial Intelligence, AAAI 2021</i> (Vol. 35,
    pp. 8209–8216). Virtual, Online: AAAI Press.'
  chicago: Kungurtsev, Vyacheslav, Malcolm Egan, Bapi Chatterjee, and Dan-Adrian Alistarh.
    “Asynchronous Optimization Methods for Efficient Training of Deep Neural Networks
    with Guarantees.” In <i>35th AAAI Conference on Artificial Intelligence, AAAI
    2021</i>, 35:8209–16. AAAI Press, 2021.
  ieee: V. Kungurtsev, M. Egan, B. Chatterjee, and D.-A. Alistarh, “Asynchronous optimization
    methods for efficient training of deep neural networks with guarantees,” in <i>35th
    AAAI Conference on Artificial Intelligence, AAAI 2021</i>, Virtual, Online, 2021,
    vol. 35, no. 9B, pp. 8209–8216.
  ista: 'Kungurtsev V, Egan M, Chatterjee B, Alistarh D-A. 2021. Asynchronous optimization
    methods for efficient training of deep neural networks with guarantees. 35th AAAI
    Conference on Artificial Intelligence, AAAI 2021. AAAI: Conference on Artificial
    Intelligence vol. 35, 8209–8216.'
  mla: Kungurtsev, Vyacheslav, et al. “Asynchronous Optimization Methods for Efficient
    Training of Deep Neural Networks with Guarantees.” <i>35th AAAI Conference on
    Artificial Intelligence, AAAI 2021</i>, vol. 35, no. 9B, AAAI Press, 2021, pp.
    8209–16.
  short: V. Kungurtsev, M. Egan, B. Chatterjee, D.-A. Alistarh, in:, 35th AAAI Conference
    on Artificial Intelligence, AAAI 2021, AAAI Press, 2021, pp. 8209–8216.
conference:
  end_date: 2021-02-09
  location: Virtual, Online
  name: 'AAAI: Conference on Artificial Intelligence'
  start_date: 2021-02-02
date_created: 2022-06-05T22:01:52Z
date_published: 2021-05-18T00:00:00Z
date_updated: 2022-06-07T06:53:36Z
day: '18'
department:
- _id: DaAl
ec_funded: 1
external_id:
  arxiv:
  - '1905.11845'
intvolume: '        35'
issue: 9B
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: ' https://doi.org/10.48550/arXiv.1905.11845'
month: '05'
oa: 1
oa_version: Preprint
page: 8209-8216
project:
- _id: 260C2330-B435-11E9-9278-68D0E5697425
  call_identifier: H2020
  grant_number: '754411'
  name: ISTplus - Postdoctoral Fellowships
- _id: 268A44D6-B435-11E9-9278-68D0E5697425
  call_identifier: H2020
  grant_number: '805223'
  name: Elastic Coordination for Scalable Machine Learning
publication: 35th AAAI Conference on Artificial Intelligence, AAAI 2021
publication_identifier:
  eissn:
  - 2374-3468
  isbn:
  - '9781713835974'
  issn:
  - 2159-5399
publication_status: published
publisher: AAAI Press
quality_controlled: '1'
scopus_import: '1'
status: public
title: Asynchronous optimization methods for efficient training of deep neural networks
  with guarantees
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 35
year: '2021'
...
---
_id: '14186'
abstract:
- lang: eng
  text: "The goal of the unsupervised learning of disentangled representations is
    to\r\nseparate the independent explanatory factors of variation in the data without\r\naccess
    to supervision. In this paper, we summarize the results of Locatello et\r\nal.,
    2019, and focus on their implications for practitioners. We discuss the\r\ntheoretical
    result showing that the unsupervised learning of disentangled\r\nrepresentations
    is fundamentally impossible without inductive biases and the\r\npractical challenges
    it entails. Finally, we comment on our experimental\r\nfindings, highlighting
    the limitations of state-of-the-art approaches and\r\ndirections for future research."
article_processing_charge: No
arxiv: 1
author:
- first_name: Francesco
  full_name: Locatello, Francesco
  id: 26cfd52f-2483-11ee-8040-88983bcc06d4
  last_name: Locatello
  orcid: 0000-0002-4850-0683
- first_name: Stefan
  full_name: Bauer, Stefan
  last_name: Bauer
- first_name: Mario
  full_name: Lucic, Mario
  last_name: Lucic
- first_name: Gunnar
  full_name: Rätsch, Gunnar
  last_name: Rätsch
- first_name: Sylvain
  full_name: Gelly, Sylvain
  last_name: Gelly
- first_name: Bernhard
  full_name: Schölkopf, Bernhard
  last_name: Schölkopf
- first_name: Olivier
  full_name: Bachem, Olivier
  last_name: Bachem
citation:
  ama: 'Locatello F, Bauer S, Lucic M, et al. A commentary on the unsupervised learning
    of disentangled representations. In: <i>The 34th AAAI Conference on Artificial
    Intelligence</i>. Vol 34. Association for the Advancement of Artificial Intelligence;
    2020:13681-13684. doi:<a href="https://doi.org/10.1609/aaai.v34i09.7120">10.1609/aaai.v34i09.7120</a>'
  apa: 'Locatello, F., Bauer, S., Lucic, M., Rätsch, G., Gelly, S., Schölkopf, B.,
    &#38; Bachem, O. (2020). A commentary on the unsupervised learning of disentangled
    representations. In <i>The 34th AAAI Conference on Artificial Intelligence</i>
    (Vol. 34, pp. 13681–13684). New York, NY, United States: Association for the Advancement
    of Artificial Intelligence. <a href="https://doi.org/10.1609/aaai.v34i09.7120">https://doi.org/10.1609/aaai.v34i09.7120</a>'
  chicago: Locatello, Francesco, Stefan Bauer, Mario Lucic, Gunnar Rätsch, Sylvain
    Gelly, Bernhard Schölkopf, and Olivier Bachem. “A Commentary on the Unsupervised
    Learning of Disentangled Representations.” In <i>The 34th AAAI Conference on Artificial
    Intelligence</i>, 34:13681–84. Association for the Advancement of Artificial Intelligence,
    2020. <a href="https://doi.org/10.1609/aaai.v34i09.7120">https://doi.org/10.1609/aaai.v34i09.7120</a>.
  ieee: F. Locatello <i>et al.</i>, “A commentary on the unsupervised learning of
    disentangled representations,” in <i>The 34th AAAI Conference on Artificial Intelligence</i>,
    New York, NY, United States, 2020, vol. 34, no. 9, pp. 13681–13684.
  ista: 'Locatello F, Bauer S, Lucic M, Rätsch G, Gelly S, Schölkopf B, Bachem O.
    2020. A commentary on the unsupervised learning of disentangled representations.
    The 34th AAAI Conference on Artificial Intelligence. AAAI: Conference on Artificial
    Intelligence vol. 34, 13681–13684.'
  mla: Locatello, Francesco, et al. “A Commentary on the Unsupervised Learning of
    Disentangled Representations.” <i>The 34th AAAI Conference on Artificial Intelligence</i>,
    vol. 34, no. 9, Association for the Advancement of Artificial Intelligence, 2020,
    pp. 13681–84, doi:<a href="https://doi.org/10.1609/aaai.v34i09.7120">10.1609/aaai.v34i09.7120</a>.
  short: F. Locatello, S. Bauer, M. Lucic, G. Rätsch, S. Gelly, B. Schölkopf, O. Bachem,
    in:, The 34th AAAI Conference on Artificial Intelligence, Association for the
    Advancement of Artificial Intelligence, 2020, pp. 13681–13684.
conference:
  end_date: 2020-02-12
  location: New York, NY, United States
  name: 'AAAI: Conference on Artificial Intelligence'
  start_date: 2020-02-07
date_created: 2023-08-22T14:07:26Z
date_published: 2020-07-28T00:00:00Z
date_updated: 2023-09-12T07:44:48Z
day: '28'
department:
- _id: FrLo
doi: 10.1609/aaai.v34i09.7120
extern: '1'
external_id:
  arxiv:
  - '2007.14184'
intvolume: '        34'
issue: '9'
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: https://arxiv.org/abs/2007.14184
month: '07'
oa: 1
oa_version: Preprint
page: 13681-13684
publication: The 34th AAAI Conference on Artificial Intelligence
publication_identifier:
  eissn:
  - 2374-3468
  isbn:
  - '9781577358350'
publication_status: published
publisher: Association for the Advancement of Artificial Intelligence
quality_controlled: '1'
scopus_import: '1'
status: public
title: A commentary on the unsupervised learning of disentangled representations
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 34
year: '2020'
...
---
_id: '9197'
abstract:
- lang: eng
  text: In this paper we introduce and study all-pay bidding games, a class of two
    player, zero-sum games on graphs. The game proceeds as follows. We place a token
    on some vertex in the graph and assign budgets to the two players. Each turn,
    each player submits a sealed legal bid (non-negative and below their remaining
    budget), which is deducted from their budget and the highest bidder moves the
    token onto an adjacent vertex. The game ends once a sink is reached, and Player
    1 pays Player 2 the outcome that is associated with the sink. The players attempt
    to maximize their expected outcome. Our games model settings where effort (of
    no inherent value) needs to be invested in an ongoing and stateful manner. On
    the negative side, we show that even in simple games on DAGs, optimal strategies
    may require a distribution over bids with infinite support. A central quantity
    in bidding games is the ratio of the players budgets. On the positive side, we
    show a simple FPTAS for DAGs, that, for each budget ratio, outputs an approximation
    for the optimal strategy for that ratio. We also implement it, show that it performs
    well, and suggests interesting properties of these games. Then, given an outcome
    c, we show an algorithm for finding the necessary and sufficient initial ratio
    for guaranteeing outcome c with probability 1 and a strategy ensuring such. Finally,
    while the general case has not previously been studied, solving the specific game
    in which Player 1 wins iff he wins the first two auctions, has been long stated
    as an open question, which we solve.
acknowledgement: This research was supported by the Austrian Science Fund (FWF) under
  grants S11402-N23 (RiSE/SHiNE), Z211-N23 (Wittgenstein Award), and M 2369-N33 (Meitner
  fellowship).
article_processing_charge: No
article_type: original
arxiv: 1
author:
- first_name: Guy
  full_name: Avni, Guy
  id: 463C8BC2-F248-11E8-B48F-1D18A9856A87
  last_name: Avni
  orcid: 0000-0001-5588-8287
- first_name: Rasmus
  full_name: Ibsen-Jensen, Rasmus
  id: 3B699956-F248-11E8-B48F-1D18A9856A87
  last_name: Ibsen-Jensen
  orcid: 0000-0003-4783-0389
- first_name: Josef
  full_name: Tkadlec, Josef
  id: 3F24CCC8-F248-11E8-B48F-1D18A9856A87
  last_name: Tkadlec
  orcid: 0000-0002-1097-9684
citation:
  ama: Avni G, Ibsen-Jensen R, Tkadlec J. All-pay bidding games on graphs. <i>Proceedings
    of the AAAI Conference on Artificial Intelligence</i>. 2020;34(02):1798-1805.
    doi:<a href="https://doi.org/10.1609/aaai.v34i02.5546">10.1609/aaai.v34i02.5546</a>
  apa: 'Avni, G., Ibsen-Jensen, R., &#38; Tkadlec, J. (2020). All-pay bidding games
    on graphs. <i>Proceedings of the AAAI Conference on Artificial Intelligence</i>.
    New York, NY, United States: Association for the Advancement of Artificial Intelligence.
    <a href="https://doi.org/10.1609/aaai.v34i02.5546">https://doi.org/10.1609/aaai.v34i02.5546</a>'
  chicago: Avni, Guy, Rasmus Ibsen-Jensen, and Josef Tkadlec. “All-Pay Bidding Games
    on Graphs.” <i>Proceedings of the AAAI Conference on Artificial Intelligence</i>.
    Association for the Advancement of Artificial Intelligence, 2020. <a href="https://doi.org/10.1609/aaai.v34i02.5546">https://doi.org/10.1609/aaai.v34i02.5546</a>.
  ieee: G. Avni, R. Ibsen-Jensen, and J. Tkadlec, “All-pay bidding games on graphs,”
    <i>Proceedings of the AAAI Conference on Artificial Intelligence</i>, vol. 34,
    no. 02. Association for the Advancement of Artificial Intelligence, pp. 1798–1805,
    2020.
  ista: Avni G, Ibsen-Jensen R, Tkadlec J. 2020. All-pay bidding games on graphs.
    Proceedings of the AAAI Conference on Artificial Intelligence. 34(02), 1798–1805.
  mla: Avni, Guy, et al. “All-Pay Bidding Games on Graphs.” <i>Proceedings of the
    AAAI Conference on Artificial Intelligence</i>, vol. 34, no. 02, Association for
    the Advancement of Artificial Intelligence, 2020, pp. 1798–805, doi:<a href="https://doi.org/10.1609/aaai.v34i02.5546">10.1609/aaai.v34i02.5546</a>.
  short: G. Avni, R. Ibsen-Jensen, J. Tkadlec, Proceedings of the AAAI Conference
    on Artificial Intelligence 34 (2020) 1798–1805.
conference:
  end_date: 2020-02-12
  location: New York, NY, United States
  name: 'AAAI: Conference on Artificial Intelligence'
  start_date: 2020-02-07
date_created: 2021-02-25T09:05:18Z
date_published: 2020-04-03T00:00:00Z
date_updated: 2023-09-05T12:40:00Z
day: '03'
department:
- _id: ToHe
- _id: KrCh
doi: 10.1609/aaai.v34i02.5546
external_id:
  arxiv:
  - '1911.08360'
intvolume: '        34'
issue: '02'
language:
- iso: eng
month: '04'
oa_version: Preprint
page: 1798-1805
project:
- _id: 25F2ACDE-B435-11E9-9278-68D0E5697425
  call_identifier: FWF
  grant_number: S11402-N23
  name: Rigorous Systems Engineering
- _id: 25F42A32-B435-11E9-9278-68D0E5697425
  call_identifier: FWF
  grant_number: Z211
  name: The Wittgenstein Prize
- _id: 264B3912-B435-11E9-9278-68D0E5697425
  call_identifier: FWF
  grant_number: M02369
  name: Formal Methods meets Algorithmic Game Theory
publication: Proceedings of the AAAI Conference on Artificial Intelligence
publication_identifier:
  eissn:
  - 2374-3468
  isbn:
  - '9781577358350'
  issn:
  - 2159-5399
publication_status: published
publisher: Association for the Advancement of Artificial Intelligence
quality_controlled: '1'
scopus_import: '1'
status: public
title: All-pay bidding games on graphs
type: journal_article
user_id: c635000d-4b10-11ee-a964-aac5a93f6ac1
volume: 34
year: '2020'
...
