---
_id: '14410'
abstract:
- lang: eng
  text: This paper focuses on the implementation details of the baseline methods and
    a recent lightweight conditional model extrapolation algorithm LIMES [5] for streaming
    data under class-prior shift. LIMES achieves superior performance over the baseline
    methods, especially concerning the minimum-across-day accuracy, which is important
    for the users of the system. In this work, the key measures to facilitate reproducibility
    and enhance the credibility of the results are described.
alternative_title:
- LNCS
article_processing_charge: No
author:
- first_name: Paulina
  full_name: Tomaszewska, Paulina
  last_name: Tomaszewska
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
citation:
  ama: 'Tomaszewska P, Lampert C. On the implementation of baselines and lightweight
    conditional model extrapolation (LIMES) under class-prior shift. In: <i>International
    Workshop on Reproducible Research in Pattern Recognition</i>. Vol 14068. Springer
    Nature; 2023:67-73. doi:<a href="https://doi.org/10.1007/978-3-031-40773-4_6">10.1007/978-3-031-40773-4_6</a>'
  apa: 'Tomaszewska, P., &#38; Lampert, C. (2023). On the implementation of baselines
    and lightweight conditional model extrapolation (LIMES) under class-prior shift.
    In <i>International Workshop on Reproducible Research in Pattern Recognition</i>
    (Vol. 14068, pp. 67–73). Montreal, Canada: Springer Nature. <a href="https://doi.org/10.1007/978-3-031-40773-4_6">https://doi.org/10.1007/978-3-031-40773-4_6</a>'
  chicago: Tomaszewska, Paulina, and Christoph Lampert. “On the Implementation of Baselines
    and Lightweight Conditional Model Extrapolation (LIMES) under Class-Prior Shift.”
    In <i>International Workshop on Reproducible Research in Pattern Recognition</i>,
    14068:67–73. Springer Nature, 2023. <a href="https://doi.org/10.1007/978-3-031-40773-4_6">https://doi.org/10.1007/978-3-031-40773-4_6</a>.
  ieee: P. Tomaszewska and C. Lampert, “On the implementation of baselines and lightweight
    conditional model extrapolation (LIMES) under class-prior shift,” in <i>International
    Workshop on Reproducible Research in Pattern Recognition</i>, Montreal, Canada,
    2023, vol. 14068, pp. 67–73.
  ista: 'Tomaszewska P, Lampert C. 2023. On the implementation of baselines and lightweight
    conditional model extrapolation (LIMES) under class-prior shift. International
    Workshop on Reproducible Research in Pattern Recognition. RRPR: Reproducible Research
    in Pattern Recognition, LNCS, vol. 14068, 67–73.'
  mla: Tomaszewska, Paulina, and Christoph Lampert. “On the Implementation of Baselines
    and Lightweight Conditional Model Extrapolation (LIMES) under Class-Prior Shift.”
    <i>International Workshop on Reproducible Research in Pattern Recognition</i>,
    vol. 14068, Springer Nature, 2023, pp. 67–73, doi:<a href="https://doi.org/10.1007/978-3-031-40773-4_6">10.1007/978-3-031-40773-4_6</a>.
  short: P. Tomaszewska, C. Lampert, in:, International Workshop on Reproducible Research
    in Pattern Recognition, Springer Nature, 2023, pp. 67–73.
conference:
  end_date: 2022-08-21
  location: Montreal, Canada
  name: 'RRPR: Reproducible Research in Pattern Recognition'
  start_date: 2022-08-21
date_created: 2023-10-08T22:01:18Z
date_published: 2023-08-20T00:00:00Z
date_updated: 2023-10-09T06:48:02Z
day: '20'
department:
- _id: ChLa
doi: 10.1007/978-3-031-40773-4_6
intvolume: '     14068'
language:
- iso: eng
month: '08'
oa_version: None
page: 67-73
publication: International Workshop on Reproducible Research in Pattern Recognition
publication_identifier:
  eissn:
  - 1611-3349
  isbn:
  - '9783031407727'
  issn:
  - 0302-9743
publication_status: published
publisher: Springer Nature
quality_controlled: '1'
scopus_import: '1'
status: public
title: On the implementation of baselines and lightweight conditional model extrapolation
  (LIMES) under class-prior shift
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 14068
year: '2023'
...
---
_id: '14921'
abstract:
- lang: eng
  text: Neural collapse (NC) refers to the surprising structure of the last layer
    of deep neural networks in the terminal phase of gradient descent training. Recently,
    an increasing amount of experimental evidence has pointed to the propagation of
    NC to earlier layers of neural networks. However, while the NC in the last layer
    is well studied theoretically, much less is known about its multi-layered counterpart
    - deep neural collapse (DNC). In particular, existing work focuses either on linear
    layers or only on the last two layers at the price of an extra assumption. Our
    paper fills this gap by generalizing the established analytical framework for
    NC - the unconstrained features model - to multiple non-linear layers. Our key
    technical contribution is to show that, in a deep unconstrained features model,
    the unique global optimum for binary classification exhibits all the properties
    typical of DNC. This explains the existing experimental evidence of DNC. We also
    empirically show that (i) by optimizing deep unconstrained features models via
    gradient descent, the resulting solution agrees well with our theory, and (ii)
    trained networks recover the unconstrained features suitable for the occurrence
    of DNC, thus supporting the validity of this modeling principle.
acknowledgement: M. M. is partially supported by the 2019 Lopez-Loreta Prize. The
  authors would like to thank Eugenia Iofinova, Bernd Prach and Simone Bombari for
  valuable feedback on the manuscript.
alternative_title:
- NeurIPS
article_processing_charge: No
arxiv: 1
author:
- first_name: Peter
  full_name: Súkeník, Peter
  id: d64d6a8d-eb8e-11eb-b029-96fd216dec3c
  last_name: Súkeník
- first_name: Marco
  full_name: Mondelli, Marco
  id: 27EB676C-8706-11E9-9510-7717E6697425
  last_name: Mondelli
  orcid: 0000-0002-3242-7020
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
citation:
  ama: 'Súkeník P, Mondelli M, Lampert C. Deep neural collapse is provably optimal
    for the deep unconstrained features model. In: <i>37th Annual Conference on Neural
    Information Processing Systems</i>.'
  apa: Súkeník, P., Mondelli, M., &#38; Lampert, C. (n.d.). Deep neural collapse is
    provably optimal for the deep unconstrained features model. In <i>37th Annual
    Conference on Neural Information Processing Systems</i>. New Orleans, LA, United
    States.
  chicago: Súkeník, Peter, Marco Mondelli, and Christoph Lampert. “Deep Neural Collapse
    Is Provably Optimal for the Deep Unconstrained Features Model.” In <i>37th Annual
    Conference on Neural Information Processing Systems</i>, n.d.
  ieee: P. Súkeník, M. Mondelli, and C. Lampert, “Deep neural collapse is provably
    optimal for the deep unconstrained features model,” in <i>37th Annual Conference
    on Neural Information Processing Systems</i>, New Orleans, LA, United States.
  ista: 'Súkeník P, Mondelli M, Lampert C. Deep neural collapse is provably optimal
    for the deep unconstrained features model. 37th Annual Conference on Neural Information
    Processing Systems. NeurIPS: Neural Information Processing Systems, NeurIPS, .'
  mla: Súkeník, Peter, et al. “Deep Neural Collapse Is Provably Optimal for the Deep
    Unconstrained Features Model.” <i>37th Annual Conference on Neural Information
    Processing Systems</i>.
  short: P. Súkeník, M. Mondelli, C. Lampert, in:, 37th Annual Conference on Neural
    Information Processing Systems, n.d.
conference:
  end_date: 2023-12-16
  location: New Orleans, LA, United States
  name: 'NeurIPS: Neural Information Processing Systems'
  start_date: 2023-12-10
date_created: 2024-02-02T11:17:41Z
date_published: 2023-12-15T00:00:00Z
date_updated: 2024-09-10T13:03:19Z
day: '15'
department:
- _id: MaMo
- _id: ChLa
external_id:
  arxiv:
  - '2305.13165'
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: ' https://doi.org/10.48550/arXiv.2305.13165'
month: '12'
oa: 1
oa_version: Preprint
project:
- _id: 059876FA-7A3F-11EA-A408-12923DDC885E
  name: Prix Lopez-Loretta 2019 - Marco Mondelli
publication: 37th Annual Conference on Neural Information Processing Systems
publication_status: inpress
quality_controlled: '1'
status: public
title: Deep neural collapse is provably optimal for the deep unconstrained features
  model
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
year: '2023'
...
---
_id: '15039'
abstract:
- lang: eng
  text: 'A crucial property for achieving secure, trustworthy and interpretable deep
    learning systems is their robustness: small changes to a system''s inputs should
    not result in large changes to its outputs. Mathematically, this means one strives
    for networks with a small Lipschitz constant. Several recent works have focused
    on how to construct such Lipschitz networks, typically by imposing constraints
    on the weight matrices. In this work, we study an orthogonal aspect, namely the
    role of the activation function. We show that commonly used activation functions,
    such as MaxMin, as well as all piece-wise linear ones with two segments unnecessarily
    restrict the class of representable functions, even in the simplest one-dimensional
    setting. We furthermore introduce the new N-activation function that is provably
    more expressive than currently popular activation functions. We provide code at
    this https URL.'
article_number: '2311.06103'
article_processing_charge: No
arxiv: 1
author:
- first_name: Bernd
  full_name: Prach, Bernd
  id: 2D561D42-C427-11E9-89B4-9C1AE6697425
  last_name: Prach
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
citation:
  ama: Prach B, Lampert C. 1-Lipschitz neural networks are more expressive with N-activations.
    <i>arXiv</i>. doi:<a href="https://doi.org/10.48550/ARXIV.2311.06103">10.48550/ARXIV.2311.06103</a>
  apa: Prach, B., &#38; Lampert, C. (n.d.). 1-Lipschitz neural networks are more expressive
    with N-activations. <i>arXiv</i>. <a href="https://doi.org/10.48550/ARXIV.2311.06103">https://doi.org/10.48550/ARXIV.2311.06103</a>
  chicago: Prach, Bernd, and Christoph Lampert. “1-Lipschitz Neural Networks Are More
    Expressive with N-Activations.” <i>ArXiv</i>, n.d. <a href="https://doi.org/10.48550/ARXIV.2311.06103">https://doi.org/10.48550/ARXIV.2311.06103</a>.
  ieee: B. Prach and C. Lampert, “1-Lipschitz neural networks are more expressive
    with N-activations,” <i>arXiv</i>. .
  ista: Prach B, Lampert C. 1-Lipschitz neural networks are more expressive with N-activations.
    arXiv, 2311.06103.
  mla: Prach, Bernd, and Christoph Lampert. “1-Lipschitz Neural Networks Are More
    Expressive with N-Activations.” <i>ArXiv</i>, 2311.06103, doi:<a href="https://doi.org/10.48550/ARXIV.2311.06103">10.48550/ARXIV.2311.06103</a>.
  short: B. Prach, C. Lampert, ArXiv (n.d.).
date_created: 2024-02-28T17:59:32Z
date_published: 2023-11-10T00:00:00Z
date_updated: 2024-03-04T07:02:39Z
day: '10'
department:
- _id: GradSch
- _id: ChLa
doi: 10.48550/ARXIV.2311.06103
external_id:
  arxiv:
  - '2311.06103'
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: https://doi.org/10.48550/arXiv.2311.06103
month: '11'
oa: 1
oa_version: Preprint
publication: arXiv
publication_status: submitted
status: public
title: 1-Lipschitz neural networks are more expressive with N-activations
type: preprint
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
year: '2023'
...
---
_id: '13053'
abstract:
- lang: eng
  text: 'Deep neural networks (DNNs) often have to be compressed, via pruning and/or
    quantization, before they can be deployed in practical settings. In this work
    we propose a new compression-aware minimizer dubbed CrAM that modifies the optimization
    step in a principled way, in order to produce models whose local loss behavior
    is stable under compression operations such as pruning. Thus, dense models trained
    via CrAM should be compressible post-training, in a single step, without significant
    accuracy loss. Experimental results on standard benchmarks, such as residual networks
    for ImageNet classification and BERT models for language modelling, show that
    CrAM produces dense models that can be more accurate than the standard SGD/Adam-based
    baselines, but which are stable under weight pruning: specifically, we can prune
    models in one-shot to 70-80% sparsity with almost no accuracy loss, and to 90%
    with reasonable (∼1%) accuracy loss, which is competitive with gradual compression
    methods. Additionally, CrAM can produce sparse models which perform well for transfer
    learning, and it also works for semi-structured 2:4 pruning patterns supported
    by GPU hardware. The code for reproducing the results is available at this https
    URL .'
acknowledged_ssus:
- _id: ScienComp
acknowledgement: "AP, EK, DA received funding from the European Research Council (ERC)
  under the European\r\nUnion’s Horizon 2020 research and innovation programme (grant
  agreement No 805223 ScaleML). AV acknowledges the support of the French Agence Nationale
  de la Recherche (ANR), under grant ANR-21-CE48-0016 (project COMCOPT). We further
  acknowledge the support from the Scientific Service Units (SSU) of ISTA through
  resources provided by Scientific Computing (SciComp)-"
article_processing_charge: No
arxiv: 1
author:
- first_name: Elena-Alexandra
  full_name: Peste, Elena-Alexandra
  id: 32D78294-F248-11E8-B48F-1D18A9856A87
  last_name: Peste
- first_name: Adrian
  full_name: Vladu, Adrian
  last_name: Vladu
- first_name: Eldar
  full_name: Kurtic, Eldar
  id: 47beb3a5-07b5-11eb-9b87-b108ec578218
  last_name: Kurtic
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
- first_name: Dan-Adrian
  full_name: Alistarh, Dan-Adrian
  id: 4A899BFC-F248-11E8-B48F-1D18A9856A87
  last_name: Alistarh
  orcid: 0000-0003-3650-940X
citation:
  ama: 'Peste E-A, Vladu A, Kurtic E, Lampert C, Alistarh D-A. CrAM: A Compression-Aware
    Minimizer. In: <i>11th International Conference on Learning Representations </i>.'
  apa: 'Peste, E.-A., Vladu, A., Kurtic, E., Lampert, C., &#38; Alistarh, D.-A. (n.d.).
    CrAM: A Compression-Aware Minimizer. In <i>11th International Conference on Learning
    Representations </i>. Kigali, Rwanda .'
  chicago: 'Peste, Elena-Alexandra, Adrian Vladu, Eldar Kurtic, Christoph Lampert,
    and Dan-Adrian Alistarh. “CrAM: A Compression-Aware Minimizer.” In <i>11th International
    Conference on Learning Representations </i>, n.d.'
  ieee: 'E.-A. Peste, A. Vladu, E. Kurtic, C. Lampert, and D.-A. Alistarh, “CrAM:
    A Compression-Aware Minimizer,” in <i>11th International Conference on Learning
    Representations </i>, Kigali, Rwanda .'
  ista: 'Peste E-A, Vladu A, Kurtic E, Lampert C, Alistarh D-A. CrAM: A Compression-Aware
    Minimizer. 11th International Conference on Learning Representations . ICLR: International
    Conference on Learning Representations.'
  mla: 'Peste, Elena-Alexandra, et al. “CrAM: A Compression-Aware Minimizer.” <i>11th
    International Conference on Learning Representations </i>.'
  short: E.-A. Peste, A. Vladu, E. Kurtic, C. Lampert, D.-A. Alistarh, in:, 11th International
    Conference on Learning Representations , n.d.
conference:
  end_date: 2023-05-05
  location: 'Kigali, Rwanda '
  name: 'ICLR: International Conference on Learning Representations'
  start_date: 2023-05-01
date_created: 2023-05-23T11:36:18Z
date_published: 2023-05-01T00:00:00Z
date_updated: 2023-06-01T12:54:45Z
department:
- _id: GradSch
- _id: DaAl
- _id: ChLa
ec_funded: 1
external_id:
  arxiv:
  - '2207.14200'
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: https://openreview.net/pdf?id=_eTZBs-yedr
month: '05'
oa: 1
oa_version: Preprint
project:
- _id: 268A44D6-B435-11E9-9278-68D0E5697425
  call_identifier: H2020
  grant_number: '805223'
  name: Elastic Coordination for Scalable Machine Learning
publication: '11th International Conference on Learning Representations '
publication_status: accepted
quality_controlled: '1'
related_material:
  record:
  - id: '13074'
    relation: dissertation_contains
    status: public
status: public
title: 'CrAM: A Compression-Aware Minimizer'
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
year: '2023'
...
---
_id: '10752'
abstract:
- lang: eng
  text: 'The digitalization of almost all aspects of our everyday lives has led to
    unprecedented amounts of data being freely available on the Internet. In particular
    social media platforms provide rich sources of user-generated data, though typically
    in unstructured form, and with high diversity, such as written in many different
    languages. Automatically identifying meaningful information in such big data resources
    and extracting it efficiently is one of the ongoing challenges of our time. A
    common step for this is sentiment analysis, which forms the foundation for tasks
    such as opinion mining or trend prediction. Unfortunately, publicly available
    tools for this task are almost exclusively available for English-language texts.
    Consequently, a large fraction of the Internet users, who do not communicate in
    English, are ignored in automatized studies, a phenomenon called rare-language
    discrimination.In this work we propose a technique to overcome this problem by
    a truly multi-lingual model, which can be trained automatically without linguistic
    knowledge or even the ability to read the many target languages. The main step
    is to combine self-annotation, specifically the use of emoticons as a proxy for
    labels, with multi-lingual sentence representations.To evaluate our method we
    curated several large datasets from data obtained via the free Twitter streaming
    API. The results show that our proposed multi-lingual training is able to achieve
    sentiment predictions at the same quality level for rare languages as for frequent
    ones, and in particular clearly better than what mono-lingual training achieves
    on the same data. '
article_processing_charge: No
author:
- first_name: Jasmin
  full_name: Lampert, Jasmin
  last_name: Lampert
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0002-4561-241X
citation:
  ama: 'Lampert J, Lampert C. Overcoming rare-language discrimination in multi-lingual
    sentiment analysis. In: <i>2021 IEEE International Conference on Big Data</i>.
    IEEE; 2022:5185-5192. doi:<a href="https://doi.org/10.1109/bigdata52589.2021.9672003">10.1109/bigdata52589.2021.9672003</a>'
  apa: 'Lampert, J., &#38; Lampert, C. (2022). Overcoming rare-language discrimination
    in multi-lingual sentiment analysis. In <i>2021 IEEE International Conference
    on Big Data</i> (pp. 5185–5192). Orlando, FL, United States: IEEE. <a href="https://doi.org/10.1109/bigdata52589.2021.9672003">https://doi.org/10.1109/bigdata52589.2021.9672003</a>'
  chicago: Lampert, Jasmin, and Christoph Lampert. “Overcoming Rare-Language Discrimination
    in Multi-Lingual Sentiment Analysis.” In <i>2021 IEEE International Conference
    on Big Data</i>, 5185–92. IEEE, 2022. <a href="https://doi.org/10.1109/bigdata52589.2021.9672003">https://doi.org/10.1109/bigdata52589.2021.9672003</a>.
  ieee: J. Lampert and C. Lampert, “Overcoming rare-language discrimination in multi-lingual
    sentiment analysis,” in <i>2021 IEEE International Conference on Big Data</i>,
    Orlando, FL, United States, 2022, pp. 5185–5192.
  ista: 'Lampert J, Lampert C. 2022. Overcoming rare-language discrimination in multi-lingual
    sentiment analysis. 2021 IEEE International Conference on Big Data. Big Data:
    International Conference on Big Data, 5185–5192.'
  mla: Lampert, Jasmin, and Christoph Lampert. “Overcoming Rare-Language Discrimination
    in Multi-Lingual Sentiment Analysis.” <i>2021 IEEE International Conference on
    Big Data</i>, IEEE, 2022, pp. 5185–92, doi:<a href="https://doi.org/10.1109/bigdata52589.2021.9672003">10.1109/bigdata52589.2021.9672003</a>.
  short: J. Lampert, C. Lampert, in:, 2021 IEEE International Conference on Big Data,
    IEEE, 2022, pp. 5185–5192.
conference:
  end_date: 2021-12-18
  location: Orlando, FL, United States
  name: 'Big Data: International Conference on Big Data'
  start_date: 2021-12-15
date_created: 2022-02-10T14:08:23Z
date_published: 2022-01-13T00:00:00Z
date_updated: 2023-08-02T14:27:50Z
day: '13'
department:
- _id: ChLa
doi: 10.1109/bigdata52589.2021.9672003
external_id:
  isi:
  - '000800559505036'
isi: 1
language:
- iso: eng
month: '01'
oa_version: None
page: 5185-5192
publication: 2021 IEEE International Conference on Big Data
publication_identifier:
  isbn:
  - '9781665439022'
publication_status: published
publisher: IEEE
quality_controlled: '1'
status: public
title: Overcoming rare-language discrimination in multi-lingual sentiment analysis
type: conference
user_id: 4359f0d1-fa6c-11eb-b949-802e58b17ae8
year: '2022'
...
---
_id: '10802'
abstract:
- lang: eng
  text: "Addressing fairness concerns about machine learning models is a crucial step
    towards their long-term adoption in real-world automated systems. While many approaches
    have been developed for training fair models from data, little is known about
    the robustness of these methods to data corruption. In this work we consider fairness-aware
    learning under worst-case data manipulations. We show that an adversary can in
    some situations force any learner to return an overly biased classifier, regardless
    of the sample size and with or without degrading\r\naccuracy, and that the strength
    of the excess bias increases for learning problems with underrepresented protected
    groups in the data. We also prove that our hardness results are tight up to constant
    factors. To this end, we study two natural learning algorithms that optimize for
    both accuracy and fairness and show that these algorithms enjoy guarantees that
    are order-optimal in terms of the corruption ratio and the protected groups frequencies
    in the large data\r\nlimit."
acknowledgement: The authors thank Eugenia Iofinova and Bernd Prach for providing
  feedback on early versions of this paper. This publication was made possible by
  an ETH AI Center postdoctoral fellowship to Nikola Konstantinov.
article_processing_charge: No
article_type: original
arxiv: 1
author:
- first_name: Nikola H
  full_name: Konstantinov, Nikola H
  id: 4B9D76E4-F248-11E8-B48F-1D18A9856A87
  last_name: Konstantinov
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0002-4561-241X
citation:
  ama: Konstantinov NH, Lampert C. Fairness-aware PAC learning from corrupted data.
    <i>Journal of Machine Learning Research</i>. 2022;23:1-60.
  apa: Konstantinov, N. H., &#38; Lampert, C. (2022). Fairness-aware PAC learning
    from corrupted data. <i>Journal of Machine Learning Research</i>. ML Research
    Press.
  chicago: Konstantinov, Nikola H, and Christoph Lampert. “Fairness-Aware PAC Learning
    from Corrupted Data.” <i>Journal of Machine Learning Research</i>. ML Research
    Press, 2022.
  ieee: N. H. Konstantinov and C. Lampert, “Fairness-aware PAC learning from corrupted
    data,” <i>Journal of Machine Learning Research</i>, vol. 23. ML Research Press,
    pp. 1–60, 2022.
  ista: Konstantinov NH, Lampert C. 2022. Fairness-aware PAC learning from corrupted
    data. Journal of Machine Learning Research. 23, 1–60.
  mla: Konstantinov, Nikola H., and Christoph Lampert. “Fairness-Aware PAC Learning
    from Corrupted Data.” <i>Journal of Machine Learning Research</i>, vol. 23, ML
    Research Press, 2022, pp. 1–60.
  short: N.H. Konstantinov, C. Lampert, Journal of Machine Learning Research 23 (2022)
    1–60.
date_created: 2022-02-28T14:05:42Z
date_published: 2022-05-01T00:00:00Z
date_updated: 2023-09-26T10:44:37Z
day: '01'
ddc:
- '004'
department:
- _id: ChLa
external_id:
  arxiv:
  - '2102.06004'
file:
- access_level: open_access
  checksum: 9cac897b54a0ddf3a553a2c33e88cfda
  content_type: application/pdf
  creator: kschuh
  date_created: 2022-07-12T15:08:28Z
  date_updated: 2022-07-12T15:08:28Z
  file_id: '11570'
  file_name: 2022_JournalMachineLearningResearch_Konstantinov.pdf
  file_size: 551862
  relation: main_file
  success: 1
file_date_updated: 2022-07-12T15:08:28Z
has_accepted_license: '1'
intvolume: '        23'
keyword:
- Fairness
- robustness
- data poisoning
- trustworthy machine learning
- PAC learning
language:
- iso: eng
license: https://creativecommons.org/licenses/by/4.0/
month: '05'
oa: 1
oa_version: Published Version
page: 1-60
publication: Journal of Machine Learning Research
publication_identifier:
  eissn:
  - 1533-7928
  issn:
  - 1532-4435
publication_status: published
publisher: ML Research Press
quality_controlled: '1'
related_material:
  record:
  - id: '10799'
    relation: dissertation_contains
    status: public
  - id: '13241'
    relation: shorter_version
    status: public
scopus_import: '1'
status: public
title: Fairness-aware PAC learning from corrupted data
tmp:
  image: /images/cc_by.png
  legal_code_url: https://creativecommons.org/licenses/by/4.0/legalcode
  name: Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)
  short: CC BY (4.0)
type: journal_article
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 23
year: '2022'
...
---
_id: '13241'
abstract:
- lang: eng
  text: Addressing fairness concerns about machine learning models is a crucial step
    towards their long-term adoption in real-world automated systems. Many approaches
    for training fair models from data have been developed and an implicit assumption
    about such algorithms is that they are able to recover a fair model, despite potential
    historical biases in the data. In this work we show a number of impossibility
    results that indicate that there is no learning algorithm that can recover a fair
    model when a proportion of the dataset is subject to arbitrary manipulations.
    Specifically, we prove that there are situations in which an adversary can force
    any learner to return a biased classifier, with or without degrading accuracy,
    and that the strength of this bias increases for learning problems with underrepresented
    protected groups in the data. Our results emphasize on the importance of studying
    further data corruption models of various strength and of establishing stricter
    data collection practices for fairness-aware learning.
acknowledgement: "This paper is a shortened, workshop version of Konstantinov and
  Lampert (2021),\r\nhttps://arxiv.org/abs/2102.06004. For further results, including
  an analysis of algorithms achieving the lower bounds from this paper, we refer to
  the full version."
article_processing_charge: No
arxiv: 1
author:
- first_name: Nikola H
  full_name: Konstantinov, Nikola H
  id: 4B9D76E4-F248-11E8-B48F-1D18A9856A87
  last_name: Konstantinov
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
citation:
  ama: 'Konstantinov NH, Lampert C. On the impossibility of fairness-aware learning
    from corrupted data. In: <i>Proceedings of Machine Learning Research</i>. Vol
    171. ML Research Press; 2022:59-83.'
  apa: Konstantinov, N. H., &#38; Lampert, C. (2022). On the impossibility of fairness-aware
    learning from corrupted data. In <i>Proceedings of Machine Learning Research</i>
    (Vol. 171, pp. 59–83). ML Research Press.
  chicago: Konstantinov, Nikola H, and Christoph Lampert. “On the Impossibility of
    Fairness-Aware Learning from Corrupted Data.” In <i>Proceedings of Machine Learning
    Research</i>, 171:59–83. ML Research Press, 2022.
  ieee: N. H. Konstantinov and C. Lampert, “On the impossibility of fairness-aware
    learning from corrupted data,” in <i>Proceedings of Machine Learning Research</i>,
    2022, vol. 171, pp. 59–83.
  ista: Konstantinov NH, Lampert C. 2022. On the impossibility of fairness-aware learning
    from corrupted data. Proceedings of Machine Learning Research. vol. 171, 59–83.
  mla: Konstantinov, Nikola H., and Christoph Lampert. “On the Impossibility of Fairness-Aware
    Learning from Corrupted Data.” <i>Proceedings of Machine Learning Research</i>,
    vol. 171, ML Research Press, 2022, pp. 59–83.
  short: N.H. Konstantinov, C. Lampert, in:, Proceedings of Machine Learning Research,
    ML Research Press, 2022, pp. 59–83.
date_created: 2023-07-16T22:01:13Z
date_published: 2022-12-01T00:00:00Z
date_updated: 2023-09-26T10:44:37Z
day: '01'
department:
- _id: ChLa
external_id:
  arxiv:
  - '2102.06004'
intvolume: '       171'
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: https://arxiv.org/abs/2102.06004
month: '12'
oa: 1
oa_version: Preprint
page: 59-83
publication: Proceedings of Machine Learning Research
publication_identifier:
  eissn:
  - 2640-3498
publication_status: published
publisher: ML Research Press
quality_controlled: '1'
related_material:
  record:
  - id: '10802'
    relation: extended_version
    status: public
scopus_import: '1'
status: public
title: On the impossibility of fairness-aware learning from corrupted data
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 171
year: '2022'
...
---
_id: '11839'
abstract:
- lang: eng
  text: "It is a highly desirable property for deep networks to be robust against\r\nsmall
    input changes. One popular way to achieve this property is by designing\r\nnetworks
    with a small Lipschitz constant. In this work, we propose a new\r\ntechnique for
    constructing such Lipschitz networks that has a number of\r\ndesirable properties:
    it can be applied to any linear network layer\r\n(fully-connected or convolutional),
    it provides formal guarantees on the\r\nLipschitz constant, it is easy to implement
    and efficient to run, and it can be\r\ncombined with any training objective and
    optimization method. In fact, our\r\ntechnique is the first one in the literature
    that achieves all of these\r\nproperties simultaneously. Our main contribution
    is a rescaling-based weight\r\nmatrix parametrization that guarantees each network
    layer to have a Lipschitz\r\nconstant of at most 1 and results in the learned
    weight matrices to be close to\r\northogonal. Hence we call such layers almost-orthogonal
    Lipschitz (AOL).\r\nExperiments and ablation studies in the context of image classification
    with\r\ncertified robust accuracy confirm that AOL layers achieve results that
    are on\r\npar with most existing methods. Yet, they are simpler to implement and
    more\r\nbroadly applicable, because they do not require computationally expensive\r\nmatrix
    orthogonalization or inversion steps as part of the network\r\narchitecture. We
    provide code at https://github.com/berndprach/AOL."
alternative_title:
- LNCS
article_processing_charge: No
arxiv: 1
author:
- first_name: Bernd
  full_name: Prach, Bernd
  id: 2D561D42-C427-11E9-89B4-9C1AE6697425
  last_name: Prach
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
citation:
  ama: 'Prach B, Lampert C. Almost-orthogonal layers for efficient general-purpose
    Lipschitz networks. In: <i>Computer Vision – ECCV 2022</i>. Vol 13681. Springer
    Nature; 2022:350-365. doi:<a href="https://doi.org/10.1007/978-3-031-19803-8_21">10.1007/978-3-031-19803-8_21</a>'
  apa: 'Prach, B., &#38; Lampert, C. (2022). Almost-orthogonal layers for efficient
    general-purpose Lipschitz networks. In <i>Computer Vision – ECCV 2022</i> (Vol.
    13681, pp. 350–365). Tel Aviv, Israel: Springer Nature. <a href="https://doi.org/10.1007/978-3-031-19803-8_21">https://doi.org/10.1007/978-3-031-19803-8_21</a>'
  chicago: Prach, Bernd, and Christoph Lampert. “Almost-Orthogonal Layers for Efficient
    General-Purpose Lipschitz Networks.” In <i>Computer Vision – ECCV 2022</i>, 13681:350–65.
    Springer Nature, 2022. <a href="https://doi.org/10.1007/978-3-031-19803-8_21">https://doi.org/10.1007/978-3-031-19803-8_21</a>.
  ieee: B. Prach and C. Lampert, “Almost-orthogonal layers for efficient general-purpose
    Lipschitz networks,” in <i>Computer Vision – ECCV 2022</i>, Tel Aviv, Israel,
    2022, vol. 13681, pp. 350–365.
  ista: 'Prach B, Lampert C. 2022. Almost-orthogonal layers for efficient general-purpose
    Lipschitz networks. Computer Vision – ECCV 2022. ECCV: European Conference on
    Computer Vision, LNCS, vol. 13681, 350–365.'
  mla: Prach, Bernd, and Christoph Lampert. “Almost-Orthogonal Layers for Efficient
    General-Purpose Lipschitz Networks.” <i>Computer Vision – ECCV 2022</i>, vol.
    13681, Springer Nature, 2022, pp. 350–65, doi:<a href="https://doi.org/10.1007/978-3-031-19803-8_21">10.1007/978-3-031-19803-8_21</a>.
  short: B. Prach, C. Lampert, in:, Computer Vision – ECCV 2022, Springer Nature,
    2022, pp. 350–365.
conference:
  end_date: 2022-10-27
  location: Tel Aviv, Israel
  name: 'ECCV: European Conference on Computer Vision'
  start_date: 2022-10-23
date_created: 2022-08-12T15:09:47Z
date_published: 2022-10-23T00:00:00Z
date_updated: 2023-05-03T08:00:46Z
day: '23'
department:
- _id: GradSch
- _id: ChLa
doi: 10.1007/978-3-031-19803-8_21
external_id:
  arxiv:
  - '2208.03160'
intvolume: '     13681'
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: ' https://doi.org/10.48550/arXiv.2208.03160'
month: '10'
oa: 1
oa_version: Preprint
page: 350-365
publication: Computer Vision – ECCV 2022
publication_identifier:
  eisbn:
  - '9783031198038'
  isbn:
  - '9783031198021'
publication_status: published
publisher: Springer Nature
quality_controlled: '1'
scopus_import: '1'
status: public
title: Almost-orthogonal layers for efficient general-purpose Lipschitz networks
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 13681
year: '2022'
...
---
_id: '12161'
abstract:
- lang: eng
  text: 'We introduce LIMES, a new method for learning with non-stationary streaming
    data, inspired by the recent success of meta-learning. The main idea is not to
    attempt to learn a single classifier that would have to work well across all occurring
    data distributions, nor many separate classifiers, but to exploit a hybrid strategy:
    we learn a single set of model parameters from which a specific classifier for
    any specific data distribution is derived via classifier adaptation. Assuming
    a multiclass classification setting with class-prior shift, the adaptation step
    can be performed analytically with only the classifier’s bias terms being affected.
    Another contribution of our work is an extrapolation step that predicts suitable
    adaptation parameters for future time steps based on the previous data. In combination,
    we obtain a lightweight procedure for learning from streaming data with varying
    class distribution that adds no trainable parameters and almost no memory or computational
    overhead compared to training a single model. Experiments on a set of exemplary
    tasks using Twitter data show that LIMES achieves higher accuracy than alternative
    approaches, especially with respect to the relevant real-world metric of lowest
    within-day accuracy.'
article_processing_charge: No
arxiv: 1
author:
- first_name: Paulina
  full_name: Tomaszewska, Paulina
  last_name: Tomaszewska
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
citation:
  ama: 'Tomaszewska P, Lampert C. Lightweight conditional model extrapolation for
    streaming data under class-prior shift. In: <i>26th International Conference on
    Pattern Recognition</i>. Vol 2022. Institute of Electrical and Electronics Engineers;
    2022:2128-2134. doi:<a href="https://doi.org/10.1109/icpr56361.2022.9956195">10.1109/icpr56361.2022.9956195</a>'
  apa: 'Tomaszewska, P., &#38; Lampert, C. (2022). Lightweight conditional model extrapolation
    for streaming data under class-prior shift. In <i>26th International Conference
    on Pattern Recognition</i> (Vol. 2022, pp. 2128–2134). Montreal, Canada: Institute
    of Electrical and Electronics Engineers. <a href="https://doi.org/10.1109/icpr56361.2022.9956195">https://doi.org/10.1109/icpr56361.2022.9956195</a>'
  chicago: Tomaszewska, Paulina, and Christoph Lampert. “Lightweight Conditional Model
    Extrapolation for Streaming Data under Class-Prior Shift.” In <i>26th International
    Conference on Pattern Recognition</i>, 2022:2128–34. Institute of Electrical and
    Electronics Engineers, 2022. <a href="https://doi.org/10.1109/icpr56361.2022.9956195">https://doi.org/10.1109/icpr56361.2022.9956195</a>.
  ieee: P. Tomaszewska and C. Lampert, “Lightweight conditional model extrapolation
    for streaming data under class-prior shift,” in <i>26th International Conference
    on Pattern Recognition</i>, Montreal, Canada, 2022, vol. 2022, pp. 2128–2134.
  ista: 'Tomaszewska P, Lampert C. 2022. Lightweight conditional model extrapolation
    for streaming data under class-prior shift. 26th International Conference on Pattern
    Recognition. ICPR: International Conference on Pattern Recognition vol. 2022,
    2128–2134.'
  mla: Tomaszewska, Paulina, and Christoph Lampert. “Lightweight Conditional Model
    Extrapolation for Streaming Data under Class-Prior Shift.” <i>26th International
    Conference on Pattern Recognition</i>, vol. 2022, Institute of Electrical and
    Electronics Engineers, 2022, pp. 2128–34, doi:<a href="https://doi.org/10.1109/icpr56361.2022.9956195">10.1109/icpr56361.2022.9956195</a>.
  short: P. Tomaszewska, C. Lampert, in:, 26th International Conference on Pattern
    Recognition, Institute of Electrical and Electronics Engineers, 2022, pp. 2128–2134.
conference:
  end_date: 2022-08-25
  location: Montreal, Canada
  name: 'ICPR: International Conference on Pattern Recognition'
  start_date: 2022-08-21
date_created: 2023-01-12T12:09:38Z
date_published: 2022-11-29T00:00:00Z
date_updated: 2023-08-04T09:06:34Z
day: '29'
department:
- _id: ChLa
doi: 10.1109/icpr56361.2022.9956195
external_id:
  arxiv:
  - '2206.05181'
  isi:
  - '000897707602018'
intvolume: '      2022'
isi: 1
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: https://doi.org/10.48550/arXiv.2206.05181
month: '11'
oa: 1
oa_version: Preprint
page: 2128-2134
publication: 26th International Conference on Pattern Recognition
publication_identifier:
  eisbn:
  - '9781665490627'
  eissn:
  - 2831-7475
publication_status: published
publisher: Institute of Electrical and Electronics Engineers
quality_controlled: '1'
scopus_import: '1'
status: public
title: Lightweight conditional model extrapolation for streaming data under class-prior
  shift
type: conference
user_id: 4359f0d1-fa6c-11eb-b949-802e58b17ae8
volume: 2022
year: '2022'
...
---
_id: '12495'
abstract:
- lang: eng
  text: "Fairness-aware learning aims at constructing classifiers that not only make
    accurate predictions, but also do not discriminate against specific groups. It
    is a fast-growing area of\r\nmachine learning with far-reaching societal impact.
    However, existing fair learning methods\r\nare vulnerable to accidental or malicious
    artifacts in the training data, which can cause\r\nthem to unknowingly produce
    unfair classifiers. In this work we address the problem of\r\nfair learning from
    unreliable training data in the robust multisource setting, where the\r\navailable
    training data comes from multiple sources, a fraction of which might not be representative
    of the true data distribution. We introduce FLEA, a filtering-based algorithm\r\nthat
    identifies and suppresses those data sources that would have a negative impact
    on\r\nfairness or accuracy if they were used for training. As such, FLEA is not
    a replacement of\r\nprior fairness-aware learning methods but rather an augmentation
    that makes any of them\r\nrobust against unreliable training data. We show the
    effectiveness of our approach by a\r\ndiverse range of experiments on multiple
    datasets. Additionally, we prove formally that\r\n–given enough data– FLEA protects
    the learner against corruptions as long as the fraction of\r\naffected data sources
    is less than half. Our source code and documentation are available at\r\nhttps://github.com/ISTAustria-CVML/FLEA."
acknowledged_ssus:
- _id: ScienComp
acknowledgement: 'The authors would like to thank Bernd Prach, Elias Frantar, Alexandra
  Peste, Mahdi Nikdan, and Peter Súkeník for their helpful feedback. This research
  was supported by the Scientific Service Units (SSU) of IST Austria through resources
  provided by Scientific Computing (SciComp). This publication was made possible by
  an ETH AI Center postdoctoral fellowship granted to Nikola Konstantinov. Eugenia
  Iofinova was supported in part by the FWF DK VGSCO, grant agreement number W1260-N35. '
article_processing_charge: No
article_type: original
arxiv: 1
author:
- first_name: Eugenia B
  full_name: Iofinova, Eugenia B
  id: f9a17499-f6e0-11ea-865d-fdf9a3f77117
  last_name: Iofinova
  orcid: 0000-0002-7778-3221
- first_name: Nikola H
  full_name: Konstantinov, Nikola H
  id: 4B9D76E4-F248-11E8-B48F-1D18A9856A87
  last_name: Konstantinov
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
citation:
  ama: 'Iofinova EB, Konstantinov NH, Lampert C. FLEA: Provably robust fair multisource
    learning from unreliable training data. <i>Transactions on Machine Learning Research</i>.
    2022.'
  apa: 'Iofinova, E. B., Konstantinov, N. H., &#38; Lampert, C. (2022). FLEA: Provably
    robust fair multisource learning from unreliable training data. <i>Transactions
    on Machine Learning Research</i>. ML Research Press.'
  chicago: 'Iofinova, Eugenia B, Nikola H Konstantinov, and Christoph Lampert. “FLEA:
    Provably Robust Fair Multisource Learning from Unreliable Training Data.” <i>Transactions
    on Machine Learning Research</i>. ML Research Press, 2022.'
  ieee: 'E. B. Iofinova, N. H. Konstantinov, and C. Lampert, “FLEA: Provably robust
    fair multisource learning from unreliable training data,” <i>Transactions on Machine
    Learning Research</i>. ML Research Press, 2022.'
  ista: 'Iofinova EB, Konstantinov NH, Lampert C. 2022. FLEA: Provably robust fair
    multisource learning from unreliable training data. Transactions on Machine Learning
    Research.'
  mla: 'Iofinova, Eugenia B., et al. “FLEA: Provably Robust Fair Multisource Learning
    from Unreliable Training Data.” <i>Transactions on Machine Learning Research</i>,
    ML Research Press, 2022.'
  short: E.B. Iofinova, N.H. Konstantinov, C. Lampert, Transactions on Machine Learning
    Research (2022).
date_created: 2023-02-02T20:29:57Z
date_published: 2022-12-22T00:00:00Z
date_updated: 2023-02-23T10:30:54Z
day: '22'
ddc:
- '000'
department:
- _id: ChLa
external_id:
  arxiv:
  - '2106.11732'
file:
- access_level: open_access
  checksum: 97c8a8470759cab597abb973ca137a3b
  content_type: application/pdf
  creator: dernst
  date_created: 2023-02-23T10:30:04Z
  date_updated: 2023-02-23T10:30:04Z
  file_id: '12673'
  file_name: 2022_TMLR_Iofinova.pdf
  file_size: 1948063
  relation: main_file
  success: 1
file_date_updated: 2023-02-23T10:30:04Z
has_accepted_license: '1'
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: https://openreview.net/forum?id=XsPopigZXV
month: '12'
oa: 1
oa_version: Published Version
project:
- _id: 9B9290DE-BA93-11EA-9121-9846C619BF3A
  grant_number: ' W1260-N35'
  name: Vienna Graduate School on Computational Optimization
publication: Transactions on Machine Learning Research
publication_identifier:
  issn:
  - 2835-8856
publication_status: published
publisher: ML Research Press
quality_controlled: '1'
related_material:
  link:
  - description: source code
    relation: software
    url: https://github.com/ISTAustria-CVML/FLEA
status: public
title: 'FLEA: Provably robust fair multisource learning from unreliable training data'
tmp:
  image: /images/cc_by.png
  legal_code_url: https://creativecommons.org/licenses/by/4.0/legalcode
  name: Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)
  short: CC BY (4.0)
type: journal_article
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
year: '2022'
...
---
_id: '12660'
abstract:
- lang: eng
  text: 'We present Cross-Client Label Propagation(XCLP), a new method for transductive
    federated learning. XCLP estimates a data graph jointly from the data of multiple
    clients and computes labels for the unlabeled data by propagating label information
    across the graph. To avoid clients having to share their data with anyone, XCLP
    employs two cryptographically secure protocols: secure Hamming distance computation
    and secure summation. We demonstrate two distinct applications of XCLP within
    federated learning. In the first, we use it in a one-shot way to predict labels
    for unseen test points. In the second, we use it to repeatedly pseudo-label unlabeled
    training data in a federated semi-supervised setting. Experiments on both real
    federated and standard benchmark datasets show that in both applications XCLP
    achieves higher classification accuracy than alternative approaches.'
article_number: '2210.06434'
article_processing_charge: No
arxiv: 1
author:
- first_name: Jonathan A
  full_name: Scott, Jonathan A
  id: e499926b-f6e0-11ea-865d-9c63db0031e8
  last_name: Scott
- first_name: Michelle X
  full_name: Yeo, Michelle X
  id: 2D82B818-F248-11E8-B48F-1D18A9856A87
  last_name: Yeo
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
citation:
  ama: Scott JA, Yeo MX, Lampert C. Cross-client Label Propagation for transductive
    federated learning. <i>arXiv</i>. doi:<a href="https://doi.org/10.48550/arXiv.2210.06434">10.48550/arXiv.2210.06434</a>
  apa: Scott, J. A., Yeo, M. X., &#38; Lampert, C. (n.d.). Cross-client Label Propagation
    for transductive federated learning. <i>arXiv</i>. <a href="https://doi.org/10.48550/arXiv.2210.06434">https://doi.org/10.48550/arXiv.2210.06434</a>
  chicago: Scott, Jonathan A, Michelle X Yeo, and Christoph Lampert. “Cross-Client
    Label Propagation for Transductive Federated Learning.” <i>ArXiv</i>, n.d. <a
    href="https://doi.org/10.48550/arXiv.2210.06434">https://doi.org/10.48550/arXiv.2210.06434</a>.
  ieee: J. A. Scott, M. X. Yeo, and C. Lampert, “Cross-client Label Propagation for
    transductive federated learning,” <i>arXiv</i>. .
  ista: Scott JA, Yeo MX, Lampert C. Cross-client Label Propagation for transductive
    federated learning. arXiv, 2210.06434.
  mla: Scott, Jonathan A., et al. “Cross-Client Label Propagation for Transductive
    Federated Learning.” <i>ArXiv</i>, 2210.06434, doi:<a href="https://doi.org/10.48550/arXiv.2210.06434">10.48550/arXiv.2210.06434</a>.
  short: J.A. Scott, M.X. Yeo, C. Lampert, ArXiv (n.d.).
date_created: 2023-02-20T08:21:50Z
date_published: 2022-10-12T00:00:00Z
date_updated: 2023-02-21T08:20:18Z
day: '12'
ddc:
- '004'
department:
- _id: ChLa
doi: 10.48550/arXiv.2210.06434
external_id:
  arxiv:
  - '2210.06434'
file:
- access_level: open_access
  checksum: 7ab20543fd4393f14fb857ce2e4f03c6
  content_type: application/pdf
  creator: chl
  date_created: 2023-02-20T08:21:35Z
  date_updated: 2023-02-20T08:21:35Z
  file_id: '12661'
  file_name: 2210.06434.pdf
  file_size: 291893
  relation: main_file
  success: 1
file_date_updated: 2023-02-20T08:21:35Z
has_accepted_license: '1'
language:
- iso: eng
month: '10'
oa: 1
oa_version: Preprint
publication: arXiv
publication_status: submitted
status: public
title: Cross-client Label Propagation for transductive federated learning
tmp:
  image: /images/cc_by.png
  legal_code_url: https://creativecommons.org/licenses/by/4.0/legalcode
  name: Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)
  short: CC BY (4.0)
type: preprint
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
year: '2022'
...
---
_id: '12662'
abstract:
- lang: eng
  text: 'Modern machine learning tasks often require considering not just one but
    multiple objectives. For example, besides the prediction quality, this could be
    the efficiency, robustness or fairness of the learned models, or any of their
    combinations. Multi-objective learning offers a natural framework for handling
    such problems without having to commit to early trade-offs. Surprisingly, statistical
    learning theory so far offers almost no insight into the generalization properties
    of multi-objective learning. In this work, we make first steps to fill this gap:
    we establish foundational generalization bounds for the multi-objective setting
    as well as generalization and excess bounds for learning with scalarizations.
    We also provide the first theoretical analysis of the relation between the Pareto-optimal
    sets of the true objectives and the Pareto-optimal sets of their empirical approximations
    from training data. In particular, we show a surprising asymmetry: all Pareto-optimal
    solutions can be approximated by empirically Pareto-optimal ones, but not vice
    versa.'
article_number: '2208.13499'
article_processing_charge: No
arxiv: 1
author:
- first_name: Peter
  full_name: Súkeník, Peter
  id: d64d6a8d-eb8e-11eb-b029-96fd216dec3c
  last_name: Súkeník
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
citation:
  ama: Súkeník P, Lampert C. Generalization in Multi-objective machine learning. <i>arXiv</i>.
    doi:<a href="https://doi.org/10.48550/arXiv.2208.13499">10.48550/arXiv.2208.13499</a>
  apa: Súkeník, P., &#38; Lampert, C. (n.d.). Generalization in Multi-objective machine
    learning. <i>arXiv</i>. <a href="https://doi.org/10.48550/arXiv.2208.13499">https://doi.org/10.48550/arXiv.2208.13499</a>
  chicago: Súkeník, Peter, and Christoph Lampert. “Generalization in Multi-Objective
    Machine Learning.” <i>ArXiv</i>, n.d. <a href="https://doi.org/10.48550/arXiv.2208.13499">https://doi.org/10.48550/arXiv.2208.13499</a>.
  ieee: P. Súkeník and C. Lampert, “Generalization in Multi-objective machine learning,”
    <i>arXiv</i>. .
  ista: Súkeník P, Lampert C. Generalization in Multi-objective machine learning.
    arXiv, 2208.13499.
  mla: Súkeník, Peter, and Christoph Lampert. “Generalization in Multi-Objective Machine
    Learning.” <i>ArXiv</i>, 2208.13499, doi:<a href="https://doi.org/10.48550/arXiv.2208.13499">10.48550/arXiv.2208.13499</a>.
  short: P. Súkeník, C. Lampert, ArXiv (n.d.).
date_created: 2023-02-20T08:23:06Z
date_published: 2022-08-29T00:00:00Z
date_updated: 2023-02-21T08:24:55Z
day: '29'
ddc:
- '004'
department:
- _id: ChLa
doi: 10.48550/arXiv.2208.13499
external_id:
  arxiv:
  - '2208.13499'
has_accepted_license: '1'
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: ' https://doi.org/10.48550/arXiv.2208.13499'
month: '08'
oa: 1
oa_version: Preprint
publication: arXiv
publication_status: submitted
status: public
title: Generalization in Multi-objective machine learning
type: preprint
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
year: '2022'
...
---
_id: '10803'
abstract:
- lang: eng
  text: Given the abundance of applications of ranking in recent years, addressing
    fairness concerns around automated ranking systems becomes necessary for increasing
    the trust among end-users. Previous work on fair ranking has mostly focused on
    application-specific fairness notions, often tailored to online advertising, and
    it rarely considers learning as part of the process. In this work, we show how
    to transfer numerous fairness notions from binary classification to a learning
    to rank setting. Our formalism allows us to design methods for incorporating fairness
    objectives with provable generalization guarantees. An extensive experimental
    evaluation shows that our method can improve ranking fairness substantially with
    no or only little loss of model quality.
article_number: '2102.05996'
article_processing_charge: No
arxiv: 1
author:
- first_name: Nikola H
  full_name: Konstantinov, Nikola H
  id: 4B9D76E4-F248-11E8-B48F-1D18A9856A87
  last_name: Konstantinov
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0002-4561-241X
citation:
  ama: Konstantinov NH, Lampert C. Fairness through regularization for learning to
    rank. <i>arXiv</i>. doi:<a href="https://doi.org/10.48550/arXiv.2102.05996">10.48550/arXiv.2102.05996</a>
  apa: Konstantinov, N. H., &#38; Lampert, C. (n.d.). Fairness through regularization
    for learning to rank. <i>arXiv</i>. <a href="https://doi.org/10.48550/arXiv.2102.05996">https://doi.org/10.48550/arXiv.2102.05996</a>
  chicago: Konstantinov, Nikola H, and Christoph Lampert. “Fairness through Regularization
    for Learning to Rank.” <i>ArXiv</i>, n.d. <a href="https://doi.org/10.48550/arXiv.2102.05996">https://doi.org/10.48550/arXiv.2102.05996</a>.
  ieee: N. H. Konstantinov and C. Lampert, “Fairness through regularization for learning
    to rank,” <i>arXiv</i>. .
  ista: Konstantinov NH, Lampert C. Fairness through regularization for learning to
    rank. arXiv, 2102.05996.
  mla: Konstantinov, Nikola H., and Christoph Lampert. “Fairness through Regularization
    for Learning to Rank.” <i>ArXiv</i>, 2102.05996, doi:<a href="https://doi.org/10.48550/arXiv.2102.05996">10.48550/arXiv.2102.05996</a>.
  short: N.H. Konstantinov, C. Lampert, ArXiv (n.d.).
date_created: 2022-02-28T14:13:59Z
date_published: 2021-06-07T00:00:00Z
date_updated: 2023-09-07T13:42:08Z
day: '07'
department:
- _id: ChLa
doi: 10.48550/arXiv.2102.05996
external_id:
  arxiv:
  - '2102.05996'
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: https://arxiv.org/abs/2102.05996
month: '06'
oa: 1
oa_version: Preprint
publication: arXiv
publication_status: submitted
related_material:
  record:
  - id: '10799'
    relation: dissertation_contains
    status: public
status: public
title: Fairness through regularization for learning to rank
type: preprint
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
year: '2021'
...
---
_id: '14987'
abstract:
- lang: eng
  text: "The goal of zero-shot learning is to construct a classifier that can identify
    object classes for which no training examples are available. When training data
    for some of the object classes is available but not for others, the name generalized
    zero-shot learning is commonly used.\r\nIn a wider sense, the phrase zero-shot
    is also used to describe other machine learning-based approaches that require
    no training data from the problem of interest, such as zero-shot action recognition
    or zero-shot machine translation."
article_processing_charge: No
author:
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
citation:
  ama: 'Lampert C. Zero-Shot Learning. In: Ikeuchi K, ed. <i>Computer Vision</i>.
    2nd ed. Cham: Springer; 2021:1395-1397. doi:<a href="https://doi.org/10.1007/978-3-030-63416-2_874">10.1007/978-3-030-63416-2_874</a>'
  apa: 'Lampert, C. (2021). Zero-Shot Learning. In K. Ikeuchi (Ed.), <i>Computer Vision</i>
    (2nd ed., pp. 1395–1397). Cham: Springer. <a href="https://doi.org/10.1007/978-3-030-63416-2_874">https://doi.org/10.1007/978-3-030-63416-2_874</a>'
  chicago: 'Lampert, Christoph. “Zero-Shot Learning.” In <i>Computer Vision</i>, edited
    by Katsushi Ikeuchi, 2nd ed., 1395–97. Cham: Springer, 2021. <a href="https://doi.org/10.1007/978-3-030-63416-2_874">https://doi.org/10.1007/978-3-030-63416-2_874</a>.'
  ieee: 'C. Lampert, “Zero-Shot Learning,” in <i>Computer Vision</i>, 2nd ed., K.
    Ikeuchi, Ed. Cham: Springer, 2021, pp. 1395–1397.'
  ista: 'Lampert C. 2021.Zero-Shot Learning. In: Computer Vision. , 1395–1397.'
  mla: Lampert, Christoph. “Zero-Shot Learning.” <i>Computer Vision</i>, edited by
    Katsushi Ikeuchi, 2nd ed., Springer, 2021, pp. 1395–97, doi:<a href="https://doi.org/10.1007/978-3-030-63416-2_874">10.1007/978-3-030-63416-2_874</a>.
  short: C. Lampert, in:, K. Ikeuchi (Ed.), Computer Vision, 2nd ed., Springer, Cham,
    2021, pp. 1395–1397.
date_created: 2024-02-14T14:05:32Z
date_published: 2021-10-13T00:00:00Z
date_updated: 2024-02-19T10:59:04Z
day: '13'
department:
- _id: ChLa
doi: 10.1007/978-3-030-63416-2_874
edition: '2'
editor:
- first_name: Katsushi
  full_name: Ikeuchi, Katsushi
  last_name: Ikeuchi
language:
- iso: eng
month: '10'
oa_version: None
page: 1395-1397
place: Cham
publication: Computer Vision
publication_identifier:
  eisbn:
  - '9783030634162'
  isbn:
  - '9783030634155'
publication_status: published
publisher: Springer
quality_controlled: '1'
status: public
title: Zero-Shot Learning
type: book_chapter
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
year: '2021'
...
---
_id: '9210'
abstract:
- lang: eng
  text: "Modern neural networks can easily fit their training set perfectly. Surprisingly,
    despite being “overfit” in this way, they tend to generalize well to future data,
    thereby defying the classic bias–variance trade-off of machine learning theory.
    Of the many possible explanations, a prevalent one is that training by stochastic
    gradient descent (SGD) imposes an implicit bias that leads it to learn simple
    functions, and these simple functions generalize well. However, the specifics
    of this implicit bias are not well understood.\r\nIn this work, we explore the
    smoothness conjecture which states that SGD is implicitly biased towards learning
    functions that are smooth. We propose several measures to formalize the intuitive
    notion of smoothness, and we conduct experiments to determine whether SGD indeed
    implicitly optimizes for these measures. Our findings rule out the possibility
    that smoothness measures based on first-order derivatives are being implicitly
    enforced. They are supportive, though, of the smoothness conjecture for measures
    based on second-order derivatives."
article_processing_charge: No
author:
- first_name: Vaclav
  full_name: Volhejn, Vaclav
  id: d5235fb4-7a6d-11eb-b254-f25d12d631a8
  last_name: Volhejn
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
citation:
  ama: 'Volhejn V, Lampert C. Does SGD implicitly optimize for smoothness? In: <i>42nd
    German Conference on Pattern Recognition</i>. Vol 12544. LNCS. Springer; 2021:246-259.
    doi:<a href="https://doi.org/10.1007/978-3-030-71278-5_18">10.1007/978-3-030-71278-5_18</a>'
  apa: 'Volhejn, V., &#38; Lampert, C. (2021). Does SGD implicitly optimize for smoothness?
    In <i>42nd German Conference on Pattern Recognition</i> (Vol. 12544, pp. 246–259).
    Tübingen, Germany: Springer. <a href="https://doi.org/10.1007/978-3-030-71278-5_18">https://doi.org/10.1007/978-3-030-71278-5_18</a>'
  chicago: Volhejn, Vaclav, and Christoph Lampert. “Does SGD Implicitly Optimize for
    Smoothness?” In <i>42nd German Conference on Pattern Recognition</i>, 12544:246–59.
    LNCS. Springer, 2021. <a href="https://doi.org/10.1007/978-3-030-71278-5_18">https://doi.org/10.1007/978-3-030-71278-5_18</a>.
  ieee: V. Volhejn and C. Lampert, “Does SGD implicitly optimize for smoothness?,”
    in <i>42nd German Conference on Pattern Recognition</i>, Tübingen, Germany, 2021,
    vol. 12544, pp. 246–259.
  ista: 'Volhejn V, Lampert C. 2021. Does SGD implicitly optimize for smoothness?
    42nd German Conference on Pattern Recognition. DAGM GCPR: German Conference on
    Pattern Recognition LNCS vol. 12544, 246–259.'
  mla: Volhejn, Vaclav, and Christoph Lampert. “Does SGD Implicitly Optimize for Smoothness?”
    <i>42nd German Conference on Pattern Recognition</i>, vol. 12544, Springer, 2021,
    pp. 246–59, doi:<a href="https://doi.org/10.1007/978-3-030-71278-5_18">10.1007/978-3-030-71278-5_18</a>.
  short: V. Volhejn, C. Lampert, in:, 42nd German Conference on Pattern Recognition,
    Springer, 2021, pp. 246–259.
conference:
  end_date: 2020-10-01
  location: Tübingen, Germany
  name: 'DAGM GCPR: German Conference on Pattern Recognition '
  start_date: 2020-09-28
date_created: 2021-03-01T09:01:16Z
date_published: 2021-03-17T00:00:00Z
date_updated: 2022-08-12T07:28:47Z
day: '17'
ddc:
- '510'
department:
- _id: ChLa
doi: 10.1007/978-3-030-71278-5_18
file:
- access_level: open_access
  checksum: 3e3628ab1cf658d82524963f808004ea
  content_type: application/pdf
  creator: dernst
  date_created: 2022-08-12T07:27:58Z
  date_updated: 2022-08-12T07:27:58Z
  file_id: '11820'
  file_name: 2020_GCPR_submitted_Volhejn.pdf
  file_size: 420234
  relation: main_file
  success: 1
file_date_updated: 2022-08-12T07:27:58Z
has_accepted_license: '1'
intvolume: '     12544'
language:
- iso: eng
month: '03'
oa: 1
oa_version: Submitted Version
page: 246-259
publication: 42nd German Conference on Pattern Recognition
publication_identifier:
  eissn:
  - 1611-3349
  isbn:
  - '9783030712778'
  issn:
  - 0302-9743
publication_status: published
publisher: Springer
quality_controlled: '1'
scopus_import: '1'
series_title: LNCS
status: public
title: Does SGD implicitly optimize for smoothness?
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 12544
year: '2021'
...
---
_id: '9416'
abstract:
- lang: eng
  text: 'We study the inductive bias of two-layer ReLU networks trained by gradient
    flow. We identify a class of easy-to-learn (`orthogonally separable'') datasets,
    and characterise the solution that ReLU networks trained on such datasets converge
    to. Irrespective of network width, the solution turns out to be a combination
    of two max-margin classifiers: one corresponding to the positive data subset and
    one corresponding to the negative data subset. The proof is based on the recently
    introduced concept of extremal sectors, for which we prove a number of properties
    in the context of orthogonal separability. In particular, we prove stationarity
    of activation patterns from some time  onwards, which enables a reduction of the
    ReLU network to an ensemble of linear subnetworks.'
article_processing_charge: No
author:
- first_name: Phuong
  full_name: Bui Thi Mai, Phuong
  id: 3EC6EE64-F248-11E8-B48F-1D18A9856A87
  last_name: Bui Thi Mai
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
citation:
  ama: 'Phuong M, Lampert C. The inductive bias of ReLU networks on orthogonally separable
    data. In: <i>9th International Conference on Learning Representations</i>. ; 2021.'
  apa: Phuong, M., &#38; Lampert, C. (2021). The inductive bias of ReLU networks on
    orthogonally separable data. In <i>9th International Conference on Learning Representations</i>.
    Virtual.
  chicago: Phuong, Mary, and Christoph Lampert. “The Inductive Bias of ReLU Networks
    on Orthogonally Separable Data.” In <i>9th International Conference on Learning
    Representations</i>, 2021.
  ieee: M. Phuong and C. Lampert, “The inductive bias of ReLU networks on orthogonally
    separable data,” in <i>9th International Conference on Learning Representations</i>,
    Virtual, 2021.
  ista: 'Phuong M, Lampert C. 2021. The inductive bias of ReLU networks on orthogonally
    separable data. 9th International Conference on Learning Representations.  ICLR:
    International Conference on Learning Representations.'
  mla: Phuong, Mary, and Christoph Lampert. “The Inductive Bias of ReLU Networks on
    Orthogonally Separable Data.” <i>9th International Conference on Learning Representations</i>,
    2021.
  short: M. Phuong, C. Lampert, in:, 9th International Conference on Learning Representations,
    2021.
conference:
  end_date: 2021-05-07
  location: Virtual
  name: ' ICLR: International Conference on Learning Representations'
  start_date: 2021-05-03
date_created: 2021-05-24T11:16:46Z
date_published: 2021-05-01T00:00:00Z
date_updated: 2023-09-07T13:29:50Z
day: '01'
ddc:
- '000'
department:
- _id: GradSch
- _id: ChLa
file:
- access_level: open_access
  checksum: f34ff17017527db5ba6927f817bdd125
  content_type: application/pdf
  creator: bphuong
  date_created: 2021-05-24T11:15:57Z
  date_updated: 2021-05-24T11:15:57Z
  file_id: '9417'
  file_name: iclr2021_conference.pdf
  file_size: 502356
  relation: main_file
file_date_updated: 2021-05-24T11:15:57Z
has_accepted_license: '1'
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: https://openreview.net/pdf?id=krz7T0xU9Z_
month: '05'
oa: 1
oa_version: Published Version
publication: 9th International Conference on Learning Representations
publication_status: published
quality_controlled: '1'
related_material:
  record:
  - id: '9418'
    relation: dissertation_contains
    status: public
scopus_import: '1'
status: public
title: The inductive bias of ReLU networks on orthogonally separable data
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
year: '2021'
...
---
_id: '7936'
abstract:
- lang: eng
  text: 'State-of-the-art detection systems are generally evaluated on their ability
    to exhaustively retrieve objects densely distributed in the image, across a wide
    variety of appearances and semantic categories. Orthogonal to this, many real-life
    object detection applications, for example in remote sensing, instead require
    dealing with large images that contain only a few small objects of a single class,
    scattered heterogeneously across the space. In addition, they are often subject
    to strict computational constraints, such as limited battery capacity and computing
    power.To tackle these more practical scenarios, we propose a novel flexible detection
    scheme that efficiently adapts to variable object sizes and densities: We rely
    on a sequence of detection stages, each of which has the ability to predict groups
    of objects as well as individuals. Similar to a detection cascade, this multi-stage
    architecture spares computational effort by discarding large irrelevant regions
    of the image early during the detection process. The ability to group objects
    provides further computational and memory savings, as it allows working with lower
    image resolutions in early stages, where groups are more easily detected than
    individuals, as they are more salient. We report experimental results on two aerial
    image datasets, and show that the proposed method is as accurate yet computationally
    more efficient than standard single-shot detectors, consistently across three
    different backbone architectures.'
article_number: 1716-1725
article_processing_charge: No
arxiv: 1
author:
- first_name: Amélie
  full_name: Royer, Amélie
  id: 3811D890-F248-11E8-B48F-1D18A9856A87
  last_name: Royer
  orcid: 0000-0002-8407-0705
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
citation:
  ama: 'Royer A, Lampert C. Localizing grouped instances for efficient detection in
    low-resource scenarios. In: <i>IEEE Winter Conference on Applications of Computer
    Vision</i>. IEEE; 2020. doi:<a href="https://doi.org/10.1109/WACV45572.2020.9093288">10.1109/WACV45572.2020.9093288</a>'
  apa: 'Royer, A., &#38; Lampert, C. (2020). Localizing grouped instances for efficient
    detection in low-resource scenarios. In <i>IEEE Winter Conference on Applications
    of Computer Vision</i>.  Snowmass Village, CO, United States: IEEE. <a href="https://doi.org/10.1109/WACV45572.2020.9093288">https://doi.org/10.1109/WACV45572.2020.9093288</a>'
  chicago: Royer, Amélie, and Christoph Lampert. “Localizing Grouped Instances for
    Efficient Detection in Low-Resource Scenarios.” In <i>IEEE Winter Conference on
    Applications of Computer Vision</i>. IEEE, 2020. <a href="https://doi.org/10.1109/WACV45572.2020.9093288">https://doi.org/10.1109/WACV45572.2020.9093288</a>.
  ieee: A. Royer and C. Lampert, “Localizing grouped instances for efficient detection
    in low-resource scenarios,” in <i>IEEE Winter Conference on Applications of Computer
    Vision</i>,  Snowmass Village, CO, United States, 2020.
  ista: 'Royer A, Lampert C. 2020. Localizing grouped instances for efficient detection
    in low-resource scenarios. IEEE Winter Conference on Applications of Computer
    Vision. WACV: Winter Conference on Applications of Computer Vision, 1716–1725.'
  mla: Royer, Amélie, and Christoph Lampert. “Localizing Grouped Instances for Efficient
    Detection in Low-Resource Scenarios.” <i>IEEE Winter Conference on Applications
    of Computer Vision</i>, 1716–1725, IEEE, 2020, doi:<a href="https://doi.org/10.1109/WACV45572.2020.9093288">10.1109/WACV45572.2020.9093288</a>.
  short: A. Royer, C. Lampert, in:, IEEE Winter Conference on Applications of Computer
    Vision, IEEE, 2020.
conference:
  end_date: 2020-03-05
  location: ' Snowmass Village, CO, United States'
  name: 'WACV: Winter Conference on Applications of Computer Vision'
  start_date: 2020-03-01
date_created: 2020-06-07T22:00:53Z
date_published: 2020-03-01T00:00:00Z
date_updated: 2023-09-07T13:16:17Z
day: '01'
department:
- _id: ChLa
doi: 10.1109/WACV45572.2020.9093288
external_id:
  arxiv:
  - '2004.12623'
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: https://arxiv.org/abs/2004.12623
month: '03'
oa: 1
oa_version: Preprint
publication: IEEE Winter Conference on Applications of Computer Vision
publication_identifier:
  isbn:
  - '9781728165530'
publication_status: published
publisher: IEEE
quality_controlled: '1'
related_material:
  record:
  - id: '8331'
    relation: dissertation_contains
    status: deleted
  - id: '8390'
    relation: dissertation_contains
    status: public
scopus_import: 1
status: public
title: Localizing grouped instances for efficient detection in low-resource scenarios
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
year: '2020'
...
---
_id: '7937'
abstract:
- lang: eng
  text: 'Fine-tuning is a popular way of exploiting knowledge contained in a pre-trained
    convolutional network for a new visual recognition task. However, the orthogonal
    setting of transferring knowledge from a pretrained network to a visually different
    yet semantically close source is rarely considered: This commonly happens with
    real-life data, which is not necessarily as clean as the training source (noise,
    geometric transformations, different modalities, etc.).To tackle such scenarios,
    we introduce a new, generalized form of fine-tuning, called flex-tuning, in which
    any individual unit (e.g. layer) of a network can be tuned, and the most promising
    one is chosen automatically. In order to make the method appealing for practical
    use, we propose two lightweight and faster selection procedures that prove to
    be good approximations in practice. We study these selection criteria empirically
    across a variety of domain shifts and data scarcity scenarios, and show that fine-tuning
    individual units, despite its simplicity, yields very good results as an adaptation
    technique. As it turns out, in contrast to common practice, rather than the last
    fully-connected unit it is best to tune an intermediate or early one in many domain-
    shift scenarios, which is accurately detected by flex-tuning.'
article_number: 2180-2189
article_processing_charge: No
arxiv: 1
author:
- first_name: Amélie
  full_name: Royer, Amélie
  id: 3811D890-F248-11E8-B48F-1D18A9856A87
  last_name: Royer
  orcid: 0000-0002-8407-0705
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
citation:
  ama: 'Royer A, Lampert C. A flexible selection scheme for minimum-effort transfer
    learning. In: <i>2020 IEEE Winter Conference on Applications of Computer Vision</i>.
    IEEE; 2020. doi:<a href="https://doi.org/10.1109/WACV45572.2020.9093635">10.1109/WACV45572.2020.9093635</a>'
  apa: 'Royer, A., &#38; Lampert, C. (2020). A flexible selection scheme for minimum-effort
    transfer learning. In <i>2020 IEEE Winter Conference on Applications of Computer
    Vision</i>. Snowmass Village, CO, United States: IEEE. <a href="https://doi.org/10.1109/WACV45572.2020.9093635">https://doi.org/10.1109/WACV45572.2020.9093635</a>'
  chicago: Royer, Amélie, and Christoph Lampert. “A Flexible Selection Scheme for
    Minimum-Effort Transfer Learning.” In <i>2020 IEEE Winter Conference on Applications
    of Computer Vision</i>. IEEE, 2020. <a href="https://doi.org/10.1109/WACV45572.2020.9093635">https://doi.org/10.1109/WACV45572.2020.9093635</a>.
  ieee: A. Royer and C. Lampert, “A flexible selection scheme for minimum-effort transfer
    learning,” in <i>2020 IEEE Winter Conference on Applications of Computer Vision</i>,
    Snowmass Village, CO, United States, 2020.
  ista: 'Royer A, Lampert C. 2020. A flexible selection scheme for minimum-effort
    transfer learning. 2020 IEEE Winter Conference on Applications of Computer Vision.
    WACV: Winter Conference on Applications of Computer Vision, 2180–2189.'
  mla: Royer, Amélie, and Christoph Lampert. “A Flexible Selection Scheme for Minimum-Effort
    Transfer Learning.” <i>2020 IEEE Winter Conference on Applications of Computer
    Vision</i>, 2180–2189, IEEE, 2020, doi:<a href="https://doi.org/10.1109/WACV45572.2020.9093635">10.1109/WACV45572.2020.9093635</a>.
  short: A. Royer, C. Lampert, in:, 2020 IEEE Winter Conference on Applications of
    Computer Vision, IEEE, 2020.
conference:
  end_date: 2020-03-05
  location: Snowmass Village, CO, United States
  name: 'WACV: Winter Conference on Applications of Computer Vision'
  start_date: 2020-03-01
date_created: 2020-06-07T22:00:53Z
date_published: 2020-03-01T00:00:00Z
date_updated: 2023-09-07T13:16:17Z
day: '01'
department:
- _id: ChLa
doi: 10.1109/WACV45572.2020.9093635
external_id:
  arxiv:
  - '2008.11995'
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: http://arxiv.org/abs/2008.11995
month: '03'
oa: 1
oa_version: Preprint
publication: 2020 IEEE Winter Conference on Applications of Computer Vision
publication_identifier:
  isbn:
  - '9781728165530'
publication_status: published
publisher: IEEE
quality_controlled: '1'
related_material:
  record:
  - id: '8331'
    relation: dissertation_contains
    status: deleted
  - id: '8390'
    relation: dissertation_contains
    status: public
scopus_import: '1'
status: public
title: A flexible selection scheme for minimum-effort transfer learning
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
year: '2020'
...
---
_id: '8063'
abstract:
- lang: eng
  text: "We present a generative model of images that explicitly reasons over the
    set\r\nof objects they show. Our model learns a structured latent representation
    that\r\nseparates objects from each other and from the background; unlike prior
    works,\r\nit explicitly represents the 2D position and depth of each object, as
    well as\r\nan embedding of its segmentation mask and appearance. The model can
    be trained\r\nfrom images alone in a purely unsupervised fashion without the need
    for object\r\nmasks or depth information. Moreover, it always generates complete
    objects,\r\neven though a significant fraction of training images contain occlusions.\r\nFinally,
    we show that our model can infer decompositions of novel images into\r\ntheir
    constituent objects, including accurate prediction of depth ordering and\r\nsegmentation
    of occluded parts."
article_number: '2004.00642'
article_processing_charge: No
arxiv: 1
author:
- first_name: Titas
  full_name: Anciukevicius, Titas
  last_name: Anciukevicius
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
- first_name: Paul M
  full_name: Henderson, Paul M
  id: 13C09E74-18D9-11E9-8878-32CFE5697425
  last_name: Henderson
  orcid: 0000-0002-5198-7445
citation:
  ama: Anciukevicius T, Lampert C, Henderson PM. Object-centric image generation with
    factored depths, locations, and appearances. <i>arXiv</i>.
  apa: Anciukevicius, T., Lampert, C., &#38; Henderson, P. M. (n.d.). Object-centric
    image generation with factored depths, locations, and appearances. <i>arXiv</i>.
  chicago: Anciukevicius, Titas, Christoph Lampert, and Paul M Henderson. “Object-Centric
    Image Generation with Factored Depths, Locations, and Appearances.” <i>ArXiv</i>,
    n.d.
  ieee: T. Anciukevicius, C. Lampert, and P. M. Henderson, “Object-centric image generation
    with factored depths, locations, and appearances,” <i>arXiv</i>. .
  ista: Anciukevicius T, Lampert C, Henderson PM. Object-centric image generation
    with factored depths, locations, and appearances. arXiv, 2004.00642.
  mla: Anciukevicius, Titas, et al. “Object-Centric Image Generation with Factored
    Depths, Locations, and Appearances.” <i>ArXiv</i>, 2004.00642.
  short: T. Anciukevicius, C. Lampert, P.M. Henderson, ArXiv (n.d.).
date_created: 2020-06-29T23:55:23Z
date_published: 2020-04-01T00:00:00Z
date_updated: 2021-01-12T08:16:44Z
day: '01'
ddc:
- '004'
department:
- _id: ChLa
external_id:
  arxiv:
  - '2004.00642'
language:
- iso: eng
license: https://creativecommons.org/licenses/by-sa/4.0/
main_file_link:
- open_access: '1'
  url: https://arxiv.org/abs/2004.00642
month: '04'
oa: 1
oa_version: Preprint
publication: arXiv
publication_status: submitted
status: public
title: Object-centric image generation with factored depths, locations, and appearances
tmp:
  image: /images/cc_by_sa.png
  legal_code_url: https://creativecommons.org/licenses/by-sa/4.0/legalcode
  name: Creative Commons Attribution-ShareAlike 4.0 International Public License (CC
    BY-SA 4.0)
  short: CC BY-SA (4.0)
type: preprint
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
year: '2020'
...
---
_id: '8186'
abstract:
- lang: eng
  text: "Numerous methods have been proposed for probabilistic generative modelling
    of\r\n3D objects. However, none of these is able to produce textured objects,
    which\r\nrenders them of limited use for practical tasks. In this work, we present
    the\r\nfirst generative model of textured 3D meshes. Training such a model would\r\ntraditionally
    require a large dataset of textured meshes, but unfortunately,\r\nexisting datasets
    of meshes lack detailed textures. We instead propose a new\r\ntraining methodology
    that allows learning from collections of 2D images without\r\nany 3D information.
    To do so, we train our model to explain a distribution of\r\nimages by modelling
    each image as a 3D foreground object placed in front of a\r\n2D background. Thus,
    it learns to generate meshes that when rendered, produce\r\nimages similar to
    those in its training set.\r\n  A well-known problem when generating meshes with
    deep networks is the\r\nemergence of self-intersections, which are problematic
    for many use-cases. As a\r\nsecond contribution we therefore introduce a new generation
    process for 3D\r\nmeshes that guarantees no self-intersections arise, based on
    the physical\r\nintuition that faces should push one another out of the way as
    they move.\r\n  We conduct extensive experiments on our approach, reporting quantitative
    and\r\nqualitative results on both synthetic data and natural images. These show
    our\r\nmethod successfully learns to generate plausible and diverse textured 3D\r\nsamples
    for five challenging object classes."
article_processing_charge: No
arxiv: 1
author:
- first_name: Paul M
  full_name: Henderson, Paul M
  id: 13C09E74-18D9-11E9-8878-32CFE5697425
  last_name: Henderson
  orcid: 0000-0002-5198-7445
- first_name: Vagia
  full_name: Tsiminaki, Vagia
  last_name: Tsiminaki
- first_name: Christoph
  full_name: Lampert, Christoph
  id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
  last_name: Lampert
  orcid: 0000-0001-8622-7887
citation:
  ama: 'Henderson PM, Tsiminaki V, Lampert C. Leveraging 2D data to learn textured
    3D mesh generation. In: <i>Proceedings of the IEEE/CVF Conference on Computer
    Vision and Pattern Recognition</i>. IEEE; 2020:7498-7507. doi:<a href="https://doi.org/10.1109/CVPR42600.2020.00752">10.1109/CVPR42600.2020.00752</a>'
  apa: 'Henderson, P. M., Tsiminaki, V., &#38; Lampert, C. (2020). Leveraging 2D data
    to learn textured 3D mesh generation. In <i>Proceedings of the IEEE/CVF Conference
    on Computer Vision and Pattern Recognition</i> (pp. 7498–7507). Virtual: IEEE.
    <a href="https://doi.org/10.1109/CVPR42600.2020.00752">https://doi.org/10.1109/CVPR42600.2020.00752</a>'
  chicago: Henderson, Paul M, Vagia Tsiminaki, and Christoph Lampert. “Leveraging
    2D Data to Learn Textured 3D Mesh Generation.” In <i>Proceedings of the IEEE/CVF
    Conference on Computer Vision and Pattern Recognition</i>, 7498–7507. IEEE, 2020.
    <a href="https://doi.org/10.1109/CVPR42600.2020.00752">https://doi.org/10.1109/CVPR42600.2020.00752</a>.
  ieee: P. M. Henderson, V. Tsiminaki, and C. Lampert, “Leveraging 2D data to learn
    textured 3D mesh generation,” in <i>Proceedings of the IEEE/CVF Conference on
    Computer Vision and Pattern Recognition</i>, Virtual, 2020, pp. 7498–7507.
  ista: 'Henderson PM, Tsiminaki V, Lampert C. 2020. Leveraging 2D data to learn textured
    3D mesh generation. Proceedings of the IEEE/CVF Conference on Computer Vision
    and Pattern Recognition. CVPR: Conference on Computer Vision and Pattern Recognition,
    7498–7507.'
  mla: Henderson, Paul M., et al. “Leveraging 2D Data to Learn Textured 3D Mesh Generation.”
    <i>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</i>,
    IEEE, 2020, pp. 7498–507, doi:<a href="https://doi.org/10.1109/CVPR42600.2020.00752">10.1109/CVPR42600.2020.00752</a>.
  short: P.M. Henderson, V. Tsiminaki, C. Lampert, in:, Proceedings of the IEEE/CVF
    Conference on Computer Vision and Pattern Recognition, IEEE, 2020, pp. 7498–7507.
conference:
  end_date: 2020-06-19
  location: Virtual
  name: 'CVPR: Conference on Computer Vision and Pattern Recognition'
  start_date: 2020-06-14
date_created: 2020-07-31T16:53:49Z
date_published: 2020-07-01T00:00:00Z
date_updated: 2023-10-17T07:37:11Z
day: '01'
ddc:
- '004'
department:
- _id: ChLa
doi: 10.1109/CVPR42600.2020.00752
external_id:
  arxiv:
  - '2004.04180'
file:
- access_level: open_access
  content_type: application/pdf
  creator: phenders
  date_created: 2020-07-31T16:57:12Z
  date_updated: 2020-07-31T16:57:12Z
  file_id: '8187'
  file_name: paper.pdf
  file_size: 10262773
  relation: main_file
  success: 1
file_date_updated: 2020-07-31T16:57:12Z
has_accepted_license: '1'
language:
- iso: eng
main_file_link:
- open_access: '1'
  url: https://openaccess.thecvf.com/content_CVPR_2020/papers/Henderson_Leveraging_2D_Data_to_Learn_Textured_3D_Mesh_Generation_CVPR_2020_paper.pdf
month: '07'
oa: 1
oa_version: Submitted Version
page: 7498-7507
publication: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern
  Recognition
publication_identifier:
  eisbn:
  - '9781728171685'
  eissn:
  - 2575-7075
publication_status: published
publisher: IEEE
quality_controlled: '1'
scopus_import: '1'
status: public
title: Leveraging 2D data to learn textured 3D mesh generation
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
year: '2020'
...
