[{"language":[{"iso":"eng"}],"oa":1,"user_id":"3E5EF7F0-F248-11E8-B48F-1D18A9856A87","citation":{"mla":"Sun, Rémy, and Christoph Lampert. “KS(Conf): A Light-Weight Test If a Multiclass Classifier Operates Outside of Its Specifications.” <i>International Journal of Computer Vision</i>, vol. 128, no. 4, Springer Nature, 2020, pp. 970–95, doi:<a href=\"https://doi.org/10.1007/s11263-019-01232-x\">10.1007/s11263-019-01232-x</a>.","apa":"Sun, R., &#38; Lampert, C. (2020). KS(conf): A light-weight test if a multiclass classifier operates outside of its specifications. <i>International Journal of Computer Vision</i>. Springer Nature. <a href=\"https://doi.org/10.1007/s11263-019-01232-x\">https://doi.org/10.1007/s11263-019-01232-x</a>","chicago":"Sun, Rémy, and Christoph Lampert. “KS(Conf): A Light-Weight Test If a Multiclass Classifier Operates Outside of Its Specifications.” <i>International Journal of Computer Vision</i>. Springer Nature, 2020. <a href=\"https://doi.org/10.1007/s11263-019-01232-x\">https://doi.org/10.1007/s11263-019-01232-x</a>.","ista":"Sun R, Lampert C. 2020. KS(conf): A light-weight test if a multiclass classifier operates outside of its specifications. International Journal of Computer Vision. 128(4), 970–995.","short":"R. Sun, C. Lampert, International Journal of Computer Vision 128 (2020) 970–995.","ieee":"R. Sun and C. Lampert, “KS(conf): A light-weight test if a multiclass classifier operates outside of its specifications,” <i>International Journal of Computer Vision</i>, vol. 128, no. 4. Springer Nature, pp. 970–995, 2020.","ama":"Sun R, Lampert C. KS(conf): A light-weight test if a multiclass classifier operates outside of its specifications. <i>International Journal of Computer Vision</i>. 2020;128(4):970-995. doi:<a href=\"https://doi.org/10.1007/s11263-019-01232-x\">10.1007/s11263-019-01232-x</a>"},"issue":"4","month":"04","file":[{"file_id":"7110","date_created":"2019-11-26T10:30:02Z","file_size":1715072,"date_updated":"2020-07-14T12:47:45Z","creator":"dernst","relation":"main_file","checksum":"155e63edf664dcacb3bdc1c2223e606f","file_name":"2019_IJCV_Sun.pdf","content_type":"application/pdf","access_level":"open_access"}],"department":[{"_id":"ChLa"}],"intvolume":"       128","tmp":{"name":"Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)","legal_code_url":"https://creativecommons.org/licenses/by/4.0/legalcode","image":"/images/cc_by.png","short":"CC BY (4.0)"},"abstract":[{"lang":"eng","text":"We study the problem of automatically detecting if a given multi-class classifier operates outside of its specifications (out-of-specs), i.e. on input data from a different distribution than what it was trained for. This is an important problem to solve on the road towards creating reliable computer vision systems for real-world applications, because the quality of a classifier’s predictions cannot be guaranteed if it operates out-of-specs. Previously proposed methods for out-of-specs detection make decisions on the level of single inputs. This, however, is insufficient to achieve low false positive rate and high false negative rates at the same time. In this work, we describe a new procedure named KS(conf), based on statistical reasoning. Its main component is a classical Kolmogorov–Smirnov test that is applied to the set of predicted confidence values for batches of samples. Working with batches instead of single samples allows increasing the true positive rate without negatively affecting the false positive rate, thereby overcoming a crucial limitation of single sample tests. We show by extensive experiments using a variety of convolutional network architectures and datasets that KS(conf) reliably detects out-of-specs situations even under conditions where other tests fail. It furthermore has a number of properties that make it an excellent candidate for practical deployment: it is easy to implement, adds almost no overhead to the system, works with any classifier that outputs confidence scores, and requires no a priori knowledge about how the data distribution could change."}],"has_accepted_license":"1","file_date_updated":"2020-07-14T12:47:45Z","publication_status":"published","publication_identifier":{"eissn":["1573-1405"],"issn":["0920-5691"]},"title":"KS(conf): A light-weight test if a multiclass classifier operates outside of its specifications","oa_version":"Published Version","day":"01","scopus_import":"1","author":[{"full_name":"Sun, Rémy","last_name":"Sun","first_name":"Rémy"},{"orcid":"0000-0001-8622-7887","first_name":"Christoph","last_name":"Lampert","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87","full_name":"Lampert, Christoph"}],"date_created":"2019-10-14T09:14:28Z","article_type":"original","volume":128,"publication":"International Journal of Computer Vision","status":"public","project":[{"call_identifier":"FP7","name":"Lifelong Learning of Visual Scene Understanding","grant_number":"308036","_id":"2532554C-B435-11E9-9278-68D0E5697425"},{"name":"IST Austria Open Access Fund","_id":"B67AFEDC-15C9-11EA-A837-991A96BB2854"}],"date_published":"2020-04-01T00:00:00Z","ec_funded":1,"related_material":{"link":[{"relation":"erratum","url":"https://doi.org/10.1007/s11263-019-01262-5"}],"record":[{"id":"6482","relation":"earlier_version","status":"public"}]},"external_id":{"isi":["000494406800001"]},"isi":1,"year":"2020","ddc":["004"],"page":"970-995","quality_controlled":"1","publisher":"Springer Nature","article_processing_charge":"Yes (via OA deal)","doi":"10.1007/s11263-019-01232-x","type":"journal_article","_id":"6944","date_updated":"2024-02-22T14:57:30Z"},{"date_created":"2020-02-11T09:06:57Z","volume":"2019-October","title":"Distillation-based training for multi-exit architectures","oa_version":"Submitted Version","author":[{"first_name":"Phuong","id":"3EC6EE64-F248-11E8-B48F-1D18A9856A87","full_name":"Bui Thi Mai, Phuong","last_name":"Bui Thi Mai"},{"last_name":"Lampert","full_name":"Lampert, Christoph","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87","first_name":"Christoph","orcid":"0000-0001-8622-7887"}],"scopus_import":"1","day":"01","publication_identifier":{"issn":["15505499"],"isbn":["9781728148038"]},"publication_status":"published","file_date_updated":"2020-07-14T12:47:59Z","abstract":[{"lang":"eng","text":"Multi-exit architectures, in which a stack of processing layers is interleaved with early output layers, allow the processing of a test example to stop early and thus save computation time and/or energy.  In this work, we propose a new training procedure for multi-exit architectures based on the principle of knowledge distillation. The method encourage searly exits to mimic later, more accurate exits, by matching their output probabilities.\r\nExperiments  on  CIFAR100  and  ImageNet  show  that distillation-based training significantly improves the accuracy of early exits while maintaining state-of-the-art accuracy  for  late  ones.   The  method  is  particularly  beneficial when  training  data  is  limited  and  it  allows  a  straightforward extension to semi-supervised learning,i.e. making use of unlabeled data at training time. Moreover, it takes only afew lines to implement and incurs almost no computational overhead at training time, and none at all at test time."}],"has_accepted_license":"1","file":[{"access_level":"open_access","content_type":"application/pdf","file_name":"main.pdf","checksum":"7b77fb5c2d27c4c37a7612ba46a66117","relation":"main_file","date_updated":"2020-07-14T12:47:59Z","creator":"bphuong","date_created":"2020-02-11T09:06:39Z","file_size":735768,"file_id":"7480"}],"department":[{"_id":"ChLa"}],"month":"10","user_id":"c635000d-4b10-11ee-a964-aac5a93f6ac1","citation":{"mla":"Phuong, Mary, and Christoph Lampert. “Distillation-Based Training for Multi-Exit Architectures.” <i>IEEE International Conference on Computer Vision</i>, vol. 2019–October, IEEE, 2019, pp. 1355–64, doi:<a href=\"https://doi.org/10.1109/ICCV.2019.00144\">10.1109/ICCV.2019.00144</a>.","apa":"Phuong, M., &#38; Lampert, C. (2019). Distillation-based training for multi-exit architectures. In <i>IEEE International Conference on Computer Vision</i> (Vol. 2019–October, pp. 1355–1364). Seoul, Korea: IEEE. <a href=\"https://doi.org/10.1109/ICCV.2019.00144\">https://doi.org/10.1109/ICCV.2019.00144</a>","chicago":"Phuong, Mary, and Christoph Lampert. “Distillation-Based Training for Multi-Exit Architectures.” In <i>IEEE International Conference on Computer Vision</i>, 2019–October:1355–64. IEEE, 2019. <a href=\"https://doi.org/10.1109/ICCV.2019.00144\">https://doi.org/10.1109/ICCV.2019.00144</a>.","ista":"Phuong M, Lampert C. 2019. Distillation-based training for multi-exit architectures. IEEE International Conference on Computer Vision. ICCV: International Conference on Computer Vision vol. 2019–October, 1355–1364.","short":"M. Phuong, C. Lampert, in:, IEEE International Conference on Computer Vision, IEEE, 2019, pp. 1355–1364.","ieee":"M. Phuong and C. Lampert, “Distillation-based training for multi-exit architectures,” in <i>IEEE International Conference on Computer Vision</i>, Seoul, Korea, 2019, vol. 2019–October, pp. 1355–1364.","ama":"Phuong M, Lampert C. Distillation-based training for multi-exit architectures. In: <i>IEEE International Conference on Computer Vision</i>. Vol 2019-October. IEEE; 2019:1355-1364. doi:<a href=\"https://doi.org/10.1109/ICCV.2019.00144\">10.1109/ICCV.2019.00144</a>"},"language":[{"iso":"eng"}],"oa":1,"type":"conference","date_updated":"2023-09-08T11:11:12Z","_id":"7479","publisher":"IEEE","doi":"10.1109/ICCV.2019.00144","article_processing_charge":"No","quality_controlled":"1","ddc":["000"],"page":"1355-1364","external_id":{"isi":["000531438101047"]},"related_material":{"record":[{"status":"public","relation":"dissertation_contains","id":"9418"}]},"year":"2019","isi":1,"conference":{"start_date":"2019-10-27","end_date":"2019-11-02","name":"ICCV: International Conference on Computer Vision","location":"Seoul, Korea"},"date_published":"2019-10-01T00:00:00Z","ec_funded":1,"project":[{"call_identifier":"FP7","name":"Lifelong Learning of Visual Scene Understanding","grant_number":"308036","_id":"2532554C-B435-11E9-9278-68D0E5697425"}],"status":"public","publication":"IEEE International Conference on Computer Vision"},{"abstract":[{"text":"We propose a new model for detecting visual relationships, such as \"person riding motorcycle\" or \"bottle on table\". This task is an important step towards comprehensive structured mage understanding, going beyond detecting individual objects. Our main novelty is a Box Attention mechanism that allows to model pairwise interactions between objects using standard object detection pipelines. The resulting model is conceptually clean, expressive and relies on well-justified training and prediction procedures. Moreover, unlike previously proposed approaches, our model does not introduce any additional complex components or hyperparameters on top of those already required by the underlying detection model. We conduct an experimental evaluation on two datasets, V-COCO and Open Images, demonstrating strong quantitative and qualitative results.","lang":"eng"}],"publication_identifier":{"isbn":["9781728150239"]},"publication_status":"published","scopus_import":"1","day":"01","author":[{"id":"2D157DB6-F248-11E8-B48F-1D18A9856A87","full_name":"Kolesnikov, Alexander","last_name":"Kolesnikov","first_name":"Alexander"},{"first_name":"Alina","last_name":"Kuznetsova","full_name":"Kuznetsova, Alina"},{"orcid":"0000-0001-8622-7887","first_name":"Christoph","full_name":"Lampert, Christoph","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87","last_name":"Lampert"},{"first_name":"Vittorio","full_name":"Ferrari, Vittorio","last_name":"Ferrari"}],"title":"Detecting visual relationships using box attention","oa_version":"Preprint","date_created":"2020-04-05T22:00:51Z","oa":1,"language":[{"iso":"eng"}],"citation":{"ama":"Kolesnikov A, Kuznetsova A, Lampert C, Ferrari V. Detecting visual relationships using box attention. In: <i>Proceedings of the 2019 International Conference on Computer Vision Workshop</i>. IEEE; 2019. doi:<a href=\"https://doi.org/10.1109/ICCVW.2019.00217\">10.1109/ICCVW.2019.00217</a>","ieee":"A. Kolesnikov, A. Kuznetsova, C. Lampert, and V. Ferrari, “Detecting visual relationships using box attention,” in <i>Proceedings of the 2019 International Conference on Computer Vision Workshop</i>, Seoul, South Korea, 2019.","short":"A. Kolesnikov, A. Kuznetsova, C. Lampert, V. Ferrari, in:, Proceedings of the 2019 International Conference on Computer Vision Workshop, IEEE, 2019.","ista":"Kolesnikov A, Kuznetsova A, Lampert C, Ferrari V. 2019. Detecting visual relationships using box attention. Proceedings of the 2019 International Conference on Computer Vision Workshop. ICCVW: International Conference on Computer Vision Workshop, 1749–1753.","chicago":"Kolesnikov, Alexander, Alina Kuznetsova, Christoph Lampert, and Vittorio Ferrari. “Detecting Visual Relationships Using Box Attention.” In <i>Proceedings of the 2019 International Conference on Computer Vision Workshop</i>. IEEE, 2019. <a href=\"https://doi.org/10.1109/ICCVW.2019.00217\">https://doi.org/10.1109/ICCVW.2019.00217</a>.","apa":"Kolesnikov, A., Kuznetsova, A., Lampert, C., &#38; Ferrari, V. (2019). Detecting visual relationships using box attention. In <i>Proceedings of the 2019 International Conference on Computer Vision Workshop</i>. Seoul, South Korea: IEEE. <a href=\"https://doi.org/10.1109/ICCVW.2019.00217\">https://doi.org/10.1109/ICCVW.2019.00217</a>","mla":"Kolesnikov, Alexander, et al. “Detecting Visual Relationships Using Box Attention.” <i>Proceedings of the 2019 International Conference on Computer Vision Workshop</i>, 1749–1753, IEEE, 2019, doi:<a href=\"https://doi.org/10.1109/ICCVW.2019.00217\">10.1109/ICCVW.2019.00217</a>."},"user_id":"c635000d-4b10-11ee-a964-aac5a93f6ac1","month":"10","arxiv":1,"department":[{"_id":"ChLa"}],"article_number":"1749-1753","main_file_link":[{"url":"https://arxiv.org/abs/1807.02136","open_access":"1"}],"quality_controlled":"1","article_processing_charge":"No","doi":"10.1109/ICCVW.2019.00217","publisher":"IEEE","_id":"7640","date_updated":"2023-09-08T11:18:37Z","type":"conference","status":"public","publication":"Proceedings of the 2019 International Conference on Computer Vision Workshop","project":[{"_id":"2532554C-B435-11E9-9278-68D0E5697425","call_identifier":"FP7","name":"Lifelong Learning of Visual Scene Understanding","grant_number":"308036"}],"ec_funded":1,"conference":{"location":"Seoul, South Korea","end_date":"2019-10-28","start_date":"2019-10-27","name":"ICCVW: International Conference on Computer Vision Workshop"},"date_published":"2019-10-01T00:00:00Z","year":"2019","isi":1,"external_id":{"isi":["000554591601098"],"arxiv":["1807.02136"]}},{"author":[{"last_name":"Sun","full_name":"Sun, Rémy","first_name":"Rémy"},{"first_name":"Christoph","orcid":"0000-0001-8622-7887","last_name":"Lampert","full_name":"Lampert, Christoph","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87"}],"day":"14","scopus_import":"1","title":"KS(conf): A light-weight test if a ConvNet operates outside of Its specifications","oa_version":"Preprint","volume":11269,"date_created":"2019-05-24T09:48:36Z","abstract":[{"text":"Computer vision systems for automatic image categorization have become accurate and reliable enough that they can run continuously for days or even years as components of real-world commercial applications. A major open problem in this context, however, is quality control. Good classification performance can only be expected if systems run under the specific conditions, in particular data distributions, that they were trained for. Surprisingly, none of the currently used deep network architectures have a built-in functionality that could detect if a network operates on data from a distribution it was not trained for, such that potentially a warning to the human users could be triggered. In this work, we describe KS(conf), a procedure for detecting such outside of specifications (out-of-specs) operation, based on statistical testing of the network outputs. We show by extensive experiments using the ImageNet, AwA2 and DAVIS datasets on a variety of ConvNets architectures that KS(conf) reliably detects out-of-specs situations. It furthermore has a number of properties that make it a promising candidate for practical deployment: it is easy to implement, adds almost no overhead to the system, works with all networks, including pretrained ones, and requires no a priori knowledge of how the data distribution could change. ","lang":"eng"}],"intvolume":"     11269","publication_identifier":{"eissn":["1611-3349"],"isbn":["9783030129385","9783030129392"],"issn":["0302-9743"]},"publication_status":"published","month":"02","arxiv":1,"department":[{"_id":"ChLa"}],"oa":1,"language":[{"iso":"eng"}],"citation":{"short":"R. Sun, C. Lampert, in:, Springer Nature, 2019, pp. 244–259.","ieee":"R. Sun and C. Lampert, “KS(conf): A light-weight test if a ConvNet operates outside of Its specifications,” presented at the GCPR: Conference on Pattern Recognition, Stuttgart, Germany, 2019, vol. 11269, pp. 244–259.","ama":"Sun R, Lampert C. KS(conf): A light-weight test if a ConvNet operates outside of Its specifications. In: Vol 11269. Springer Nature; 2019:244-259. doi:<a href=\"https://doi.org/10.1007/978-3-030-12939-2_18\">10.1007/978-3-030-12939-2_18</a>","mla":"Sun, Rémy, and Christoph Lampert. <i>KS(Conf): A Light-Weight Test If a ConvNet Operates Outside of Its Specifications</i>. Vol. 11269, Springer Nature, 2019, pp. 244–59, doi:<a href=\"https://doi.org/10.1007/978-3-030-12939-2_18\">10.1007/978-3-030-12939-2_18</a>.","apa":"Sun, R., &#38; Lampert, C. (2019). KS(conf): A light-weight test if a ConvNet operates outside of Its specifications (Vol. 11269, pp. 244–259). Presented at the GCPR: Conference on Pattern Recognition, Stuttgart, Germany: Springer Nature. <a href=\"https://doi.org/10.1007/978-3-030-12939-2_18\">https://doi.org/10.1007/978-3-030-12939-2_18</a>","chicago":"Sun, Rémy, and Christoph Lampert. “KS(Conf): A Light-Weight Test If a ConvNet Operates Outside of Its Specifications,” 11269:244–59. Springer Nature, 2019. <a href=\"https://doi.org/10.1007/978-3-030-12939-2_18\">https://doi.org/10.1007/978-3-030-12939-2_18</a>.","ista":"Sun R, Lampert C. 2019. KS(conf): A light-weight test if a ConvNet operates outside of Its specifications. GCPR: Conference on Pattern Recognition, LNCS, vol. 11269, 244–259."},"user_id":"c635000d-4b10-11ee-a964-aac5a93f6ac1","doi":"10.1007/978-3-030-12939-2_18","article_processing_charge":"No","alternative_title":["LNCS"],"publisher":"Springer Nature","date_updated":"2024-02-22T14:57:29Z","_id":"6482","type":"conference","page":"244-259","main_file_link":[{"url":"https://arxiv.org/abs/1804.04171","open_access":"1"}],"quality_controlled":"1","year":"2019","external_id":{"arxiv":["1804.04171"]},"related_material":{"record":[{"id":"6944","status":"public","relation":"later_version"}]},"project":[{"_id":"2532554C-B435-11E9-9278-68D0E5697425","call_identifier":"FP7","name":"Lifelong Learning of Visual Scene Understanding","grant_number":"308036"}],"status":"public","ec_funded":1,"conference":{"location":"Stuttgart, Germany","end_date":"2018-10-12","start_date":"2018-10-09","name":"GCPR: Conference on Pattern Recognition"},"date_published":"2019-02-14T00:00:00Z"},{"publication_status":"published","intvolume":"        97","abstract":[{"text":"Modern machine learning methods often require more data for training than a single expert can provide. Therefore, it has become a standard procedure to collect data from external sources, e.g. via crowdsourcing. Unfortunately, the quality of these sources is not always guaranteed. As additional complications, the data might be stored in a distributed way, or might even have to remain private. In this work, we address the question of how to learn robustly in such scenarios. Studying the problem through the lens of statistical learning theory, we derive a procedure that allows for learning from all available sources, yet automatically suppresses irrelevant or corrupted data. We show by extensive experiments that our method provides significant improvements over alternative approaches from robust statistics and distributed optimization. ","lang":"eng"}],"volume":97,"date_created":"2019-06-27T14:18:23Z","day":"01","scopus_import":"1","author":[{"first_name":"Nikola H","full_name":"Konstantinov, Nikola H","id":"4B9D76E4-F248-11E8-B48F-1D18A9856A87","last_name":"Konstantinov"},{"last_name":"Lampert","full_name":"Lampert, Christoph","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87","orcid":"0000-0001-8622-7887","first_name":"Christoph"}],"title":"Robust learning from untrusted sources","oa_version":"Preprint","citation":{"ama":"Konstantinov NH, Lampert C. Robust learning from untrusted sources. In: <i>Proceedings of the 36th International Conference on Machine Learning</i>. Vol 97. ML Research Press; 2019:3488-3498.","short":"N.H. Konstantinov, C. Lampert, in:, Proceedings of the 36th International Conference on Machine Learning, ML Research Press, 2019, pp. 3488–3498.","ieee":"N. H. Konstantinov and C. Lampert, “Robust learning from untrusted sources,” in <i>Proceedings of the 36th International Conference on Machine Learning</i>, Long Beach, CA, USA, 2019, vol. 97, pp. 3488–3498.","chicago":"Konstantinov, Nikola H, and Christoph Lampert. “Robust Learning from Untrusted Sources.” In <i>Proceedings of the 36th International Conference on Machine Learning</i>, 97:3488–98. ML Research Press, 2019.","ista":"Konstantinov NH, Lampert C. 2019. Robust learning from untrusted sources. Proceedings of the 36th International Conference on Machine Learning. ICML: International Conference on Machine Learning vol. 97, 3488–3498.","mla":"Konstantinov, Nikola H., and Christoph Lampert. “Robust Learning from Untrusted Sources.” <i>Proceedings of the 36th International Conference on Machine Learning</i>, vol. 97, ML Research Press, 2019, pp. 3488–98.","apa":"Konstantinov, N. H., &#38; Lampert, C. (2019). Robust learning from untrusted sources. In <i>Proceedings of the 36th International Conference on Machine Learning</i> (Vol. 97, pp. 3488–3498). Long Beach, CA, USA: ML Research Press."},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","oa":1,"language":[{"iso":"eng"}],"department":[{"_id":"ChLa"}],"month":"06","arxiv":1,"main_file_link":[{"url":"https://arxiv.org/abs/1901.10310","open_access":"1"}],"quality_controlled":"1","page":"3488-3498","_id":"6590","date_updated":"2023-10-17T12:31:55Z","type":"conference","article_processing_charge":"No","publisher":"ML Research Press","ec_funded":1,"date_published":"2019-06-01T00:00:00Z","conference":{"end_date":"2919-06-15","start_date":"2019-06-10","name":"ICML: International Conference on Machine Learning","location":"Long Beach, CA, USA"},"status":"public","publication":"Proceedings of the 36th International Conference on Machine Learning","project":[{"name":"Lifelong Learning of Visual Scene Understanding","grant_number":"308036","call_identifier":"FP7","_id":"2532554C-B435-11E9-9278-68D0E5697425"},{"name":"International IST Doctoral Program","grant_number":"665385","call_identifier":"H2020","_id":"2564DBCA-B435-11E9-9278-68D0E5697425"}],"year":"2019","related_material":{"record":[{"relation":"dissertation_contains","status":"public","id":"10799"}]},"external_id":{"arxiv":["1901.10310"]}},{"ddc":["004"],"page":"113","type":"dissertation","_id":"197","date_updated":"2023-09-07T12:51:46Z","publisher":"Institute of Science and Technology Austria","article_processing_charge":"No","alternative_title":["ISTA Thesis"],"doi":"10.15479/AT:ISTA:th_1021","acknowledgement":"I also gratefully acknowledge the support of NVIDIA Corporation with the donation of the GPUs used for this research.","date_published":"2018-05-25T00:00:00Z","ec_funded":1,"status":"public","project":[{"call_identifier":"FP7","grant_number":"308036","name":"Lifelong Learning of Visual Scene Understanding","_id":"2532554C-B435-11E9-9278-68D0E5697425"}],"degree_awarded":"PhD","publist_id":"7718","year":"2018","file_date_updated":"2020-07-14T12:45:22Z","publication_status":"published","publication_identifier":{"issn":["2663-337X"]},"abstract":[{"lang":"eng","text":"Modern computer vision systems heavily rely on statistical machine learning models, which typically require large amounts of labeled data to be learned reliably. Moreover, very recently computer vision research widely adopted techniques for representation learning, which further increase the demand for labeled data. However, for many important practical problems there is relatively small amount of labeled data available, so it is problematic to leverage full potential of the representation learning methods. One way to overcome this obstacle is to invest substantial resources into producing large labelled datasets. Unfortunately, this can be prohibitively expensive in practice. In this thesis we focus on the alternative way of tackling the aforementioned issue. We concentrate on methods, which make use of weakly-labeled or even unlabeled data. Specifically, the first half of the thesis is dedicated to the semantic image segmentation task. We develop a technique, which achieves competitive segmentation performance and only requires annotations in a form of global image-level labels instead of dense segmentation masks. Subsequently, we present a new methodology, which further improves segmentation performance by leveraging tiny additional feedback from a human annotator. By using our methods practitioners can greatly reduce the amount of data annotation effort, which is required to learn modern image segmentation models. In the second half of the thesis we focus on methods for learning from unlabeled visual data. We study a family of autoregressive models for modeling structure of natural images and discuss potential applications of these models. Moreover, we conduct in-depth study of one of these applications, where we develop the state-of-the-art model for the probabilistic image colorization task."}],"has_accepted_license":"1","date_created":"2018-12-11T11:45:09Z","title":"Weakly-Supervised Segmentation and Unsupervised Modeling of Natural Images","oa_version":"Published Version","day":"25","author":[{"full_name":"Kolesnikov, Alexander","id":"2D157DB6-F248-11E8-B48F-1D18A9856A87","last_name":"Kolesnikov","first_name":"Alexander"}],"user_id":"c635000d-4b10-11ee-a964-aac5a93f6ac1","citation":{"ama":"Kolesnikov A. Weakly-Supervised Segmentation and Unsupervised Modeling of Natural Images. 2018. doi:<a href=\"https://doi.org/10.15479/AT:ISTA:th_1021\">10.15479/AT:ISTA:th_1021</a>","ieee":"A. Kolesnikov, “Weakly-Supervised Segmentation and Unsupervised Modeling of Natural Images,” Institute of Science and Technology Austria, 2018.","short":"A. Kolesnikov, Weakly-Supervised Segmentation and Unsupervised Modeling of Natural Images, Institute of Science and Technology Austria, 2018.","chicago":"Kolesnikov, Alexander. “Weakly-Supervised Segmentation and Unsupervised Modeling of Natural Images.” Institute of Science and Technology Austria, 2018. <a href=\"https://doi.org/10.15479/AT:ISTA:th_1021\">https://doi.org/10.15479/AT:ISTA:th_1021</a>.","ista":"Kolesnikov A. 2018. Weakly-Supervised Segmentation and Unsupervised Modeling of Natural Images. Institute of Science and Technology Austria.","apa":"Kolesnikov, A. (2018). <i>Weakly-Supervised Segmentation and Unsupervised Modeling of Natural Images</i>. Institute of Science and Technology Austria. <a href=\"https://doi.org/10.15479/AT:ISTA:th_1021\">https://doi.org/10.15479/AT:ISTA:th_1021</a>","mla":"Kolesnikov, Alexander. <i>Weakly-Supervised Segmentation and Unsupervised Modeling of Natural Images</i>. Institute of Science and Technology Austria, 2018, doi:<a href=\"https://doi.org/10.15479/AT:ISTA:th_1021\">10.15479/AT:ISTA:th_1021</a>."},"pubrep_id":"1021","language":[{"iso":"eng"}],"oa":1,"file":[{"access_level":"open_access","content_type":"application/pdf","file_name":"IST-2018-1021-v1+1_thesis-unsigned-pdfa.pdf","checksum":"bc678e02468d8ebc39dc7267dfb0a1c4","relation":"main_file","date_updated":"2020-07-14T12:45:22Z","creator":"system","date_created":"2018-12-12T10:14:57Z","file_size":12918758,"file_id":"5113"},{"file_id":"6225","date_created":"2019-04-05T09:34:49Z","file_size":55973760,"date_updated":"2020-07-14T12:45:22Z","creator":"dernst","relation":"source_file","checksum":"bc66973b086da5a043f1162dcfb1fde4","file_name":"2018_Thesis_Kolesnikov_source.zip","access_level":"closed","content_type":"application/zip"}],"department":[{"_id":"ChLa"}],"month":"05","supervisor":[{"last_name":"Lampert","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87","full_name":"Lampert, Christoph","orcid":"0000-0001-8622-7887","first_name":"Christoph"}]},{"oa_version":"Published Version","title":"Learning from dependent data","day":"01","author":[{"last_name":"Zimin","id":"37099E9C-F248-11E8-B48F-1D18A9856A87","full_name":"Zimin, Alexander","first_name":"Alexander"}],"date_created":"2018-12-11T11:44:27Z","abstract":[{"lang":"eng","text":"The most common assumption made in statistical learning theory is the assumption of the independent and identically distributed (i.i.d.) data. While being very convenient mathematically, it is often very clearly violated in practice. This disparity between the machine learning theory and applications underlies a growing demand in the development of algorithms that learn from dependent data and theory that can provide generalization guarantees similar to the independent situations. This thesis is dedicated to two variants of dependencies that can arise in practice. One is a dependence on the level of samples in a single learning task. Another dependency type arises in the multi-task setting when the tasks are dependent on each other even though the data for them can be i.i.d. In both cases we model the data (samples or tasks) as stochastic processes and introduce new algorithms for both settings that take into account and exploit the resulting dependencies. We prove the theoretical guarantees on the performance of the introduced algorithms under different evaluation criteria and, in addition, we compliment the theoretical study by the empirical one, where we evaluate some of the algorithms on two real world datasets to highlight their practical applicability."}],"has_accepted_license":"1","file_date_updated":"2020-07-14T12:47:40Z","publication_identifier":{"issn":["2663-337X"]},"publication_status":"published","month":"09","supervisor":[{"first_name":"Christoph","orcid":"0000-0001-8622-7887","last_name":"Lampert","full_name":"Lampert, Christoph","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87"}],"file":[{"checksum":"e849dd40a915e4d6c5572b51b517f098","relation":"main_file","content_type":"application/pdf","access_level":"open_access","file_name":"2018_Thesis_Zimin.pdf","file_id":"6253","creator":"dernst","date_updated":"2020-07-14T12:47:40Z","file_size":1036137,"date_created":"2019-04-09T07:32:47Z"},{"date_created":"2019-04-09T07:32:47Z","file_size":637490,"date_updated":"2020-07-14T12:47:40Z","creator":"dernst","file_id":"6254","file_name":"2018_Thesis_Zimin_Source.zip","content_type":"application/zip","access_level":"closed","relation":"source_file","checksum":"da092153cec55c97461bd53c45c5d139"}],"department":[{"_id":"ChLa"}],"language":[{"iso":"eng"}],"pubrep_id":"1048","oa":1,"user_id":"c635000d-4b10-11ee-a964-aac5a93f6ac1","citation":{"apa":"Zimin, A. (2018). <i>Learning from dependent data</i>. Institute of Science and Technology Austria. <a href=\"https://doi.org/10.15479/AT:ISTA:TH1048\">https://doi.org/10.15479/AT:ISTA:TH1048</a>","mla":"Zimin, Alexander. <i>Learning from Dependent Data</i>. Institute of Science and Technology Austria, 2018, doi:<a href=\"https://doi.org/10.15479/AT:ISTA:TH1048\">10.15479/AT:ISTA:TH1048</a>.","chicago":"Zimin, Alexander. “Learning from Dependent Data.” Institute of Science and Technology Austria, 2018. <a href=\"https://doi.org/10.15479/AT:ISTA:TH1048\">https://doi.org/10.15479/AT:ISTA:TH1048</a>.","ista":"Zimin A. 2018. Learning from dependent data. Institute of Science and Technology Austria.","ieee":"A. Zimin, “Learning from dependent data,” Institute of Science and Technology Austria, 2018.","short":"A. Zimin, Learning from Dependent Data, Institute of Science and Technology Austria, 2018.","ama":"Zimin A. Learning from dependent data. 2018. doi:<a href=\"https://doi.org/10.15479/AT:ISTA:TH1048\">10.15479/AT:ISTA:TH1048</a>"},"publisher":"Institute of Science and Technology Austria","article_processing_charge":"No","alternative_title":["ISTA Thesis"],"doi":"10.15479/AT:ISTA:TH1048","type":"dissertation","_id":"68","date_updated":"2023-09-07T12:29:07Z","ddc":["004","519"],"page":"92","year":"2018","publist_id":"7986","status":"public","project":[{"_id":"2532554C-B435-11E9-9278-68D0E5697425","call_identifier":"FP7","grant_number":"308036","name":"Lifelong Learning of Visual Scene Understanding"}],"degree_awarded":"PhD","date_published":"2018-09-01T00:00:00Z","ec_funded":1},{"main_file_link":[{"open_access":"1","url":"https://arxiv.org/abs/1703.01678"}],"quality_controlled":"1","page":"2815-2824","_id":"6011","date_updated":"2023-10-17T09:51:13Z","type":"conference","article_processing_charge":"No","publisher":"ML Research Press","ec_funded":1,"date_published":"2018-02-01T00:00:00Z","conference":{"location":"Stockholm, Sweden","start_date":"2018-07-10","end_date":"2018-07-15","name":"ICML: International Conference on Machine Learning"},"status":"public","publication":"Proceedings of the 35 th International Conference on Machine Learning","project":[{"_id":"2532554C-B435-11E9-9278-68D0E5697425","call_identifier":"FP7","name":"Lifelong Learning of Visual Scene Understanding","grant_number":"308036"}],"isi":1,"year":"2018","external_id":{"arxiv":["1703.01678"],"isi":["000683379202095"]},"publication_status":"published","intvolume":"        80","abstract":[{"text":"We establish a data-dependent notion of algorithmic stability for Stochastic Gradient Descent (SGD), and employ it to develop novel generalization bounds. This is in contrast to previous distribution-free algorithmic stability results for SGD which depend on the worst-case constants. By virtue of the data-dependent argument, our bounds provide new insights into learning with SGD on convex and non-convex problems. In the convex case, we show that the bound on the generalization error depends on the risk at the initialization point. In the non-convex case, we prove that the expected curvature of the objective function around the initialization point has crucial influence on the generalization error. In both cases, our results suggest a simple data-driven strategy to stabilize SGD by pre-screening its initialization. As a corollary, our results allow us to show optimistic generalization bounds that exhibit fast convergence rates for SGD subject to a vanishing empirical risk and low noise of stochastic gradient. ","lang":"eng"}],"volume":80,"date_created":"2019-02-14T14:51:57Z","day":"01","scopus_import":"1","author":[{"full_name":"Kuzborskij, Ilja","last_name":"Kuzborskij","first_name":"Ilja"},{"full_name":"Lampert, Christoph","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87","last_name":"Lampert","orcid":"0000-0001-8622-7887","first_name":"Christoph"}],"title":"Data-dependent stability of stochastic gradient descent","oa_version":"Preprint","citation":{"ieee":"I. Kuzborskij and C. Lampert, “Data-dependent stability of stochastic gradient descent,” in <i>Proceedings of the 35 th International Conference on Machine Learning</i>, Stockholm, Sweden, 2018, vol. 80, pp. 2815–2824.","short":"I. Kuzborskij, C. Lampert, in:, Proceedings of the 35 Th International Conference on Machine Learning, ML Research Press, 2018, pp. 2815–2824.","ama":"Kuzborskij I, Lampert C. Data-dependent stability of stochastic gradient descent. In: <i>Proceedings of the 35 Th International Conference on Machine Learning</i>. Vol 80. ML Research Press; 2018:2815-2824.","apa":"Kuzborskij, I., &#38; Lampert, C. (2018). Data-dependent stability of stochastic gradient descent. In <i>Proceedings of the 35 th International Conference on Machine Learning</i> (Vol. 80, pp. 2815–2824). Stockholm, Sweden: ML Research Press.","mla":"Kuzborskij, Ilja, and Christoph Lampert. “Data-Dependent Stability of Stochastic Gradient Descent.” <i>Proceedings of the 35 Th International Conference on Machine Learning</i>, vol. 80, ML Research Press, 2018, pp. 2815–24.","chicago":"Kuzborskij, Ilja, and Christoph Lampert. “Data-Dependent Stability of Stochastic Gradient Descent.” In <i>Proceedings of the 35 Th International Conference on Machine Learning</i>, 80:2815–24. ML Research Press, 2018.","ista":"Kuzborskij I, Lampert C. 2018. Data-dependent stability of stochastic gradient descent. Proceedings of the 35 th International Conference on Machine Learning. ICML: International Conference on Machine Learning vol. 80, 2815–2824."},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","oa":1,"language":[{"iso":"eng"}],"department":[{"_id":"ChLa"}],"arxiv":1,"month":"02"},{"alternative_title":["PMLR"],"article_processing_charge":"No","publisher":"ML Research Press","date_updated":"2023-10-17T10:01:12Z","_id":"1108","type":"conference","page":"213 - 222","main_file_link":[{"url":"http://proceedings.mlr.press/v54/zimin17a/zimin17a.pdf","open_access":"1"}],"quality_controlled":"1","year":"2017","isi":1,"external_id":{"isi":["000509368500024"]},"publist_id":"6261","project":[{"grant_number":"308036","name":"Lifelong Learning of Visual Scene Understanding","call_identifier":"FP7","_id":"2532554C-B435-11E9-9278-68D0E5697425"}],"status":"public","ec_funded":1,"conference":{"name":"AISTATS: Artificial Intelligence and Statistics","end_date":"2017-04-22","start_date":"2017-04-20","location":"Fort Lauderdale, FL, United States"},"date_published":"2017-04-01T00:00:00Z","author":[{"first_name":"Alexander","last_name":"Zimin","full_name":"Zimin, Alexander","id":"37099E9C-F248-11E8-B48F-1D18A9856A87"},{"first_name":"Christoph","orcid":"0000-0001-8622-7887","last_name":"Lampert","full_name":"Lampert, Christoph","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87"}],"day":"01","oa_version":"Submitted Version","title":"Learning theory for conditional risk minimization","volume":54,"date_created":"2018-12-11T11:50:11Z","intvolume":"        54","abstract":[{"text":"In this work we study the learnability of stochastic processes with respect to the conditional risk, i.e. the existence of a learning algorithm that improves its next-step performance with the amount of observed data. We introduce a notion of pairwise discrepancy between conditional distributions at different times steps and show how certain properties of these discrepancies can be used to construct a successful learning algorithm. Our main results are two theorems that establish criteria for learnability for many classes of stochastic processes, including all special cases studied previously in the literature.","lang":"eng"}],"publication_status":"published","month":"04","department":[{"_id":"ChLa"}],"oa":1,"language":[{"iso":"eng"}],"citation":{"ieee":"A. Zimin and C. Lampert, “Learning theory for conditional risk minimization,” presented at the AISTATS: Artificial Intelligence and Statistics, Fort Lauderdale, FL, United States, 2017, vol. 54, pp. 213–222.","short":"A. Zimin, C. Lampert, in:, ML Research Press, 2017, pp. 213–222.","ama":"Zimin A, Lampert C. Learning theory for conditional risk minimization. In: Vol 54. ML Research Press; 2017:213-222.","apa":"Zimin, A., &#38; Lampert, C. (2017). Learning theory for conditional risk minimization (Vol. 54, pp. 213–222). Presented at the AISTATS: Artificial Intelligence and Statistics, Fort Lauderdale, FL, United States: ML Research Press.","mla":"Zimin, Alexander, and Christoph Lampert. <i>Learning Theory for Conditional Risk Minimization</i>. Vol. 54, ML Research Press, 2017, pp. 213–22.","ista":"Zimin A, Lampert C. 2017. Learning theory for conditional risk minimization. AISTATS: Artificial Intelligence and Statistics, PMLR, vol. 54, 213–222.","chicago":"Zimin, Alexander, and Christoph Lampert. “Learning Theory for Conditional Risk Minimization,” 54:213–22. ML Research Press, 2017."},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87"},{"year":"2017","month":"02","arxiv":1,"external_id":{"arxiv":["1610.02995"]},"department":[{"_id":"ChLa"}],"oa":1,"status":"public","language":[{"iso":"eng"}],"publication":"5th International Conference on Learning Representations, ICLR 2017 - Workshop Track Proceedings","project":[{"_id":"2532554C-B435-11E9-9278-68D0E5697425","call_identifier":"FP7","name":"Lifelong Learning of Visual Scene Understanding","grant_number":"308036"}],"citation":{"ista":"Martius GS, Lampert C. 2017. Extrapolation and learning equations. 5th International Conference on Learning Representations, ICLR 2017 - Workshop Track Proceedings. ICLR: International Conference on Learning Representations.","chicago":"Martius, Georg S, and Christoph Lampert. “Extrapolation and Learning Equations.” In <i>5th International Conference on Learning Representations, ICLR 2017 - Workshop Track Proceedings</i>. International Conference on Learning Representations, 2017.","apa":"Martius, G. S., &#38; Lampert, C. (2017). Extrapolation and learning equations. In <i>5th International Conference on Learning Representations, ICLR 2017 - Workshop Track Proceedings</i>. Toulon, France: International Conference on Learning Representations.","mla":"Martius, Georg S., and Christoph Lampert. “Extrapolation and Learning Equations.” <i>5th International Conference on Learning Representations, ICLR 2017 - Workshop Track Proceedings</i>, International Conference on Learning Representations, 2017.","ama":"Martius GS, Lampert C. Extrapolation and learning equations. In: <i>5th International Conference on Learning Representations, ICLR 2017 - Workshop Track Proceedings</i>. International Conference on Learning Representations; 2017.","ieee":"G. S. Martius and C. Lampert, “Extrapolation and learning equations,” in <i>5th International Conference on Learning Representations, ICLR 2017 - Workshop Track Proceedings</i>, Toulon, France, 2017.","short":"G.S. Martius, C. Lampert, in:, 5th International Conference on Learning Representations, ICLR 2017 - Workshop Track Proceedings, International Conference on Learning Representations, 2017."},"ec_funded":1,"user_id":"3E5EF7F0-F248-11E8-B48F-1D18A9856A87","date_published":"2017-02-21T00:00:00Z","conference":{"location":"Toulon, France","start_date":"2017-04-24","end_date":"2017-04-26","name":"ICLR: International Conference on Learning Representations"},"day":"21","scopus_import":1,"author":[{"first_name":"Georg S","id":"3A276B68-F248-11E8-B48F-1D18A9856A87","full_name":"Martius, Georg S","last_name":"Martius"},{"orcid":"0000-0001-8622-7887","first_name":"Christoph","last_name":"Lampert","full_name":"Lampert, Christoph","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87"}],"publisher":"International Conference on Learning Representations","title":"Extrapolation and learning equations","oa_version":"Preprint","_id":"6841","date_updated":"2021-01-12T08:09:17Z","date_created":"2019-09-01T22:01:00Z","type":"conference","abstract":[{"lang":"eng","text":"In classical machine learning, regression is treated as a black box process of identifying a suitable function from a hypothesis set without attempting to gain insight into the mechanism connecting inputs and outputs. In the natural sciences, however, finding an interpretable function for a phenomenon is the prime goal as it allows to understand and generalize results. This paper proposes a novel type of function learning network, called equation learner (EQL), that can learn analytical expressions and is able to extrapolate to unseen domains. It is implemented as an end-to-end differentiable feed-forward network and allows for efficient gradient based training. Due to sparsity regularization concise interpretable expressions can be obtained. Often the true underlying source expression is identified."}],"main_file_link":[{"url":"https://arxiv.org/abs/1610.02995","open_access":"1"}],"quality_controlled":"1","publication_status":"published"},{"year":"2017","external_id":{"arxiv":["1705.04258"]},"related_material":{"record":[{"relation":"dissertation_contains","status":"public","id":"8390"}]},"publist_id":"6532","project":[{"call_identifier":"FP7","grant_number":"308036","name":"Lifelong Learning of Visual Scene Understanding","_id":"2532554C-B435-11E9-9278-68D0E5697425"}],"status":"public","ec_funded":1,"date_published":"2017-09-01T00:00:00Z","conference":{"end_date":"2017-09-07","start_date":"2017-09-04","name":"BMVC: British Machine Vision Conference","location":"London, United Kingdom"},"doi":"10.5244/c.31.85","article_processing_charge":"No","publisher":"BMVA Press","date_updated":"2023-10-16T10:04:02Z","_id":"911","type":"conference","page":"85.1-85.12","ddc":["000"],"quality_controlled":"1","arxiv":1,"month":"09","department":[{"_id":"ChLa"}],"file":[{"file_id":"8224","file_size":1625363,"date_created":"2020-08-10T07:14:33Z","creator":"dernst","date_updated":"2020-08-10T07:14:33Z","relation":"main_file","file_name":"2017_BMVC_Royer.pdf","success":1,"content_type":"application/pdf","access_level":"open_access"}],"oa":1,"language":[{"iso":"eng"}],"citation":{"apa":"Royer, A., Kolesnikov, A., &#38; Lampert, C. (2017). Probabilistic image colorization (p. 85.1-85.12). Presented at the BMVC: British Machine Vision Conference, London, United Kingdom: BMVA Press. <a href=\"https://doi.org/10.5244/c.31.85\">https://doi.org/10.5244/c.31.85</a>","mla":"Royer, Amélie, et al. <i>Probabilistic Image Colorization</i>. BMVA Press, 2017, p. 85.1-85.12, doi:<a href=\"https://doi.org/10.5244/c.31.85\">10.5244/c.31.85</a>.","chicago":"Royer, Amélie, Alexander Kolesnikov, and Christoph Lampert. “Probabilistic Image Colorization,” 85.1-85.12. BMVA Press, 2017. <a href=\"https://doi.org/10.5244/c.31.85\">https://doi.org/10.5244/c.31.85</a>.","ista":"Royer A, Kolesnikov A, Lampert C. 2017. Probabilistic image colorization. BMVC: British Machine Vision Conference, 85.1-85.12.","ieee":"A. Royer, A. Kolesnikov, and C. Lampert, “Probabilistic image colorization,” presented at the BMVC: British Machine Vision Conference, London, United Kingdom, 2017, p. 85.1-85.12.","short":"A. Royer, A. Kolesnikov, C. Lampert, in:, BMVA Press, 2017, p. 85.1-85.12.","ama":"Royer A, Kolesnikov A, Lampert C. Probabilistic image colorization. In: BMVA Press; 2017:85.1-85.12. doi:<a href=\"https://doi.org/10.5244/c.31.85\">10.5244/c.31.85</a>"},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","author":[{"full_name":"Royer, Amélie","id":"3811D890-F248-11E8-B48F-1D18A9856A87","last_name":"Royer","orcid":"0000-0002-8407-0705","first_name":"Amélie"},{"last_name":"Kolesnikov","id":"2D157DB6-F248-11E8-B48F-1D18A9856A87","full_name":"Kolesnikov, Alexander","first_name":"Alexander"},{"first_name":"Christoph","orcid":"0000-0001-8622-7887","last_name":"Lampert","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87","full_name":"Lampert, Christoph"}],"scopus_import":"1","day":"01","oa_version":"Published Version","title":"Probabilistic image colorization","date_created":"2018-12-11T11:49:09Z","has_accepted_license":"1","abstract":[{"text":"We develop a probabilistic technique for colorizing grayscale natural images. In light of the intrinsic uncertainty of this task, the proposed probabilistic framework has numerous desirable properties. In particular, our model is able to produce multiple plausible and vivid colorizations for a given grayscale image and is one of the first colorization models to provide a proper stochastic sampling scheme. Moreover, our training procedure is supported by a rigorous theoretical framework that does not require any ad hoc heuristics and allows for efficient modeling and learning of the joint pixel color distribution.We demonstrate strong quantitative and qualitative experimental results on the CIFAR-10 dataset and the challenging ILSVRC 2012 dataset.","lang":"eng"}],"publication_status":"published","publication_identifier":{"eisbn":["190172560X"]},"file_date_updated":"2020-08-10T07:14:33Z"},{"article_processing_charge":"No","publisher":"JMLR","date_updated":"2023-09-22T09:50:41Z","_id":"1000","type":"conference","page":"1905 - 1914","main_file_link":[{"open_access":"1","url":"https://arxiv.org/abs/1612.08185"}],"quality_controlled":"1","year":"2017","isi":1,"external_id":{"isi":["000683309501102"],"arxiv":["1612.08185"]},"publist_id":"6398","project":[{"call_identifier":"FP7","grant_number":"308036","name":"Lifelong Learning of Visual Scene Understanding","_id":"2532554C-B435-11E9-9278-68D0E5697425"}],"publication":"34th International Conference on Machine Learning","status":"public","ec_funded":1,"conference":{"end_date":"2017-08-11","start_date":"2017-08-06","name":"ICML: International Conference on Machine Learning","location":"Sydney, Australia"},"date_published":"2017-08-01T00:00:00Z","acknowledgement":"We thank Tim Salimans for spotting a mistake in our preliminary arXiv manuscript. This work was funded by the European Research Council under the European Unions Seventh Framework Programme (FP7/2007-2013)/ERC grant agreement no 308036.","author":[{"id":"2D157DB6-F248-11E8-B48F-1D18A9856A87","full_name":"Kolesnikov, Alexander","last_name":"Kolesnikov","first_name":"Alexander"},{"last_name":"Lampert","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87","full_name":"Lampert, Christoph","orcid":"0000-0001-8622-7887","first_name":"Christoph"}],"scopus_import":"1","day":"01","title":"PixelCNN models with auxiliary variables for natural image modeling","oa_version":"Submitted Version","volume":70,"date_created":"2018-12-11T11:49:37Z","has_accepted_license":"1","abstract":[{"lang":"eng","text":"We study probabilistic models of natural images and extend the autoregressive family of PixelCNN models by incorporating latent variables. Subsequently, we describe two new generative image models that exploit different image transformations as latent variables: a quantized grayscale view of the image or a multi-resolution image pyramid. The proposed models tackle two known shortcomings of existing PixelCNN models: 1) their tendency to focus on low-level image details, while largely ignoring high-level image information, such as object shapes, and 2) their computationally costly procedure for image sampling. We experimentally demonstrate benefits of our LatentPixelCNN models, in particular showing that they produce much more realistically looking image samples than previous state-of-the-art probabilistic models. "}],"intvolume":"        70","publication_identifier":{"isbn":["978-151085514-4"]},"publication_status":"published","month":"08","arxiv":1,"department":[{"_id":"ChLa"}],"oa":1,"language":[{"iso":"eng"}],"citation":{"mla":"Kolesnikov, Alexander, and Christoph Lampert. “PixelCNN Models with Auxiliary Variables for Natural Image Modeling.” <i>34th International Conference on Machine Learning</i>, vol. 70, JMLR, 2017, pp. 1905–14.","apa":"Kolesnikov, A., &#38; Lampert, C. (2017). PixelCNN models with auxiliary variables for natural image modeling. In <i>34th International Conference on Machine Learning</i> (Vol. 70, pp. 1905–1914). Sydney, Australia: JMLR.","ista":"Kolesnikov A, Lampert C. 2017. PixelCNN models with auxiliary variables for natural image modeling. 34th International Conference on Machine Learning. ICML: International Conference on Machine Learning vol. 70, 1905–1914.","chicago":"Kolesnikov, Alexander, and Christoph Lampert. “PixelCNN Models with Auxiliary Variables for Natural Image Modeling.” In <i>34th International Conference on Machine Learning</i>, 70:1905–14. JMLR, 2017.","short":"A. Kolesnikov, C. Lampert, in:, 34th International Conference on Machine Learning, JMLR, 2017, pp. 1905–1914.","ieee":"A. Kolesnikov and C. Lampert, “PixelCNN models with auxiliary variables for natural image modeling,” in <i>34th International Conference on Machine Learning</i>, Sydney, Australia, 2017, vol. 70, pp. 1905–1914.","ama":"Kolesnikov A, Lampert C. PixelCNN models with auxiliary variables for natural image modeling. In: <i>34th International Conference on Machine Learning</i>. Vol 70. JMLR; 2017:1905-1914."},"user_id":"c635000d-4b10-11ee-a964-aac5a93f6ac1"},{"department":[{"_id":"ChLa"},{"_id":"ChWo"}],"month":"04","citation":{"mla":"Rebuffi, Sylvestre Alvise, et al. <i>ICaRL: Incremental Classifier and Representation Learning</i>. Vol. 2017, IEEE, 2017, pp. 5533–42, doi:<a href=\"https://doi.org/10.1109/CVPR.2017.587\">10.1109/CVPR.2017.587</a>.","apa":"Rebuffi, S. A., Kolesnikov, A., Sperl, G., &#38; Lampert, C. (2017). iCaRL: Incremental classifier and representation learning (Vol. 2017, pp. 5533–5542). Presented at the CVPR: Computer Vision and Pattern Recognition, Honolulu, HA, United States: IEEE. <a href=\"https://doi.org/10.1109/CVPR.2017.587\">https://doi.org/10.1109/CVPR.2017.587</a>","chicago":"Rebuffi, Sylvestre Alvise, Alexander Kolesnikov, Georg Sperl, and Christoph Lampert. “ICaRL: Incremental Classifier and Representation Learning,” 2017:5533–42. IEEE, 2017. <a href=\"https://doi.org/10.1109/CVPR.2017.587\">https://doi.org/10.1109/CVPR.2017.587</a>.","ista":"Rebuffi SA, Kolesnikov A, Sperl G, Lampert C. 2017. iCaRL: Incremental classifier and representation learning. CVPR: Computer Vision and Pattern Recognition vol. 2017, 5533–5542.","short":"S.A. Rebuffi, A. Kolesnikov, G. Sperl, C. Lampert, in:, IEEE, 2017, pp. 5533–5542.","ieee":"S. A. Rebuffi, A. Kolesnikov, G. Sperl, and C. Lampert, “iCaRL: Incremental classifier and representation learning,” presented at the CVPR: Computer Vision and Pattern Recognition, Honolulu, HA, United States, 2017, vol. 2017, pp. 5533–5542.","ama":"Rebuffi SA, Kolesnikov A, Sperl G, Lampert C. iCaRL: Incremental classifier and representation learning. In: Vol 2017. IEEE; 2017:5533-5542. doi:<a href=\"https://doi.org/10.1109/CVPR.2017.587\">10.1109/CVPR.2017.587</a>"},"user_id":"c635000d-4b10-11ee-a964-aac5a93f6ac1","oa":1,"language":[{"iso":"eng"}],"volume":2017,"date_created":"2018-12-11T11:49:37Z","day":"14","scopus_import":"1","author":[{"last_name":"Rebuffi","full_name":"Rebuffi, Sylvestre Alvise","first_name":"Sylvestre Alvise"},{"full_name":"Kolesnikov, Alexander","id":"2D157DB6-F248-11E8-B48F-1D18A9856A87","last_name":"Kolesnikov","first_name":"Alexander"},{"first_name":"Georg","id":"4DD40360-F248-11E8-B48F-1D18A9856A87","full_name":"Sperl, Georg","last_name":"Sperl"},{"orcid":"0000-0001-8622-7887","first_name":"Christoph","last_name":"Lampert","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87","full_name":"Lampert, Christoph"}],"oa_version":"Submitted Version","title":"iCaRL: Incremental classifier and representation learning","publication_status":"published","publication_identifier":{"isbn":["978-153860457-1"]},"abstract":[{"lang":"eng","text":"A major open problem on the road to artificial intelligence is the development of incrementally learning systems that learn about more and more concepts over time from a stream of data. In this work, we introduce a new training strategy, iCaRL, that allows learning in such a class-incremental way: only the training data for a small number of classes has to be present at the same time and new classes can be added progressively. iCaRL learns strong classifiers and a data representation simultaneously. This distinguishes it from earlier works that were fundamentally limited to fixed data representations and therefore incompatible with deep learning architectures. We show by experiments on CIFAR-100 and ImageNet ILSVRC 2012 data that iCaRL can learn many classes incrementally over a long period of time where other strategies quickly fail. "}],"intvolume":"      2017","publist_id":"6400","isi":1,"year":"2017","external_id":{"isi":["000418371405066"]},"ec_funded":1,"conference":{"name":"CVPR: Computer Vision and Pattern Recognition","end_date":"2017-07-26","start_date":"2017-07-21","location":"Honolulu, HA, United States"},"date_published":"2017-04-14T00:00:00Z","status":"public","project":[{"_id":"2532554C-B435-11E9-9278-68D0E5697425","name":"Lifelong Learning of Visual Scene Understanding","grant_number":"308036","call_identifier":"FP7"}],"_id":"998","date_updated":"2023-09-22T09:51:58Z","type":"conference","article_processing_charge":"No","doi":"10.1109/CVPR.2017.587","publisher":"IEEE","main_file_link":[{"open_access":"1","url":"https://arxiv.org/abs/1611.07725"}],"quality_controlled":"1","page":"5533 - 5542"},{"intvolume":"        70","abstract":[{"lang":"eng","text":"In multi-task learning, a learner is given a collection of prediction tasks and needs to solve all of them. In contrast to previous work, which required that annotated training data must be available for all tasks, we consider a new setting, in which for some tasks, potentially most of them, only unlabeled training data is provided. Consequently, to solve all tasks, information must be transferred between tasks with labels and tasks without labels. Focusing on an instance-based transfer method we analyze two variants of this setting: when the set of labeled tasks is fixed, and when it can be actively selected by the learner. We state and prove a generalization bound that covers both scenarios and derive from it an algorithm for making the choice of labeled tasks (in the active case) and for transferring information between the tasks in a principled way. We also illustrate the effectiveness of the algorithm on synthetic and real data. "}],"publication_identifier":{"isbn":["9781510855144"]},"publication_status":"published","author":[{"first_name":"Anastasia","full_name":"Pentina, Anastasia","id":"42E87FC6-F248-11E8-B48F-1D18A9856A87","last_name":"Pentina"},{"last_name":"Lampert","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87","full_name":"Lampert, Christoph","orcid":"0000-0001-8622-7887","first_name":"Christoph"}],"day":"08","scopus_import":"1","title":"Multi-task learning with labeled and unlabeled tasks","oa_version":"Submitted Version","volume":70,"date_created":"2018-12-11T11:49:37Z","oa":1,"language":[{"iso":"eng"}],"citation":{"mla":"Pentina, Anastasia, and Christoph Lampert. <i>Multi-Task Learning with Labeled and Unlabeled Tasks</i>. Vol. 70, ML Research Press, 2017, pp. 2807–16.","apa":"Pentina, A., &#38; Lampert, C. (2017). Multi-task learning with labeled and unlabeled tasks (Vol. 70, pp. 2807–2816). Presented at the ICML: International Conference on Machine Learning, Sydney, Australia: ML Research Press.","chicago":"Pentina, Anastasia, and Christoph Lampert. “Multi-Task Learning with Labeled and Unlabeled Tasks,” 70:2807–16. ML Research Press, 2017.","ista":"Pentina A, Lampert C. 2017. Multi-task learning with labeled and unlabeled tasks. ICML: International Conference on Machine Learning, PMLR, vol. 70, 2807–2816.","short":"A. Pentina, C. Lampert, in:, ML Research Press, 2017, pp. 2807–2816.","ieee":"A. Pentina and C. Lampert, “Multi-task learning with labeled and unlabeled tasks,” presented at the ICML: International Conference on Machine Learning, Sydney, Australia, 2017, vol. 70, pp. 2807–2816.","ama":"Pentina A, Lampert C. Multi-task learning with labeled and unlabeled tasks. In: Vol 70. ML Research Press; 2017:2807-2816."},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","month":"06","department":[{"_id":"ChLa"}],"page":"2807 - 2816","main_file_link":[{"url":"https://arxiv.org/abs/1602.06518","open_access":"1"}],"quality_controlled":"1","alternative_title":["PMLR"],"article_processing_charge":"No","publisher":"ML Research Press","date_updated":"2023-10-17T11:53:32Z","_id":"999","type":"conference","project":[{"_id":"2532554C-B435-11E9-9278-68D0E5697425","call_identifier":"FP7","name":"Lifelong Learning of Visual Scene Understanding","grant_number":"308036"}],"status":"public","ec_funded":1,"conference":{"location":"Sydney, Australia","name":"ICML: International Conference on Machine Learning","start_date":"2017-08-06","end_date":"2017-08-11"},"date_published":"2017-06-08T00:00:00Z","year":"2017","isi":1,"external_id":{"isi":["000683309502093"]},"publist_id":"6399"},{"date_created":"2018-12-11T11:50:08Z","volume":29,"title":"Lifelong learning with weighted majority votes","oa_version":"Published Version","author":[{"first_name":"Anastasia","full_name":"Pentina, Anastasia","id":"42E87FC6-F248-11E8-B48F-1D18A9856A87","last_name":"Pentina"},{"first_name":"Ruth","last_name":"Urner","full_name":"Urner, Ruth"}],"scopus_import":1,"day":"01","publication_status":"published","file_date_updated":"2018-12-12T10:12:43Z","abstract":[{"text":"Better understanding of the potential benefits of information transfer and representation learning is an important step towards the goal of building intelligent systems that are able to persist in the world and learn over time. In this work, we consider a setting where the learner encounters a stream of tasks but is able to retain only limited information from each encountered task, such as a learned predictor. In contrast to most previous works analyzing this scenario, we do not make any distributional assumptions on the task generating process. Instead, we formulate a complexity measure that captures the diversity of the observed tasks. We provide a lifelong learning algorithm with error guarantees for every observed task (rather than on average). We show sample complexity reductions in comparison to solving every task in isolation in terms of our task complexity measure. Further, our algorithmic framework can naturally be viewed as learning a representation from encountered tasks with a neural network.","lang":"eng"}],"intvolume":"        29","has_accepted_license":"1","file":[{"creator":"system","date_updated":"2018-12-12T10:12:42Z","file_size":237111,"date_created":"2018-12-12T10:12:42Z","file_id":"4961","access_level":"open_access","content_type":"application/pdf","file_name":"IST-2017-775-v1+1_main.pdf","relation":"main_file"},{"access_level":"open_access","content_type":"application/pdf","file_name":"IST-2017-775-v1+2_supplementary.pdf","relation":"main_file","creator":"system","date_updated":"2018-12-12T10:12:43Z","date_created":"2018-12-12T10:12:43Z","file_size":185818,"file_id":"4962"}],"department":[{"_id":"ChLa"}],"month":"12","user_id":"3E5EF7F0-F248-11E8-B48F-1D18A9856A87","citation":{"ista":"Pentina A, Urner R. 2016. Lifelong learning with weighted majority votes. NIPS: Neural Information Processing Systems, Advances in Neural Information Processing Systems, vol. 29, 3619–3627.","chicago":"Pentina, Anastasia, and Ruth Urner. “Lifelong Learning with Weighted Majority Votes,” 29:3619–27. Neural Information Processing Systems, 2016.","mla":"Pentina, Anastasia, and Ruth Urner. <i>Lifelong Learning with Weighted Majority Votes</i>. Vol. 29, Neural Information Processing Systems, 2016, pp. 3619–27.","apa":"Pentina, A., &#38; Urner, R. (2016). Lifelong learning with weighted majority votes (Vol. 29, pp. 3619–3627). Presented at the NIPS: Neural Information Processing Systems, Barcelona, Spain: Neural Information Processing Systems.","ama":"Pentina A, Urner R. Lifelong learning with weighted majority votes. In: Vol 29. Neural Information Processing Systems; 2016:3619-3627.","short":"A. Pentina, R. Urner, in:, Neural Information Processing Systems, 2016, pp. 3619–3627.","ieee":"A. Pentina and R. Urner, “Lifelong learning with weighted majority votes,” presented at the NIPS: Neural Information Processing Systems, Barcelona, Spain, 2016, vol. 29, pp. 3619–3627."},"pubrep_id":"775","language":[{"iso":"eng"}],"oa":1,"type":"conference","date_updated":"2021-01-12T06:48:15Z","_id":"1098","publisher":"Neural Information Processing Systems","alternative_title":["Advances in Neural Information Processing Systems"],"quality_controlled":"1","ddc":["006"],"page":"3619-3627","publist_id":"6277","year":"2016","date_published":"2016-12-01T00:00:00Z","acknowledgement":"This work was in parts funded by the European Research Council under the European Union’s Seventh Framework Programme (FP7/2007-2013)/ERC grant agreement no 308036.\r\n\r\n","conference":{"name":"NIPS: Neural Information Processing Systems","end_date":"2016-12-10","start_date":"2016-12-05","location":"Barcelona, Spain"},"ec_funded":1,"project":[{"name":"Lifelong Learning of Visual Scene Understanding","grant_number":"308036","call_identifier":"FP7","_id":"2532554C-B435-11E9-9278-68D0E5697425"}],"status":"public"},{"abstract":[{"lang":"eng","text":"Weakly-supervised object localization methods tend to fail for object classes that consistently co-occur with the same background elements, e.g. trains on tracks. We propose a method to overcome these failures by adding a very small amount of model-specific additional annotation. The main idea is to cluster a deep network\\'s mid-level representations and assign object or distractor labels to each cluster. Experiments show substantially improved localization results on the challenging ILSVC2014 dataset for bounding box detection and the PASCAL VOC2012 dataset for semantic segmentation."}],"publication_status":"published","author":[{"full_name":"Kolesnikov, Alexander","id":"2D157DB6-F248-11E8-B48F-1D18A9856A87","last_name":"Kolesnikov","first_name":"Alexander"},{"last_name":"Lampert","full_name":"Lampert, Christoph","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87","orcid":"0000-0001-8622-7887","first_name":"Christoph"}],"day":"01","scopus_import":1,"oa_version":"Published Version","title":"Improving weakly-supervised object localization by micro-annotation","volume":"2016-September","date_created":"2018-12-11T11:50:09Z","oa":1,"language":[{"iso":"eng"}],"citation":{"ama":"Kolesnikov A, Lampert C. Improving weakly-supervised object localization by micro-annotation. In: <i>Proceedings of the British Machine Vision Conference 2016</i>. Vol 2016-September. BMVA Press; 2016:92.1-92.12. doi:<a href=\"https://doi.org/10.5244/C.30.92\">10.5244/C.30.92</a>","ieee":"A. Kolesnikov and C. Lampert, “Improving weakly-supervised object localization by micro-annotation,” in <i>Proceedings of the British Machine Vision Conference 2016</i>, York, United Kingdom, 2016, vol. 2016–September, p. 92.1-92.12.","short":"A. Kolesnikov, C. Lampert, in:, Proceedings of the British Machine Vision Conference 2016, BMVA Press, 2016, p. 92.1-92.12.","chicago":"Kolesnikov, Alexander, and Christoph Lampert. “Improving Weakly-Supervised Object Localization by Micro-Annotation.” In <i>Proceedings of the British Machine Vision Conference 2016</i>, 2016–September:92.1-92.12. BMVA Press, 2016. <a href=\"https://doi.org/10.5244/C.30.92\">https://doi.org/10.5244/C.30.92</a>.","ista":"Kolesnikov A, Lampert C. 2016. Improving weakly-supervised object localization by micro-annotation. Proceedings of the British Machine Vision Conference 2016. BMVC: British Machine Vision Conference vol. 2016–September, 92.1-92.12.","apa":"Kolesnikov, A., &#38; Lampert, C. (2016). Improving weakly-supervised object localization by micro-annotation. In <i>Proceedings of the British Machine Vision Conference 2016</i> (Vol. 2016–September, p. 92.1-92.12). York, United Kingdom: BMVA Press. <a href=\"https://doi.org/10.5244/C.30.92\">https://doi.org/10.5244/C.30.92</a>","mla":"Kolesnikov, Alexander, and Christoph Lampert. “Improving Weakly-Supervised Object Localization by Micro-Annotation.” <i>Proceedings of the British Machine Vision Conference 2016</i>, vol. 2016–September, BMVA Press, 2016, p. 92.1-92.12, doi:<a href=\"https://doi.org/10.5244/C.30.92\">10.5244/C.30.92</a>."},"user_id":"3E5EF7F0-F248-11E8-B48F-1D18A9856A87","month":"09","department":[{"_id":"ChLa"}],"page":"92.1-92.12","main_file_link":[{"open_access":"1","url":"http://www.bmva.org/bmvc/2016/papers/paper092/paper092.pdf"}],"quality_controlled":"1","doi":"10.5244/C.30.92","publisher":"BMVA Press","date_updated":"2021-01-12T06:48:18Z","_id":"1102","type":"conference","project":[{"name":"Lifelong Learning of Visual Scene Understanding","grant_number":"308036","call_identifier":"FP7","_id":"2532554C-B435-11E9-9278-68D0E5697425"}],"status":"public","publication":"Proceedings of the British Machine Vision Conference 2016","ec_funded":1,"conference":{"location":"York, United Kingdom","end_date":"2016-09-22","start_date":"2016-09-19","name":"BMVC: British Machine Vision Conference"},"acknowledgement":"This work was funded in parts by the European Research Council\r\nunder the European Union’s Seventh Framework Programme (FP7/2007-2013)/ERC grant\r\nagreement no 308036. We gratefully acknowledge the support of NVIDIA Corporation with\r\nthe donation of the GPUs used for this research.","date_published":"2016-09-01T00:00:00Z","year":"2016","publist_id":"6273"},{"date_created":"2018-12-11T11:50:17Z","oa_version":"Published Version","title":"Theoretical foundations of multi-task lifelong learning","day":"01","author":[{"first_name":"Anastasia","last_name":"Pentina","id":"42E87FC6-F248-11E8-B48F-1D18A9856A87","full_name":"Pentina, Anastasia"}],"file_date_updated":"2018-12-12T10:14:07Z","publication_status":"published","publication_identifier":{"issn":["2663-337X"]},"abstract":[{"text":"Traditionally machine learning has been focusing on the problem of solving a single\r\ntask in isolation. While being quite well understood, this approach disregards an\r\nimportant aspect of human learning: when facing a new problem, humans are able to\r\nexploit knowledge acquired from previously learned tasks. Intuitively, access to several\r\nproblems simultaneously or sequentially could also be advantageous for a machine\r\nlearning system, especially if these tasks are closely related. Indeed, results of many\r\nempirical studies have provided justification for this intuition. However, theoretical\r\njustifications of this idea are rather limited.\r\nThe focus of this thesis is to expand the understanding of potential benefits of information\r\ntransfer between several related learning problems. We provide theoretical\r\nanalysis for three scenarios of multi-task learning - multiple kernel learning, sequential\r\nlearning and active task selection. We also provide a PAC-Bayesian perspective on\r\nlifelong learning and investigate how the task generation process influences the generalization\r\nguarantees in this scenario. In addition, we show how some of the obtained\r\ntheoretical results can be used to derive principled multi-task and lifelong learning\r\nalgorithms and illustrate their performance on various synthetic and real-world datasets.","lang":"eng"}],"has_accepted_license":"1","file":[{"date_updated":"2018-12-12T10:14:07Z","creator":"system","file_size":2140062,"date_created":"2018-12-12T10:14:07Z","file_id":"5056","access_level":"open_access","content_type":"application/pdf","file_name":"IST-2017-776-v1+1_Pentina_Thesis_2016.pdf","relation":"main_file"}],"department":[{"_id":"ChLa"}],"month":"11","supervisor":[{"first_name":"Christoph","orcid":"0000-0001-8622-7887","full_name":"Lampert, Christoph","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87","last_name":"Lampert"}],"user_id":"c635000d-4b10-11ee-a964-aac5a93f6ac1","citation":{"mla":"Pentina, Anastasia. <i>Theoretical Foundations of Multi-Task Lifelong Learning</i>. Institute of Science and Technology Austria, 2016, doi:<a href=\"https://doi.org/10.15479/AT:ISTA:TH_776\">10.15479/AT:ISTA:TH_776</a>.","apa":"Pentina, A. (2016). <i>Theoretical foundations of multi-task lifelong learning</i>. Institute of Science and Technology Austria. <a href=\"https://doi.org/10.15479/AT:ISTA:TH_776\">https://doi.org/10.15479/AT:ISTA:TH_776</a>","ista":"Pentina A. 2016. Theoretical foundations of multi-task lifelong learning. Institute of Science and Technology Austria.","chicago":"Pentina, Anastasia. “Theoretical Foundations of Multi-Task Lifelong Learning.” Institute of Science and Technology Austria, 2016. <a href=\"https://doi.org/10.15479/AT:ISTA:TH_776\">https://doi.org/10.15479/AT:ISTA:TH_776</a>.","short":"A. Pentina, Theoretical Foundations of Multi-Task Lifelong Learning, Institute of Science and Technology Austria, 2016.","ieee":"A. Pentina, “Theoretical foundations of multi-task lifelong learning,” Institute of Science and Technology Austria, 2016.","ama":"Pentina A. Theoretical foundations of multi-task lifelong learning. 2016. doi:<a href=\"https://doi.org/10.15479/AT:ISTA:TH_776\">10.15479/AT:ISTA:TH_776</a>"},"pubrep_id":"776","language":[{"iso":"eng"}],"oa":1,"type":"dissertation","_id":"1126","date_updated":"2023-09-07T11:52:03Z","publisher":"Institute of Science and Technology Austria","article_processing_charge":"No","alternative_title":["ISTA Thesis"],"doi":"10.15479/AT:ISTA:TH_776","ddc":["006"],"page":"127","publist_id":"6234","year":"2016","date_published":"2016-11-01T00:00:00Z","acknowledgement":"First and foremost I would like to express my gratitude to my supervisor, Christoph\r\nLampert. Thank you for your patience in teaching me all aspects of doing research\r\n(including English grammar), for your trust in my capabilities and endless support. Thank\r\nyou for granting me freedom in my research and, at the same time, having time and\r\nhelping me cope with the consequences whenever I needed it. Thank you for creating\r\nan excellent atmosphere in the group, it was a great pleasure and honor to be a part of\r\nit. There could not have been a better and more inspiring adviser and mentor.\r\nI thank Shai Ben-David for welcoming me into his group at the University of Waterloo,\r\nfor inspiring discussions and support. It was a great pleasure to work together. I am\r\nalso thankful to Ruth Urner for hosting me at the Max-Planck Institute Tübingen, for the\r\nfruitful collaboration and for taking care of me during that not-so-sunny month of May.\r\nI thank Jan Maas for kindly joining my thesis committee despite the short notice and\r\nproviding me with insightful comments.\r\nI would like to thank my colleagues for their support, entertaining conversations and\r\nendless table soccer games we shared together: Georg, Jan, Amelie and Emilie, Michal\r\nand Alex, Alex K. and Alex Z., Thomas, Sameh, Vlad, Mayu, Nathaniel, Silvester, Neel,\r\nCsaba, Vladimir, Morten. Thank you, Mabel and Ram, for the wonderful time we spent\r\ntogether. I am thankful to Shrinu and Samira for taking care of me during my stay at the\r\nUniversity of Waterloo. Special thanks to Viktoriia for her never-ending optimism and for\r\nbeing so inspiring and supportive, especially at the beginning of my PhD journey.\r\nThanks to IST administration, in particular, Vlad and Elisabeth for shielding me from\r\nmost of the bureaucratic paperwork.\r\n\r\nThis dissertation would not have been possible without funding from the European\r\nResearch Council under the European Union's Seventh Framework Programme\r\n(FP7/2007-2013)/ERC grant agreement no 308036.","ec_funded":1,"status":"public","project":[{"_id":"2532554C-B435-11E9-9278-68D0E5697425","grant_number":"308036","name":"Lifelong Learning of Visual Scene Understanding","call_identifier":"FP7"}],"degree_awarded":"PhD"},{"intvolume":"      9908","abstract":[{"text":"We introduce a new loss function for the weakly-supervised training of semantic image segmentation models based on three guiding principles: to seed with weak localization cues, to expand objects based on the information about which classes can occur in an image, and to constrain the segmentations to coincide with object boundaries. We show experimentally that training a deep convolutional neural network using the proposed loss function leads to substantially better segmentations than previous state-of-the-art methods on the challenging PASCAL VOC 2012 dataset. We furthermore give insight into the working mechanism of our method by a detailed experimental study that illustrates how the segmentation quality is affected by each term of the proposed loss function as well as their combinations.","lang":"eng"}],"publication_status":"published","title":"Seed, expand and constrain: Three principles for weakly-supervised image segmentation","oa_version":"Preprint","author":[{"first_name":"Alexander","last_name":"Kolesnikov","full_name":"Kolesnikov, Alexander","id":"2D157DB6-F248-11E8-B48F-1D18A9856A87"},{"last_name":"Lampert","full_name":"Lampert, Christoph","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87","first_name":"Christoph","orcid":"0000-0001-8622-7887"}],"scopus_import":1,"day":"15","date_created":"2018-12-11T11:51:37Z","volume":9908,"language":[{"iso":"eng"}],"oa":1,"user_id":"3E5EF7F0-F248-11E8-B48F-1D18A9856A87","citation":{"ista":"Kolesnikov A, Lampert C. 2016. Seed, expand and constrain: Three principles for weakly-supervised image segmentation. ECCV: European Conference on Computer Vision, LNCS, vol. 9908, 695–711.","chicago":"Kolesnikov, Alexander, and Christoph Lampert. “Seed, Expand and Constrain: Three Principles for Weakly-Supervised Image Segmentation,” 9908:695–711. Springer, 2016. <a href=\"https://doi.org/10.1007/978-3-319-46493-0_42\">https://doi.org/10.1007/978-3-319-46493-0_42</a>.","mla":"Kolesnikov, Alexander, and Christoph Lampert. <i>Seed, Expand and Constrain: Three Principles for Weakly-Supervised Image Segmentation</i>. Vol. 9908, Springer, 2016, pp. 695–711, doi:<a href=\"https://doi.org/10.1007/978-3-319-46493-0_42\">10.1007/978-3-319-46493-0_42</a>.","apa":"Kolesnikov, A., &#38; Lampert, C. (2016). Seed, expand and constrain: Three principles for weakly-supervised image segmentation (Vol. 9908, pp. 695–711). Presented at the ECCV: European Conference on Computer Vision, Amsterdam, The Netherlands: Springer. <a href=\"https://doi.org/10.1007/978-3-319-46493-0_42\">https://doi.org/10.1007/978-3-319-46493-0_42</a>","ama":"Kolesnikov A, Lampert C. Seed, expand and constrain: Three principles for weakly-supervised image segmentation. In: Vol 9908. Springer; 2016:695-711. doi:<a href=\"https://doi.org/10.1007/978-3-319-46493-0_42\">10.1007/978-3-319-46493-0_42</a>","short":"A. Kolesnikov, C. Lampert, in:, Springer, 2016, pp. 695–711.","ieee":"A. Kolesnikov and C. Lampert, “Seed, expand and constrain: Three principles for weakly-supervised image segmentation,” presented at the ECCV: European Conference on Computer Vision, Amsterdam, The Netherlands, 2016, vol. 9908, pp. 695–711."},"month":"09","department":[{"_id":"ChLa"}],"page":"695 - 711","quality_controlled":"1","main_file_link":[{"url":"https://arxiv.org/abs/1603.06098","open_access":"1"}],"publisher":"Springer","doi":"10.1007/978-3-319-46493-0_42","alternative_title":["LNCS"],"type":"conference","date_updated":"2021-01-12T06:50:12Z","_id":"1369","project":[{"_id":"2532554C-B435-11E9-9278-68D0E5697425","call_identifier":"FP7","name":"Lifelong Learning of Visual Scene Understanding","grant_number":"308036"}],"status":"public","date_published":"2016-09-15T00:00:00Z","conference":{"location":"Amsterdam, The Netherlands","start_date":"2016-10-11","end_date":"2016-10-14","name":"ECCV: European Conference on Computer Vision"},"ec_funded":1,"year":"2016","publist_id":"5842"},{"status":"public","project":[{"grant_number":"308036","name":"Lifelong Learning of Visual Scene Understanding","call_identifier":"FP7","_id":"2532554C-B435-11E9-9278-68D0E5697425"}],"conference":{"location":"Banff, AB, Canada","name":"ALT: Algorithmic Learning Theory","end_date":"2015-10-06","start_date":"2015-10-04"},"date_published":"2015-01-01T00:00:00Z","ec_funded":1,"year":"2015","publist_id":"5430","page":"194 - 208","quality_controlled":"1","main_file_link":[{"url":"http://arxiv.org/abs/1602.06531","open_access":"1"}],"publisher":"Springer","alternative_title":["LNCS"],"doi":"10.1007/978-3-319-24486-0_13","type":"conference","_id":"1706","date_updated":"2021-01-12T06:52:39Z","language":[{"iso":"eng"}],"oa":1,"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","citation":{"apa":"Pentina, A., &#38; Ben David, S. (2015). Multi-task and lifelong learning of kernels (Vol. 9355, pp. 194–208). Presented at the ALT: Algorithmic Learning Theory, Banff, AB, Canada: Springer. <a href=\"https://doi.org/10.1007/978-3-319-24486-0_13\">https://doi.org/10.1007/978-3-319-24486-0_13</a>","mla":"Pentina, Anastasia, and Shai Ben David. <i>Multi-Task and Lifelong Learning of Kernels</i>. Vol. 9355, Springer, 2015, pp. 194–208, doi:<a href=\"https://doi.org/10.1007/978-3-319-24486-0_13\">10.1007/978-3-319-24486-0_13</a>.","chicago":"Pentina, Anastasia, and Shai Ben David. “Multi-Task and Lifelong Learning of Kernels,” 9355:194–208. Springer, 2015. <a href=\"https://doi.org/10.1007/978-3-319-24486-0_13\">https://doi.org/10.1007/978-3-319-24486-0_13</a>.","ista":"Pentina A, Ben David S. 2015. Multi-task and lifelong learning of kernels. ALT: Algorithmic Learning Theory, LNCS, vol. 9355, 194–208.","ieee":"A. Pentina and S. Ben David, “Multi-task and lifelong learning of kernels,” presented at the ALT: Algorithmic Learning Theory, Banff, AB, Canada, 2015, vol. 9355, pp. 194–208.","short":"A. Pentina, S. Ben David, in:, Springer, 2015, pp. 194–208.","ama":"Pentina A, Ben David S. Multi-task and lifelong learning of kernels. In: Vol 9355. Springer; 2015:194-208. doi:<a href=\"https://doi.org/10.1007/978-3-319-24486-0_13\">10.1007/978-3-319-24486-0_13</a>"},"month":"01","department":[{"_id":"ChLa"}],"abstract":[{"lang":"eng","text":"We consider a problem of learning kernels for use in SVM classification in the multi-task and lifelong scenarios and provide generalization bounds on the error of a large margin classifier. Our results show that, under mild conditions on the family of kernels used for learning, solving several related tasks simultaneously is beneficial over single task learning. In particular, as the number of observed tasks grows, assuming that in the considered family of kernels there exists one that yields low approximation error on all tasks, the overhead associated with learning such a kernel vanishes and the complexity converges to that of learning when this good kernel is given to the learner."}],"intvolume":"      9355","publication_status":"published","title":"Multi-task and lifelong learning of kernels","oa_version":"Preprint","day":"01","scopus_import":1,"author":[{"first_name":"Anastasia","id":"42E87FC6-F248-11E8-B48F-1D18A9856A87","full_name":"Pentina, Anastasia","last_name":"Pentina"},{"full_name":"Ben David, Shai","last_name":"Ben David","first_name":"Shai"}],"date_created":"2018-12-11T11:53:35Z","volume":9355},{"abstract":[{"text":"Structural support vector machines (SSVMs) are amongst the best performing models for structured computer vision tasks, such as semantic image segmentation or human pose estimation. Training SSVMs, however, is computationally costly, because it requires repeated calls to a structured prediction subroutine (called \\emph{max-oracle}), which has to solve an optimization problem itself, e.g. a graph cut.\r\nIn this work, we introduce a new algorithm for SSVM training that is more efficient than earlier techniques when the max-oracle is computationally expensive, as it is frequently the case in computer vision tasks. The main idea is to (i) combine the recent stochastic Block-Coordinate Frank-Wolfe algorithm with efficient hyperplane caching, and (ii) use an automatic selection rule for deciding whether to call the exact max-oracle or to rely on an approximate one based on the cached hyperplanes.\r\nWe show experimentally that this strategy leads to faster convergence to the optimum with respect to the number of requires oracle calls, and that this translates into faster convergence with respect to the total runtime when the max-oracle is slow compared to the other steps of the algorithm. ","lang":"eng"}],"page":"2737 - 2745","quality_controlled":"1","publication_status":"published","main_file_link":[{"url":"http://arxiv.org/abs/1408.6804","open_access":"1"}],"oa_version":"Preprint","publisher":"IEEE","title":"A multi-plane block-coordinate Frank-Wolfe algorithm for training structural SVMs with a costly max-oracle","scopus_import":1,"day":"01","doi":"10.1109/CVPR.2015.7298890","author":[{"first_name":"Neel","full_name":"Shah, Neel","id":"31ABAF80-F248-11E8-B48F-1D18A9856A87","last_name":"Shah"},{"first_name":"Vladimir","last_name":"Kolmogorov","full_name":"Kolmogorov, Vladimir","id":"3D50B0BA-F248-11E8-B48F-1D18A9856A87"},{"first_name":"Christoph","orcid":"0000-0001-8622-7887","last_name":"Lampert","id":"40C20FD2-F248-11E8-B48F-1D18A9856A87","full_name":"Lampert, Christoph"}],"date_created":"2018-12-11T11:54:24Z","type":"conference","_id":"1859","date_updated":"2021-01-12T06:53:40Z","language":[{"iso":"eng"}],"status":"public","project":[{"grant_number":"308036","name":"Lifelong Learning of Visual Scene Understanding","call_identifier":"FP7","_id":"2532554C-B435-11E9-9278-68D0E5697425"},{"call_identifier":"FP7","name":"Discrete Optimization in Computer Vision: Theory and Practice","grant_number":"616160","_id":"25FBA906-B435-11E9-9278-68D0E5697425"}],"oa":1,"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","date_published":"2015-06-01T00:00:00Z","conference":{"end_date":"2015-06-12","start_date":"2015-06-07","name":"CVPR: Computer Vision and Pattern Recognition","location":"Boston, MA, USA"},"citation":{"mla":"Shah, Neel, et al. <i>A Multi-Plane Block-Coordinate Frank-Wolfe Algorithm for Training Structural SVMs with a Costly Max-Oracle</i>. IEEE, 2015, pp. 2737–45, doi:<a href=\"https://doi.org/10.1109/CVPR.2015.7298890\">10.1109/CVPR.2015.7298890</a>.","apa":"Shah, N., Kolmogorov, V., &#38; Lampert, C. (2015). A multi-plane block-coordinate Frank-Wolfe algorithm for training structural SVMs with a costly max-oracle (pp. 2737–2745). Presented at the CVPR: Computer Vision and Pattern Recognition, Boston, MA, USA: IEEE. <a href=\"https://doi.org/10.1109/CVPR.2015.7298890\">https://doi.org/10.1109/CVPR.2015.7298890</a>","ista":"Shah N, Kolmogorov V, Lampert C. 2015. A multi-plane block-coordinate Frank-Wolfe algorithm for training structural SVMs with a costly max-oracle. CVPR: Computer Vision and Pattern Recognition, 2737–2745.","chicago":"Shah, Neel, Vladimir Kolmogorov, and Christoph Lampert. “A Multi-Plane Block-Coordinate Frank-Wolfe Algorithm for Training Structural SVMs with a Costly Max-Oracle,” 2737–45. IEEE, 2015. <a href=\"https://doi.org/10.1109/CVPR.2015.7298890\">https://doi.org/10.1109/CVPR.2015.7298890</a>.","short":"N. Shah, V. Kolmogorov, C. Lampert, in:, IEEE, 2015, pp. 2737–2745.","ieee":"N. Shah, V. Kolmogorov, and C. Lampert, “A multi-plane block-coordinate Frank-Wolfe algorithm for training structural SVMs with a costly max-oracle,” presented at the CVPR: Computer Vision and Pattern Recognition, Boston, MA, USA, 2015, pp. 2737–2745.","ama":"Shah N, Kolmogorov V, Lampert C. A multi-plane block-coordinate Frank-Wolfe algorithm for training structural SVMs with a costly max-oracle. In: IEEE; 2015:2737-2745. doi:<a href=\"https://doi.org/10.1109/CVPR.2015.7298890\">10.1109/CVPR.2015.7298890</a>"},"ec_funded":1,"month":"06","year":"2015","publist_id":"5240","department":[{"_id":"VlKo"},{"_id":"ChLa"}]}]
