[{"scopus_import":"1","article_processing_charge":"Yes (via OA deal)","day":"31","doi":"10.1007/s11263-023-01899-3","author":[{"last_name":"Rao","full_name":"Rao, Pramod","first_name":"Pramod"},{"first_name":"B. R.","full_name":"Mallikarjun, B. R.","last_name":"Mallikarjun"},{"full_name":"Fox, Gereon","last_name":"Fox","first_name":"Gereon"},{"last_name":"Weyrich","full_name":"Weyrich, Tim","first_name":"Tim"},{"first_name":"Bernd","orcid":"0000-0001-6511-9385","last_name":"Bickel","id":"49876194-F248-11E8-B48F-1D18A9856A87","full_name":"Bickel, Bernd"},{"first_name":"Hanspeter","last_name":"Pfister","full_name":"Pfister, Hanspeter"},{"first_name":"Wojciech","full_name":"Matusik, Wojciech","last_name":"Matusik"},{"first_name":"Fangneng","full_name":"Zhan, Fangneng","last_name":"Zhan"},{"full_name":"Tewari, Ayush","last_name":"Tewari","first_name":"Ayush"},{"last_name":"Theobalt","full_name":"Theobalt, Christian","first_name":"Christian"},{"full_name":"Elgharib, Mohamed","last_name":"Elgharib","first_name":"Mohamed"}],"publisher":"Springer Nature","oa_version":"Published Version","title":"A deeper analysis of volumetric relightiable faces","_id":"14488","date_updated":"2023-11-06T08:52:30Z","date_created":"2023-11-05T23:00:54Z","type":"journal_article","article_type":"original","abstract":[{"lang":"eng","text":"Portrait viewpoint and illumination editing is an important problem with several applications in VR/AR, movies, and photography. Comprehensive knowledge of geometry and illumination is critical for obtaining photorealistic results. Current methods are unable to explicitly model in 3D while handling both viewpoint and illumination editing from a single image. In this paper, we propose VoRF, a novel approach that can take even a single portrait image as input and relight human heads under novel illuminations that can be viewed from arbitrary viewpoints. VoRF represents a human head as a continuous volumetric field and learns a prior model of human heads using a coordinate-based MLP with individual latent spaces for identity and illumination. The prior model is learned in an auto-decoder manner over a diverse class of head shapes and appearances, allowing VoRF to generalize to novel test identities from a single input image. Additionally, VoRF has a reflectance MLP that uses the intermediate features of the prior model for rendering One-Light-at-A-Time (OLAT) images under novel views. We synthesize novel illuminations by combining these OLAT images with target environment maps. Qualitative and quantitative evaluations demonstrate the effectiveness of VoRF for relighting and novel view synthesis, even when applied to unseen subjects under uncontrolled illumination. This work is an extension of Rao et al. (VoRF: Volumetric Relightable Faces 2022). We provide extensive evaluation and ablative studies of our model and also provide an application, where any face can be relighted using textual input."}],"main_file_link":[{"url":"https://doi.org/10.1007/s11263-023-01899-3","open_access":"1"}],"publication_identifier":{"issn":["0920-5691"],"eissn":["1573-1405"]},"publication_status":"epub_ahead","quality_controlled":"1","year":"2023","month":"10","department":[{"_id":"BeBi"}],"oa":1,"publication":"International Journal of Computer Vision","status":"public","language":[{"iso":"eng"}],"citation":{"ista":"Rao P, Mallikarjun BR, Fox G, Weyrich T, Bickel B, Pfister H, Matusik W, Zhan F, Tewari A, Theobalt C, Elgharib M. 2023. A deeper analysis of volumetric relightiable faces. International Journal of Computer Vision.","chicago":"Rao, Pramod, B. R. Mallikarjun, Gereon Fox, Tim Weyrich, Bernd Bickel, Hanspeter Pfister, Wojciech Matusik, et al. “A Deeper Analysis of Volumetric Relightiable Faces.” <i>International Journal of Computer Vision</i>. Springer Nature, 2023. <a href=\"https://doi.org/10.1007/s11263-023-01899-3\">https://doi.org/10.1007/s11263-023-01899-3</a>.","mla":"Rao, Pramod, et al. “A Deeper Analysis of Volumetric Relightiable Faces.” <i>International Journal of Computer Vision</i>, Springer Nature, 2023, doi:<a href=\"https://doi.org/10.1007/s11263-023-01899-3\">10.1007/s11263-023-01899-3</a>.","apa":"Rao, P., Mallikarjun, B. R., Fox, G., Weyrich, T., Bickel, B., Pfister, H., … Elgharib, M. (2023). A deeper analysis of volumetric relightiable faces. <i>International Journal of Computer Vision</i>. Springer Nature. <a href=\"https://doi.org/10.1007/s11263-023-01899-3\">https://doi.org/10.1007/s11263-023-01899-3</a>","ama":"Rao P, Mallikarjun BR, Fox G, et al. A deeper analysis of volumetric relightiable faces. <i>International Journal of Computer Vision</i>. 2023. doi:<a href=\"https://doi.org/10.1007/s11263-023-01899-3\">10.1007/s11263-023-01899-3</a>","short":"P. Rao, B.R. Mallikarjun, G. Fox, T. Weyrich, B. Bickel, H. Pfister, W. Matusik, F. Zhan, A. Tewari, C. Theobalt, M. Elgharib, International Journal of Computer Vision (2023).","ieee":"P. Rao <i>et al.</i>, “A deeper analysis of volumetric relightiable faces,” <i>International Journal of Computer Vision</i>. Springer Nature, 2023."},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","acknowledgement":"Open Access funding enabled and organized by Projekt DEAL.","date_published":"2023-10-31T00:00:00Z"},{"article_type":"original","date_created":"2023-11-29T15:02:03Z","volume":42,"title":"Procedural metamaterials: A unified procedural graph for metamaterial design","oa_version":"Published Version","author":[{"last_name":"Makatura","full_name":"Makatura, Liane","first_name":"Liane"},{"first_name":"Bohan","last_name":"Wang","full_name":"Wang, Bohan"},{"id":"0b467602-dbcd-11ea-9d1d-ed480aa46b70","full_name":"Chen, Yi-Lu","last_name":"Chen","first_name":"Yi-Lu"},{"first_name":"Bolei","full_name":"Deng, Bolei","last_name":"Deng"},{"orcid":"0000-0001-6646-5546","first_name":"Christopher J","full_name":"Wojtan, Christopher J","id":"3C61F1D2-F248-11E8-B48F-1D18A9856A87","last_name":"Wojtan"},{"first_name":"Bernd","orcid":"0000-0001-6511-9385","full_name":"Bickel, Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87","last_name":"Bickel"},{"last_name":"Matusik","full_name":"Matusik, Wojciech","first_name":"Wojciech"}],"day":"01","publication_identifier":{"issn":["0730-0301","1557-7368"]},"publication_status":"published","file_date_updated":"2023-12-04T08:04:14Z","abstract":[{"lang":"eng","text":"We introduce a compact, intuitive procedural graph representation for cellular metamaterials, which are small-scale, tileable structures that can be architected to exhibit many useful material properties. Because the structures’ “architectures” vary widely—with elements such as beams, thin shells, and solid bulks—it is difficult to explore them using existing representations. Generic approaches like voxel grids are versatile, but it is cumbersome to represent and edit individual structures; architecture-specific approaches address these issues, but are incompatible with one another. By contrast, our procedural graph succinctly represents the construction process for any structure using a simple skeleton annotated with spatially varying thickness. To express the highly constrained triply periodic minimal surfaces (TPMS) in this manner, we present the first fully automated version of the conjugate surface construction method, which allows novices to create complex TPMS from intuitive input. We demonstrate our representation’s expressiveness, accuracy, and compactness by constructing a wide range of established structures and hundreds of novel structures with diverse architectures and material properties. We also conduct a user study to verify our representation’s ease-of-use and ability to expand engineers’ capacity for exploration."}],"intvolume":"        42","has_accepted_license":"1","article_number":"168","file":[{"access_level":"open_access","content_type":"application/zip","success":1,"file_name":"tog-22-0089-File004.zip","checksum":"0192f597d7a2ceaf89baddfd6190d4c8","relation":"main_file","date_updated":"2023-11-29T15:16:01Z","creator":"yichen","file_size":95467870,"date_created":"2023-11-29T15:16:01Z","file_id":"14630"},{"success":1,"file_name":"tog-22-0089-File005.zip","access_level":"open_access","content_type":"application/zip","relation":"main_file","checksum":"7fb024963be81933494f38de191e4710","file_size":103731880,"date_created":"2023-11-29T15:16:01Z","date_updated":"2023-11-29T15:16:01Z","creator":"yichen","file_id":"14631"},{"file_id":"14638","date_updated":"2023-12-04T08:04:14Z","creator":"dernst","date_created":"2023-12-04T08:04:14Z","file_size":57067476,"checksum":"b7d6829ce396e21cac9fae0ec7130a6b","relation":"main_file","content_type":"application/pdf","access_level":"open_access","file_name":"2023_ACMToG_Makatura.pdf","success":1}],"department":[{"_id":"GradSch"},{"_id":"ChWo"},{"_id":"BeBi"}],"month":"10","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","issue":"5","citation":{"mla":"Makatura, Liane, et al. “Procedural Metamaterials: A Unified Procedural Graph for Metamaterial Design.” <i>ACM Transactions on Graphics</i>, vol. 42, no. 5, 168, Association for Computing Machinery, 2023, doi:<a href=\"https://doi.org/10.1145/3605389\">10.1145/3605389</a>.","apa":"Makatura, L., Wang, B., Chen, Y.-L., Deng, B., Wojtan, C., Bickel, B., &#38; Matusik, W. (2023). Procedural metamaterials: A unified procedural graph for metamaterial design. <i>ACM Transactions on Graphics</i>. Association for Computing Machinery. <a href=\"https://doi.org/10.1145/3605389\">https://doi.org/10.1145/3605389</a>","ista":"Makatura L, Wang B, Chen Y-L, Deng B, Wojtan C, Bickel B, Matusik W. 2023. Procedural metamaterials: A unified procedural graph for metamaterial design. ACM Transactions on Graphics. 42(5), 168.","chicago":"Makatura, Liane, Bohan Wang, Yi-Lu Chen, Bolei Deng, Chris Wojtan, Bernd Bickel, and Wojciech Matusik. “Procedural Metamaterials: A Unified Procedural Graph for Metamaterial Design.” <i>ACM Transactions on Graphics</i>. Association for Computing Machinery, 2023. <a href=\"https://doi.org/10.1145/3605389\">https://doi.org/10.1145/3605389</a>.","short":"L. Makatura, B. Wang, Y.-L. Chen, B. Deng, C. Wojtan, B. Bickel, W. Matusik, ACM Transactions on Graphics 42 (2023).","ieee":"L. Makatura <i>et al.</i>, “Procedural metamaterials: A unified procedural graph for metamaterial design,” <i>ACM Transactions on Graphics</i>, vol. 42, no. 5. Association for Computing Machinery, 2023.","ama":"Makatura L, Wang B, Chen Y-L, et al. Procedural metamaterials: A unified procedural graph for metamaterial design. <i>ACM Transactions on Graphics</i>. 2023;42(5). doi:<a href=\"https://doi.org/10.1145/3605389\">10.1145/3605389</a>"},"language":[{"iso":"eng"}],"oa":1,"type":"journal_article","date_updated":"2023-12-04T08:09:05Z","_id":"14628","publisher":"Association for Computing Machinery","doi":"10.1145/3605389","article_processing_charge":"Yes (in subscription journal)","quality_controlled":"1","ddc":["531","006"],"keyword":["Computer Graphics and Computer-Aided Design"],"year":"2023","date_published":"2023-10-01T00:00:00Z","acknowledgement":"The authors thank Mina Konaković Luković and Michael Foshey for their early contributions to this project, David Palmer and Paul Zhang for their insightful discussions about minimal surfaces and the CSCM, Julian Panetta for providing the Elastic Textures code, and Hannes Hergeth for his feedback and support. We also thank our user study participants and anonymous reviewers.\r\nThis material is based upon work supported by the National Science Foundation\r\n(NSF) Graduate Research Fellowship under Grant No. 2141064; the MIT Morningside\r\nAcademy for Design Fellowship; the Defense Advanced Research Projects Agency\r\n(DARPA) Grant No. FA8750-20-C-0075; the ERC Consolidator Grant No. 101045083,\r\n“CoDiNA: Computational Discovery of Numerical Algorithms for Animation and Simulation of Natural Phenomena”; and the NewSat project, which is co-funded by the Operational Program for Competitiveness and Internationalisation (COMPETE2020), Portugal 2020, the European Regional Development Fund (ERDF), and the Portuguese Foundation for Science and Technology (FTC) under the MIT Portugal program.","project":[{"_id":"34bc2376-11ca-11ed-8bc3-9a3b3961a088","name":"Computational Discovery of Numerical Algorithms for Animation and Simulation of Natural Phenomena","grant_number":"101045083"}],"status":"public","publication":"ACM Transactions on Graphics"},{"acknowledgement":"We thank the anonymous reviewers for their generous feedback, and Julian Fischer for his help in proving Proposition 1. This project has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement No. 715767).","date_published":"2023-09-20T00:00:00Z","ec_funded":1,"publication":"ACM Transactions on Graphics","status":"public","project":[{"_id":"24F9549A-B435-11E9-9278-68D0E5697425","name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","grant_number":"715767","call_identifier":"H2020"}],"keyword":["Computer Graphics","Computational Design","Computational Geometry","Shape Modeling"],"related_material":{"record":[{"status":"public","relation":"part_of_dissertation","id":"12897"}]},"external_id":{"isi":["001086833300010"]},"isi":1,"year":"2023","quality_controlled":"1","ddc":["516"],"type":"journal_article","_id":"13188","date_updated":"2024-03-25T23:30:26Z","publisher":"Association for Computing Machinery","article_processing_charge":"No","doi":"10.1145/3606033","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","citation":{"chicago":"Hafner, Christian, and Bernd Bickel. “The Design Space of Kirchhoff Rods.” <i>ACM Transactions on Graphics</i>. Association for Computing Machinery, 2023. <a href=\"https://doi.org/10.1145/3606033\">https://doi.org/10.1145/3606033</a>.","ista":"Hafner C, Bickel B. 2023. The design space of Kirchhoff rods. ACM Transactions on Graphics. 42(5), 171.","mla":"Hafner, Christian, and Bernd Bickel. “The Design Space of Kirchhoff Rods.” <i>ACM Transactions on Graphics</i>, vol. 42, no. 5, 171, Association for Computing Machinery, 2023, doi:<a href=\"https://doi.org/10.1145/3606033\">10.1145/3606033</a>.","apa":"Hafner, C., &#38; Bickel, B. (2023). The design space of Kirchhoff rods. <i>ACM Transactions on Graphics</i>. Association for Computing Machinery. <a href=\"https://doi.org/10.1145/3606033\">https://doi.org/10.1145/3606033</a>","ama":"Hafner C, Bickel B. The design space of Kirchhoff rods. <i>ACM Transactions on Graphics</i>. 2023;42(5). doi:<a href=\"https://doi.org/10.1145/3606033\">10.1145/3606033</a>","short":"C. Hafner, B. Bickel, ACM Transactions on Graphics 42 (2023).","ieee":"C. Hafner and B. Bickel, “The design space of Kirchhoff rods,” <i>ACM Transactions on Graphics</i>, vol. 42, no. 5. Association for Computing Machinery, 2023."},"issue":"5","language":[{"iso":"eng"}],"oa":1,"file":[{"checksum":"4954c1cfa487725bc156dcfec872478a","relation":"main_file","access_level":"open_access","content_type":"application/pdf","success":1,"file_name":"kirchhoff-rods.pdf","file_id":"13194","date_updated":"2023-07-04T08:11:28Z","creator":"chafner","date_created":"2023-07-04T08:11:28Z","file_size":19635168},{"file_name":"supp-main.pdf","access_level":"open_access","content_type":"application/pdf","relation":"supplementary_material","checksum":"79c9975fbc82ff71f1767331d2204cca","date_created":"2023-07-04T07:46:28Z","file_size":420909,"date_updated":"2023-07-04T07:46:28Z","creator":"chafner","file_id":"13190","title":"Supplemental Material with Proofs"},{"date_updated":"2023-07-04T07:46:30Z","creator":"chafner","date_created":"2023-07-04T07:46:30Z","file_size":430086,"file_id":"13191","title":"Cheat Sheet for Notation","access_level":"open_access","content_type":"application/pdf","file_name":"supp-cheat.pdf","checksum":"4ab647e4f03c711e1e6a5fc1eb8684db","relation":"supplementary_material"},{"title":"Supplemental Video","file_id":"13192","date_updated":"2023-07-04T07:46:39Z","creator":"chafner","date_created":"2023-07-04T07:46:39Z","file_size":268088064,"checksum":"c0fd9a57d012046de90c185ffa904b76","relation":"supplementary_material","content_type":"video/mp4","access_level":"open_access","file_name":"kirchhoff-video-final.mp4"},{"file_id":"13193","title":"Matlab Source Code with Example","file_size":25790,"date_created":"2023-07-04T07:47:10Z","creator":"chafner","date_updated":"2023-07-04T07:47:10Z","relation":"supplementary_material","checksum":"71b00712b489ada2cd9815910ee180a9","file_name":"matlab-submission.zip","access_level":"open_access","content_type":"application/x-zip-compressed"}],"article_number":"171","department":[{"_id":"BeBi"}],"month":"09","file_date_updated":"2023-07-04T08:11:28Z","publication_status":"published","publication_identifier":{"eissn":["1557-7368"],"issn":["0730-0301"]},"acknowledged_ssus":[{"_id":"M-Shop"}],"abstract":[{"text":"The Kirchhoff rod model describes the bending and twisting of slender elastic rods in three dimensions, and has been widely studied to enable the prediction of how a rod will deform, given its geometry and boundary conditions. In this work, we study a number of inverse problems with the goal of computing the geometry of a straight rod that will automatically deform to match a curved target shape after attaching its endpoints to a support structure. Our solution lets us finely control the static equilibrium state of a rod by varying the cross-sectional profiles along its length.\r\nWe also show that the set of physically realizable equilibrium states admits a concise geometric description in terms of linear line complexes, which leads to very efficient computational design algorithms. Implemented in an interactive software tool, they allow us to convert three-dimensional hand-drawn spline curves to elastic rods, and give feedback about the feasibility and practicality of a design in real time. We demonstrate the efficacy of our method by designing and manufacturing several physical prototypes with applications to interior design and soft robotics.","lang":"eng"}],"intvolume":"        42","has_accepted_license":"1","date_created":"2023-07-04T07:41:30Z","article_type":"original","volume":42,"title":"The design space of Kirchhoff rods","oa_version":"Submitted Version","day":"20","author":[{"first_name":"Christian","last_name":"Hafner","id":"400429CC-F248-11E8-B48F-1D18A9856A87","full_name":"Hafner, Christian"},{"last_name":"Bickel","full_name":"Bickel, Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87","orcid":"0000-0001-6511-9385","first_name":"Bernd"}]},{"publisher":"Springer Nature","article_processing_charge":"Yes","doi":"10.1038/s41592-023-01936-6","type":"journal_article","_id":"13267","date_updated":"2024-01-10T08:37:48Z","page":"1256-1265","quality_controlled":"1","main_file_link":[{"open_access":"1","url":"https://doi.org/10.1038/s41592-023-01936-6"}],"related_material":{"link":[{"url":"https://github.com/danzllab/LIONESS","relation":"software"}],"record":[{"id":"12817","status":"public","relation":"research_data"},{"relation":"shorter_version","status":"public","id":"14770"}]},"external_id":{"pmid":["37429995"],"isi":["001025621500001"]},"isi":1,"year":"2023","publication":"Nature Methods","status":"public","project":[{"_id":"265CB4D0-B435-11E9-9278-68D0E5697425","name":"Optical control of synaptic function via adhesion molecules","grant_number":"I03600","call_identifier":"FWF"},{"_id":"2548AE96-B435-11E9-9278-68D0E5697425","name":"Molecular Drug Targets","grant_number":"W1232-B24","call_identifier":"FWF"},{"name":"The Wittgenstein Prize","grant_number":"Z00312","call_identifier":"FWF","_id":"25C5A090-B435-11E9-9278-68D0E5697425"},{"_id":"23889792-32DE-11EA-91FC-C7463DDC885E","name":"High content imaging to decode human immune cell interactions in health and allergic disease"},{"_id":"2564DBCA-B435-11E9-9278-68D0E5697425","call_identifier":"H2020","grant_number":"665385","name":"International IST Doctoral Program"},{"grant_number":"715767","name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","call_identifier":"H2020","_id":"24F9549A-B435-11E9-9278-68D0E5697425"},{"_id":"25444568-B435-11E9-9278-68D0E5697425","grant_number":"715508","name":"Probing the Reversibility of Autism Spectrum Disorders by Employing in vivo and in vitro Models","call_identifier":"H2020"},{"_id":"25B7EB9E-B435-11E9-9278-68D0E5697425","grant_number":"692692","name":"Biophysics and circuit function of a giant cortical glumatergic synapse","call_identifier":"H2020"},{"name":"Synaptic computations of the hippocampal CA3 circuitry","grant_number":"101026635","call_identifier":"H2020","_id":"fc2be41b-9c52-11eb-aca3-faa90aa144e9"},{"_id":"2668BFA0-B435-11E9-9278-68D0E5697425","name":"High-speed 3D-nanoscopy to study the role of adhesion during 3D cell migration","grant_number":"LT00057"}],"acknowledgement":"We thank J. Vorlaufer, N. Agudelo and A. Wartak for microscope maintenance and troubleshooting, C. Kreuzinger and A. Freeman for technical assistance, M. Šuplata for hardware control support and M. Cunha dos Santos for initial exploration of software. We\r\nthank P. Henderson for advice on deep-learning training and M. Sixt, S. Boyd and T. Weiss for discussions and critical reading of the manuscript. L. Lavis (Janelia Research Campus) generously provided the JF585-HaloTag ligand. We acknowledge expert support by IST\r\nAustria’s scientific computing, imaging and optics, preclinical, library and laboratory support facilities and by the Miba machine shop. We gratefully acknowledge funding by the following sources: Austrian Science Fund (F.W.F.) grant no. I3600-B27 (J.G.D.), grant no. DK W1232\r\n(J.G.D. and J.M.M.) and grant no. Z 312-B27, Wittgenstein award (P.J.); the Gesellschaft für Forschungsförderung NÖ grant no. LSC18-022 (J.G.D.); an ISTA Interdisciplinary project grant (J.G.D. and B.B.); the European Union’s Horizon 2020 research and innovation programme,\r\nMarie-Skłodowska Curie grant 665385 (J.M.M. and J.L.); the European Union’s Horizon 2020 research and innovation programme, European Research Council grant no. 715767, MATERIALIZABLE (B.B.); grant no. 715508, REVERSEAUTISM (G.N.); grant no. 695568, SYNNOVATE (S.G.N.G.); and grant no. 692692, GIANTSYN (P.J.); the Simons\r\nFoundation Autism Research Initiative grant no. 529085 (S.G.N.G.); the Wellcome Trust Technology Development grant no. 202932 (S.G.N.G.); the Marie Skłodowska-Curie Actions Individual Fellowship no. 101026635 under the EU Horizon 2020 program (J.F.W.);\r\nthe Human Frontier Science Program postdoctoral fellowship LT000557/2018 (W.J.); and the National Science Foundation grant no. IIS-1835231 (H.P.) and NCS-FO-2124179 (H.P.).","date_published":"2023-08-01T00:00:00Z","ec_funded":1,"pmid":1,"oa_version":"Published Version","title":"Dense 4D nanoscale reconstruction of living brain tissue","day":"01","scopus_import":"1","author":[{"id":"39BDC62C-F248-11E8-B48F-1D18A9856A87","full_name":"Velicky, Philipp","last_name":"Velicky","first_name":"Philipp","orcid":"0000-0002-2340-7431"},{"full_name":"Miguel Villalba, Eder","id":"3FB91342-F248-11E8-B48F-1D18A9856A87","last_name":"Miguel Villalba","orcid":"0000-0001-5665-0430","first_name":"Eder"},{"orcid":"0000-0003-3862-1235","first_name":"Julia M","id":"443DB6DE-F248-11E8-B48F-1D18A9856A87","full_name":"Michalska, Julia M","last_name":"Michalska"},{"full_name":"Lyudchik, Julia","id":"46E28B80-F248-11E8-B48F-1D18A9856A87","last_name":"Lyudchik","first_name":"Julia"},{"full_name":"Wei, Donglai","last_name":"Wei","first_name":"Donglai"},{"full_name":"Lin, Zudi","last_name":"Lin","first_name":"Zudi"},{"full_name":"Watson, Jake","id":"63836096-4690-11EA-BD4E-32803DDC885E","last_name":"Watson","orcid":"0000-0002-8698-3823","first_name":"Jake"},{"full_name":"Troidl, Jakob","last_name":"Troidl","first_name":"Jakob"},{"last_name":"Beyer","full_name":"Beyer, Johanna","first_name":"Johanna"},{"first_name":"Yoav","last_name":"Ben Simon","full_name":"Ben Simon, Yoav","id":"43DF3136-F248-11E8-B48F-1D18A9856A87"},{"first_name":"Christoph M","orcid":"0000-0003-1216-9105","last_name":"Sommer","id":"4DF26D8C-F248-11E8-B48F-1D18A9856A87","full_name":"Sommer, Christoph M"},{"first_name":"Wiebke","full_name":"Jahr, Wiebke","id":"425C1CE8-F248-11E8-B48F-1D18A9856A87","last_name":"Jahr"},{"first_name":"Alban","id":"9ac8f577-2357-11eb-997a-e566c5550886","full_name":"Cenameri, Alban","last_name":"Cenameri"},{"first_name":"Johannes","full_name":"Broichhagen, Johannes","last_name":"Broichhagen"},{"last_name":"Grant","full_name":"Grant, Seth G.N.","first_name":"Seth G.N."},{"first_name":"Peter M","orcid":"0000-0001-5001-4804","last_name":"Jonas","full_name":"Jonas, Peter M","id":"353C1B58-F248-11E8-B48F-1D18A9856A87"},{"first_name":"Gaia","orcid":"0000-0002-7673-7178","last_name":"Novarino","full_name":"Novarino, Gaia","id":"3E57A680-F248-11E8-B48F-1D18A9856A87"},{"full_name":"Pfister, Hanspeter","last_name":"Pfister","first_name":"Hanspeter"},{"last_name":"Bickel","full_name":"Bickel, Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87","orcid":"0000-0001-6511-9385","first_name":"Bernd"},{"orcid":"0000-0001-8559-3973","first_name":"Johann G","last_name":"Danzl","full_name":"Danzl, Johann G","id":"42EFD3B6-F248-11E8-B48F-1D18A9856A87"}],"date_created":"2023-07-23T22:01:13Z","article_type":"original","volume":20,"acknowledged_ssus":[{"_id":"ScienComp"},{"_id":"Bio"},{"_id":"PreCl"},{"_id":"E-Lib"},{"_id":"LifeSc"},{"_id":"M-Shop"}],"intvolume":"        20","abstract":[{"lang":"eng","text":"Three-dimensional (3D) reconstruction of living brain tissue down to an individual synapse level would create opportunities for decoding the dynamics and structure–function relationships of the brain’s complex and dense information processing network; however, this has been hindered by insufficient 3D resolution, inadequate signal-to-noise ratio and prohibitive light burden in optical imaging, whereas electron microscopy is inherently static. Here we solved these challenges by developing an integrated optical/machine-learning technology, LIONESS (live information-optimized nanoscopy enabling saturated segmentation). This leverages optical modifications to stimulated emission depletion microscopy in comprehensively, extracellularly labeled tissue and previous information on sample structure via machine learning to simultaneously achieve isotropic super-resolution, high signal-to-noise ratio and compatibility with living tissue. This allows dense deep-learning-based instance segmentation and 3D reconstruction at a synapse level, incorporating molecular, activity and morphodynamic information. LIONESS opens up avenues for studying the dynamic functional (nano-)architecture of living brain tissue."}],"publication_status":"published","publication_identifier":{"eissn":["1548-7105"],"issn":["1548-7091"]},"month":"08","department":[{"_id":"PeJo"},{"_id":"GaNo"},{"_id":"BeBi"},{"_id":"JoDa"},{"_id":"Bio"}],"language":[{"iso":"eng"}],"oa":1,"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","citation":{"ama":"Velicky P, Miguel Villalba E, Michalska JM, et al. Dense 4D nanoscale reconstruction of living brain tissue. <i>Nature Methods</i>. 2023;20:1256-1265. doi:<a href=\"https://doi.org/10.1038/s41592-023-01936-6\">10.1038/s41592-023-01936-6</a>","ieee":"P. Velicky <i>et al.</i>, “Dense 4D nanoscale reconstruction of living brain tissue,” <i>Nature Methods</i>, vol. 20. Springer Nature, pp. 1256–1265, 2023.","short":"P. Velicky, E. Miguel Villalba, J.M. Michalska, J. Lyudchik, D. Wei, Z. Lin, J. Watson, J. Troidl, J. Beyer, Y. Ben Simon, C.M. Sommer, W. Jahr, A. Cenameri, J. Broichhagen, S.G.N. Grant, P.M. Jonas, G. Novarino, H. Pfister, B. Bickel, J.G. Danzl, Nature Methods 20 (2023) 1256–1265.","chicago":"Velicky, Philipp, Eder Miguel Villalba, Julia M Michalska, Julia Lyudchik, Donglai Wei, Zudi Lin, Jake Watson, et al. “Dense 4D Nanoscale Reconstruction of Living Brain Tissue.” <i>Nature Methods</i>. Springer Nature, 2023. <a href=\"https://doi.org/10.1038/s41592-023-01936-6\">https://doi.org/10.1038/s41592-023-01936-6</a>.","ista":"Velicky P, Miguel Villalba E, Michalska JM, Lyudchik J, Wei D, Lin Z, Watson J, Troidl J, Beyer J, Ben Simon Y, Sommer CM, Jahr W, Cenameri A, Broichhagen J, Grant SGN, Jonas PM, Novarino G, Pfister H, Bickel B, Danzl JG. 2023. Dense 4D nanoscale reconstruction of living brain tissue. Nature Methods. 20, 1256–1265.","apa":"Velicky, P., Miguel Villalba, E., Michalska, J. M., Lyudchik, J., Wei, D., Lin, Z., … Danzl, J. G. (2023). Dense 4D nanoscale reconstruction of living brain tissue. <i>Nature Methods</i>. Springer Nature. <a href=\"https://doi.org/10.1038/s41592-023-01936-6\">https://doi.org/10.1038/s41592-023-01936-6</a>","mla":"Velicky, Philipp, et al. “Dense 4D Nanoscale Reconstruction of Living Brain Tissue.” <i>Nature Methods</i>, vol. 20, Springer Nature, 2023, pp. 1256–65, doi:<a href=\"https://doi.org/10.1038/s41592-023-01936-6\">10.1038/s41592-023-01936-6</a>."}},{"oa":1,"language":[{"iso":"eng"}],"citation":{"ista":"Tojo K, Shamir A, Bickel B, Umetani N. 2023. Stealth shaper: Reflectivity optimization as surface stylization. SIGGRAPH 2023 Conference Proceedings. SIGGRAPH: Computer Graphics and Interactive Techniques Conference, 20.","chicago":"Tojo, Kenji, Ariel Shamir, Bernd Bickel, and Nobuyuki Umetani. “Stealth Shaper: Reflectivity Optimization as Surface Stylization.” In <i>SIGGRAPH 2023 Conference Proceedings</i>. Association for Computing Machinery, 2023. <a href=\"https://doi.org/10.1145/3588432.3591542\">https://doi.org/10.1145/3588432.3591542</a>.","apa":"Tojo, K., Shamir, A., Bickel, B., &#38; Umetani, N. (2023). Stealth shaper: Reflectivity optimization as surface stylization. In <i>SIGGRAPH 2023 Conference Proceedings</i>. Los Angeles, CA, United States: Association for Computing Machinery. <a href=\"https://doi.org/10.1145/3588432.3591542\">https://doi.org/10.1145/3588432.3591542</a>","mla":"Tojo, Kenji, et al. “Stealth Shaper: Reflectivity Optimization as Surface Stylization.” <i>SIGGRAPH 2023 Conference Proceedings</i>, 20, Association for Computing Machinery, 2023, doi:<a href=\"https://doi.org/10.1145/3588432.3591542\">10.1145/3588432.3591542</a>.","ama":"Tojo K, Shamir A, Bickel B, Umetani N. Stealth shaper: Reflectivity optimization as surface stylization. In: <i>SIGGRAPH 2023 Conference Proceedings</i>. Association for Computing Machinery; 2023. doi:<a href=\"https://doi.org/10.1145/3588432.3591542\">10.1145/3588432.3591542</a>","ieee":"K. Tojo, A. Shamir, B. Bickel, and N. Umetani, “Stealth shaper: Reflectivity optimization as surface stylization,” in <i>SIGGRAPH 2023 Conference Proceedings</i>, Los Angeles, CA, United States, 2023.","short":"K. Tojo, A. Shamir, B. Bickel, N. Umetani, in:, SIGGRAPH 2023 Conference Proceedings, Association for Computing Machinery, 2023."},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","month":"07","arxiv":1,"department":[{"_id":"BeBi"}],"article_number":"20","abstract":[{"lang":"eng","text":"We present a technique to optimize the reflectivity of a surface while preserving its overall shape. The naïve optimization of the mesh vertices using the gradients of reflectivity simulations results in undesirable distortion. In contrast, our robust formulation optimizes the surface normal as an independent variable that bridges the reflectivity term with differential rendering, and the regularization term with as-rigid-as-possible elastic energy. We further adaptively subdivide the input mesh to improve the convergence. Consequently, our method can minimize the retroreflectivity of a wide range of input shapes, resulting in sharply creased shapes ubiquitous among stealth aircraft and Sci-Fi vehicles. Furthermore, by changing the reward for the direction of the outgoing light directions, our method can be applied to other reflectivity design tasks, such as the optimization of architectural walls to concentrate light in a specific region. We have tested the proposed method using light-transport simulations and real-world 3D-printed objects."}],"publication_status":"published","publication_identifier":{"isbn":["9798400701597"]},"author":[{"first_name":"Kenji","full_name":"Tojo, Kenji","last_name":"Tojo"},{"first_name":"Ariel","full_name":"Shamir, Ariel","last_name":"Shamir"},{"last_name":"Bickel","id":"49876194-F248-11E8-B48F-1D18A9856A87","full_name":"Bickel, Bernd","orcid":"0000-0001-6511-9385","first_name":"Bernd"},{"last_name":"Umetani","full_name":"Umetani, Nobuyuki","first_name":"Nobuyuki"}],"day":"23","scopus_import":"1","title":"Stealth shaper: Reflectivity optimization as surface stylization","oa_version":"Preprint","date_created":"2023-08-27T22:01:17Z","publication":"SIGGRAPH 2023 Conference Proceedings","status":"public","conference":{"start_date":"2023-08-06","end_date":"2023-08-10","name":"SIGGRAPH: Computer Graphics and Interactive Techniques Conference","location":"Los Angeles, CA, United States"},"acknowledgement":"The authors would like to thank Yuki Koyama and Takeo Igarashi for early discussions, and Yuta Yaguchi for support in 3D printing. This research is partially supported by the Israel Science Foundation grant number 1390/19.\r\n","date_published":"2023-07-23T00:00:00Z","year":"2023","external_id":{"arxiv":["2305.05944"]},"main_file_link":[{"open_access":"1","url":"https://doi.org/10.48550/arXiv.2305.05944"}],"quality_controlled":"1","doi":"10.1145/3588432.3591542","article_processing_charge":"No","publisher":"Association for Computing Machinery","date_updated":"2023-09-05T07:22:03Z","_id":"14241","type":"conference"},{"file_date_updated":"2023-05-16T08:28:37Z","publication_identifier":{"issn":["1467-8659"]},"publication_status":"published","tmp":{"image":"/images/cc_by_nc_nd.png","name":"Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0)","legal_code_url":"https://creativecommons.org/licenses/by-nc-nd/4.0/legalcode","short":"CC BY-NC-ND (4.0)"},"intvolume":"        42","abstract":[{"lang":"eng","text":"Embroidery is a long-standing and high-quality approach to making logos and images on textiles. Nowadays, it can also be performed via automated machines that weave threads with high spatial accuracy. A characteristic feature of the appearance of the threads is a high degree of anisotropy. The anisotropic behavior is caused by depositing thin but long strings of thread. As a result, the stitched patterns convey both color and direction. Artists leverage this anisotropic behavior to enhance pure color images with textures, illusions of motion, or depth cues. However, designing colorful embroidery patterns with prescribed directionality is a challenging task, one usually requiring an expert designer. In this work, we propose an interactive algorithm that generates machine-fabricable embroidery patterns from multi-chromatic images equipped with user-specified directionality fields.We cast the problem of finding a stitching pattern into vector theory. To find a suitable stitching pattern, we extract sources and sinks from the divergence field of the vector field extracted from the input and use them to trace streamlines. We further optimize the streamlines to guarantee a smooth and connected stitching pattern. The generated patterns approximate the color distribution constrained by the directionality field. To allow for further artistic control, the trade-off between color match and directionality match can be interactively explored via an intuitive slider. We showcase our approach by fabricating several embroidery paths."}],"has_accepted_license":"1","date_created":"2023-05-16T08:47:25Z","article_type":"original","volume":42,"title":"Directionality-aware design of embroidery patterns","oa_version":"Published Version","day":"08","author":[{"last_name":"Liu","full_name":"Liu, Zhenyuan","id":"70f0d7cf-ae65-11ec-a14f-89dfc5505b19","first_name":"Zhenyuan","orcid":"0000-0001-9200-5690"},{"last_name":"Piovarci","full_name":"Piovarci, Michael","id":"62E473F4-5C99-11EA-A40E-AF823DDC885E","first_name":"Michael"},{"last_name":"Hafner","id":"400429CC-F248-11E8-B48F-1D18A9856A87","full_name":"Hafner, Christian","first_name":"Christian"},{"last_name":"Charrondiere","id":"a3a24133-2cc7-11ec-be88-8ddaf6f464b1","full_name":"Charrondiere, Raphael","first_name":"Raphael"},{"orcid":"0000-0001-6511-9385","first_name":"Bernd","last_name":"Bickel","full_name":"Bickel, Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87"}],"user_id":"4359f0d1-fa6c-11eb-b949-802e58b17ae8","citation":{"apa":"Liu, Z., Piovarci, M., Hafner, C., Charrondiere, R., &#38; Bickel, B. (2023). Directionality-aware design of embroidery patterns. <i>Computer Graphics Forum</i>. Saarbrucken, Germany: Wiley. <a href=\"https://doi.org/10.1111/cgf.14770 \">https://doi.org/10.1111/cgf.14770 </a>","mla":"Liu, Zhenyuan, et al. “Directionality-Aware Design of Embroidery Patterns.” <i>Computer Graphics Forum</i>, vol. 42, no. 2, Wiley, 2023, pp. 397–409, doi:<a href=\"https://doi.org/10.1111/cgf.14770 \">10.1111/cgf.14770 </a>.","chicago":"Liu, Zhenyuan, Michael Piovarci, Christian Hafner, Raphael Charrondiere, and Bernd Bickel. “Directionality-Aware Design of Embroidery Patterns.” <i>Computer Graphics Forum</i>. Wiley, 2023. <a href=\"https://doi.org/10.1111/cgf.14770 \">https://doi.org/10.1111/cgf.14770 </a>.","ista":"Liu Z, Piovarci M, Hafner C, Charrondiere R, Bickel B. 2023. Directionality-aware design of embroidery patterns. Computer Graphics Forum. 42(2), 397–409.","ieee":"Z. Liu, M. Piovarci, C. Hafner, R. Charrondiere, and B. Bickel, “Directionality-aware design of embroidery patterns,” <i>Computer Graphics Forum</i>, vol. 42, no. 2. Wiley, pp. 397–409, 2023.","short":"Z. Liu, M. Piovarci, C. Hafner, R. Charrondiere, B. Bickel, Computer Graphics Forum 42 (2023) 397–409.","ama":"Liu Z, Piovarci M, Hafner C, Charrondiere R, Bickel B. Directionality-aware design of embroidery patterns. <i>Computer Graphics Forum</i>. 2023;42(2):397-409. doi:<a href=\"https://doi.org/10.1111/cgf.14770 \">10.1111/cgf.14770 </a>"},"issue":"2","language":[{"iso":"eng"}],"oa":1,"file":[{"file_id":"12974","date_created":"2023-05-16T08:28:37Z","file_size":24003702,"date_updated":"2023-05-16T08:28:37Z","creator":"mpiovarc","relation":"main_file","checksum":"4c188c2be4745467a8790bbf5d6491aa","success":1,"file_name":"Zhenyuan2023.pdf","access_level":"open_access","content_type":"application/pdf"}],"department":[{"_id":"BeBi"}],"month":"05","quality_controlled":"1","ddc":["004"],"page":"397-409","type":"journal_article","_id":"12972","date_updated":"2023-08-01T14:47:05Z","publisher":"Wiley","article_processing_charge":"No","doi":"10.1111/cgf.14770 ","conference":{"end_date":"2023-05-12","start_date":"2023-05-08","name":"EG: Eurographics","location":"Saarbrucken, Germany"},"acknowledgement":"This work was supported by the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation program (grant agreement No 715767 – MATERIALIZABLE), and FWF Lise Meitner (Grant M 3319). We thank the anonymous reviewers for their insightful feedback; Solal Pirelli, Shardul Chiplunkar, and Paola Mejia for proofreading; everyone in the visual computing group at ISTA for inspiring lunch and coffee breaks; Thibault Tricard for help producing the results of Phasor Noise.","date_published":"2023-05-08T00:00:00Z","ec_funded":1,"status":"public","publication":"Computer Graphics Forum","project":[{"grant_number":"M03319","name":"Perception-Aware Appearance Fabrication","_id":"eb901961-77a9-11ec-83b8-f5c883a62027"},{"grant_number":"715767","name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","call_identifier":"H2020","_id":"24F9549A-B435-11E9-9278-68D0E5697425"}],"keyword":["embroidery","design","directionality","density","image"],"external_id":{"isi":["001000062600033"]},"isi":1,"year":"2023"},{"date_created":"2023-05-16T09:34:13Z","title":"Gloss-aware color correction for 3D printing","oa_version":"Published Version","author":[{"last_name":"Condor","full_name":"Condor, Jorge","first_name":"Jorge"},{"first_name":"Michael","orcid":"0000-0002-5062-4474","last_name":"Piovarci","id":"62E473F4-5C99-11EA-A40E-AF823DDC885E","full_name":"Piovarci, Michael"},{"id":"49876194-F248-11E8-B48F-1D18A9856A87","full_name":"Bickel, Bernd","last_name":"Bickel","first_name":"Bernd","orcid":"0000-0001-6511-9385"},{"full_name":"Didyk, Piotr","last_name":"Didyk","first_name":"Piotr"}],"day":"23","publication_identifier":{"isbn":["9798400701597"]},"publication_status":"published","file_date_updated":"2024-01-29T10:14:10Z","tmp":{"name":"Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)","legal_code_url":"https://creativecommons.org/licenses/by/4.0/legalcode","image":"/images/cc_by.png","short":"CC BY (4.0)"},"abstract":[{"lang":"eng","text":"Color and gloss are fundamental aspects of surface appearance. State-of-the-art fabrication techniques can manipulate both properties of the printed 3D objects. However, in the context of appearance reproduction, perceptual aspects of color and gloss are usually handled separately, even though previous perceptual studies suggest their interaction. Our work is motivated by previous studies demonstrating a perceived color shift due to a change in the object's gloss, i.e., two samples with the same color but different surface gloss appear as they have different colors. In this paper, we conduct new experiments which support this observation and provide insights into the magnitude and direction of the perceived color change. We use the observations as guidance to design a new method that estimates and corrects the color shift enabling the fabrication of objects with the same perceived color but different surface gloss. We formulate the problem as an optimization procedure solved using differentiable rendering. We evaluate the effectiveness of our method in perceptual experiments with 3D objects fabricated using a multi-material 3D printer and demonstrate potential applications. "}],"has_accepted_license":"1","file":[{"file_id":"12983","date_updated":"2023-05-16T09:32:50Z","creator":"mpiovarc","date_created":"2023-05-16T09:32:50Z","file_size":42323971,"checksum":"84a437739af5d46507928939b20c0c28","relation":"main_file","access_level":"open_access","content_type":"application/pdf","success":1,"file_name":"Condor2023_supplemental.pdf"},{"file_id":"14893","creator":"dernst","date_updated":"2024-01-29T10:14:10Z","date_created":"2024-01-29T10:14:10Z","file_size":26079404,"checksum":"0f5c8b242e8e7c153c04888c4d0c6f37","relation":"main_file","access_level":"open_access","content_type":"application/pdf","success":1,"file_name":"2023_Siggraph_Condor.pdf"}],"article_number":"21","department":[{"_id":"BeBi"}],"month":"07","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","citation":{"mla":"Condor, Jorge, et al. “Gloss-Aware Color Correction for 3D Printing.” <i>SIGGRAPH ’23 Conference Proceedings</i>, 21, Association for Computing Machinery, 2023, doi:<a href=\"https://doi.org/10.1145/3588432.3591546\">10.1145/3588432.3591546</a>.","apa":"Condor, J., Piovarci, M., Bickel, B., &#38; Didyk, P. (2023). Gloss-aware color correction for 3D printing. In <i>SIGGRAPH ’23 Conference Proceedings</i>. Los Angeles, CA, United States: Association for Computing Machinery. <a href=\"https://doi.org/10.1145/3588432.3591546\">https://doi.org/10.1145/3588432.3591546</a>","ista":"Condor J, Piovarci M, Bickel B, Didyk P. 2023. Gloss-aware color correction for 3D printing. SIGGRAPH ’23 Conference Proceedings. SIGGRAPH: Computer Graphics and Interactive Techniques Conference, 21.","chicago":"Condor, Jorge, Michael Piovarci, Bernd Bickel, and Piotr Didyk. “Gloss-Aware Color Correction for 3D Printing.” In <i>SIGGRAPH ’23 Conference Proceedings</i>. Association for Computing Machinery, 2023. <a href=\"https://doi.org/10.1145/3588432.3591546\">https://doi.org/10.1145/3588432.3591546</a>.","short":"J. Condor, M. Piovarci, B. Bickel, P. Didyk, in:, SIGGRAPH ’23 Conference Proceedings, Association for Computing Machinery, 2023.","ieee":"J. Condor, M. Piovarci, B. Bickel, and P. Didyk, “Gloss-aware color correction for 3D printing,” in <i>SIGGRAPH ’23 Conference Proceedings</i>, Los Angeles, CA, United States, 2023.","ama":"Condor J, Piovarci M, Bickel B, Didyk P. Gloss-aware color correction for 3D printing. In: <i>SIGGRAPH ’23 Conference Proceedings</i>. Association for Computing Machinery; 2023. doi:<a href=\"https://doi.org/10.1145/3588432.3591546\">10.1145/3588432.3591546</a>"},"language":[{"iso":"eng"}],"oa":1,"type":"conference","date_updated":"2024-02-28T12:52:04Z","_id":"12979","publisher":"Association for Computing Machinery","doi":"10.1145/3588432.3591546","article_processing_charge":"No","quality_controlled":"1","ddc":["004"],"keyword":["color","gloss","perception","color compensation","color management"],"external_id":{"isi":["001117690500021"]},"year":"2023","isi":1,"acknowledgement":"We thank Matthew S Zurawski for the 3D model of the car speed shape. This research has been supported by the Swiss National Science Foundation (SNSF, Grant 200502) and the FWF Lise Meitner (Grant M 3319).","date_published":"2023-07-23T00:00:00Z","conference":{"name":"SIGGRAPH: Computer Graphics and Interactive Techniques Conference","end_date":"2023-08-10","start_date":"2023-08-06","location":"Los Angeles, CA, United States"},"project":[{"_id":"eb901961-77a9-11ec-83b8-f5c883a62027","name":"Perception-Aware Appearance Fabrication","grant_number":"M03319"}],"status":"public","publication":"SIGGRAPH ’23 Conference Proceedings"},{"department":[{"_id":"BeBi"}],"file":[{"checksum":"5f0a6867689e025a661bd0b4fd90b821","relation":"main_file","content_type":"application/pdf","access_level":"open_access","file_name":"Piovarci2023.pdf","success":1,"file_id":"12985","creator":"mpiovarc","date_updated":"2023-05-16T09:38:25Z","file_size":30817343,"date_created":"2023-05-16T09:38:25Z"}],"article_number":"67","month":"07","citation":{"apa":"Piovarci, M., Chapiro, A., &#38; Bickel, B. (2023). Skin-Screen: A computational fabrication framework for color tattoos. <i>Transactions on Graphics</i>. Los Angeles, CA, United States: Association for Computing Machinery. <a href=\"https://doi.org/10.1145/3592432\">https://doi.org/10.1145/3592432</a>","mla":"Piovarci, Michael, et al. “Skin-Screen: A Computational Fabrication Framework for Color Tattoos.” <i>Transactions on Graphics</i>, vol. 42, no. 4, 67, Association for Computing Machinery, 2023, doi:<a href=\"https://doi.org/10.1145/3592432\">10.1145/3592432</a>.","chicago":"Piovarci, Michael, Alexandre Chapiro, and Bernd Bickel. “Skin-Screen: A Computational Fabrication Framework for Color Tattoos.” <i>Transactions on Graphics</i>. Association for Computing Machinery, 2023. <a href=\"https://doi.org/10.1145/3592432\">https://doi.org/10.1145/3592432</a>.","ista":"Piovarci M, Chapiro A, Bickel B. 2023. Skin-Screen: A computational fabrication framework for color tattoos. Transactions on Graphics. 42(4), 67.","ieee":"M. Piovarci, A. Chapiro, and B. Bickel, “Skin-Screen: A computational fabrication framework for color tattoos,” <i>Transactions on Graphics</i>, vol. 42, no. 4. Association for Computing Machinery, 2023.","short":"M. Piovarci, A. Chapiro, B. Bickel, Transactions on Graphics 42 (2023).","ama":"Piovarci M, Chapiro A, Bickel B. Skin-Screen: A computational fabrication framework for color tattoos. <i>Transactions on Graphics</i>. 2023;42(4). doi:<a href=\"https://doi.org/10.1145/3592432\">10.1145/3592432</a>"},"issue":"4","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","oa":1,"language":[{"iso":"eng"}],"volume":42,"date_created":"2023-05-16T09:39:14Z","article_type":"original","day":"26","author":[{"first_name":"Michael","orcid":"0000-0002-5062-4474","full_name":"Piovarci, Michael","id":"62E473F4-5C99-11EA-A40E-AF823DDC885E","last_name":"Piovarci"},{"last_name":"Chapiro","full_name":"Chapiro, Alexandre","first_name":"Alexandre"},{"first_name":"Bernd","orcid":"0000-0001-6511-9385","full_name":"Bickel, Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87","last_name":"Bickel"}],"title":"Skin-Screen: A computational fabrication framework for color tattoos","oa_version":"Submitted Version","file_date_updated":"2023-05-16T09:38:25Z","publication_identifier":{"issn":["0730-0301"],"eissn":["1557-7368"]},"publication_status":"published","has_accepted_license":"1","acknowledged_ssus":[{"_id":"M-Shop"}],"intvolume":"        42","abstract":[{"lang":"eng","text":"Tattoos are a highly popular medium, with both artistic and medical applications. Although the mechanical process of tattoo application has evolved historically, the results are reliant on the artisanal skill of the artist. This can be especially challenging for some skin tones, or in cases where artists lack experience. We provide the first systematic overview of tattooing as a computational fabrication technique. We built an automated tattooing rig and a recipe for the creation of silicone sheets mimicking realistic skin tones, which allowed us to create an accurate model predicting tattoo appearance. This enables several exciting applications including tattoo previewing, color retargeting, novel ink spectra optimization, color-accurate prosthetics, and more."}],"keyword":["appearance","modeling","reproduction","tattoo","skin color","gamut mapping","ink-optimization","prosthetic"],"isi":1,"year":"2023","external_id":{"isi":["001044671300033"]},"conference":{"location":"Los Angeles, CA, United States","name":"SIGGRAPH: Computer Graphics and Interactive Techniques Conference","start_date":"2023-08-06","end_date":"2023-08-10"},"date_published":"2023-07-26T00:00:00Z","acknowledgement":"We thank Todor Asenov and the Miba Machine Shop for their help in assembling the tattoo machine and manufacturing the substrates. We thank Geysler Rodrigues for the insightful discussions on tattooing practices from a professional artist's perspective. We thank Maria Fernanda Portugal for sharing a doctor's perspective on medical applications of tattoos. This work is graciously supported by the FWF Lise Meitner (Grant M 3319).","status":"public","publication":"Transactions on Graphics","project":[{"grant_number":"M03319","name":"Perception-Aware Appearance Fabrication","_id":"eb901961-77a9-11ec-83b8-f5c883a62027"}],"_id":"12984","date_updated":"2024-01-29T10:27:23Z","type":"journal_article","article_processing_charge":"No","doi":"10.1145/3592432","publisher":"Association for Computing Machinery","quality_controlled":"1","ddc":["004"]},{"day":"26","author":[{"first_name":"Marco","full_name":"Freire, Marco","last_name":"Freire"},{"first_name":"Manas","orcid":"0009-0007-6138-6890","full_name":"Bhargava, Manas","id":"FF8FA64C-AA6A-11E9-99AD-50D4E5697425","last_name":"Bhargava"},{"id":"2B14B676-F248-11E8-B48F-1D18A9856A87","full_name":"Schreck, Camille","last_name":"Schreck","first_name":"Camille"},{"first_name":"Pierre-Alexandre","last_name":"Hugron","full_name":"Hugron, Pierre-Alexandre"},{"orcid":"0000-0001-6511-9385","first_name":"Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87","full_name":"Bickel, Bernd","last_name":"Bickel"},{"last_name":"Lefebvre","full_name":"Lefebvre, Sylvain","first_name":"Sylvain"}],"oa_version":"Submitted Version","title":"PCBend: Light up your 3D shapes with foldable circuit boards","volume":42,"date_created":"2023-05-22T08:37:04Z","article_type":"original","has_accepted_license":"1","acknowledged_ssus":[{"_id":"M-Shop"}],"abstract":[{"lang":"eng","text":"We propose a computational design approach for covering a surface with individually addressable RGB LEDs, effectively forming a low-resolution surface screen. To achieve a low-cost and scalable approach, we propose creating designs from flat PCB panels bent in-place along the surface of a 3D printed core. Working with standard rigid PCBs enables the use of\r\nestablished PCB manufacturing services, allowing the fabrication of designs with several hundred LEDs. \r\nOur approach optimizes the PCB geometry for folding, and then jointly optimizes the LED packing, circuit and routing, solving a challenging layout problem under strict manufacturing requirements. Unlike paper, PCBs cannot bend beyond a certain point without breaking. Therefore, we introduce parametric cut patterns acting as hinges, designed to allow bending while remaining compact. To tackle the joint optimization of placement, circuit and routing, we propose a specialized algorithm that splits the global problem into one sub-problem per triangle, which is then individually solved.\r\nOur technique generates PCB blueprints in a completely automated way. After being fabricated by a PCB manufacturing service, the boards are bent and glued by the user onto the 3D printed support. We demonstrate our technique on a range of physical models and virtual examples, creating intricate surface light patterns from hundreds of LEDs."}],"intvolume":"        42","file_date_updated":"2023-06-20T12:20:51Z","publication_status":"published","publication_identifier":{"issn":["0730-0301"],"eissn":["1557-7368"]},"month":"07","department":[{"_id":"GradSch"},{"_id":"BeBi"}],"article_number":"142","file":[{"relation":"main_file","checksum":"a0b0ba3b36f43a94388e8824613d812a","success":1,"file_name":"2023_ACMToG_Freire.pdf","access_level":"open_access","content_type":"application/pdf","file_id":"13156","file_size":78940724,"date_created":"2023-06-19T11:02:23Z","creator":"dernst","date_updated":"2023-06-19T11:02:23Z"},{"content_type":"application/pdf","access_level":"open_access","file_name":"2023_ACMToG_SuppMaterial_Freire.pdf","success":1,"checksum":"b9206bbb67af82df49b7e7cdbde3410c","relation":"main_file","creator":"dernst","date_updated":"2023-06-20T12:20:51Z","file_size":34345905,"date_created":"2023-06-20T12:20:51Z","file_id":"13157"}],"oa":1,"language":[{"iso":"eng"}],"citation":{"apa":"Freire, M., Bhargava, M., Schreck, C., Hugron, P.-A., Bickel, B., &#38; Lefebvre, S. (2023). PCBend: Light up your 3D shapes with foldable circuit boards. <i>Transactions on Graphics</i>. Los Angeles, CA, United States: Association for Computing Machinery. <a href=\"https://doi.org/10.1145/3592411\">https://doi.org/10.1145/3592411</a>","mla":"Freire, Marco, et al. “PCBend: Light up Your 3D Shapes with Foldable Circuit Boards.” <i>Transactions on Graphics</i>, vol. 42, no. 4, 142, Association for Computing Machinery, 2023, doi:<a href=\"https://doi.org/10.1145/3592411\">10.1145/3592411</a>.","ista":"Freire M, Bhargava M, Schreck C, Hugron P-A, Bickel B, Lefebvre S. 2023. PCBend: Light up your 3D shapes with foldable circuit boards. Transactions on Graphics. 42(4), 142.","chicago":"Freire, Marco, Manas Bhargava, Camille Schreck, Pierre-Alexandre Hugron, Bernd Bickel, and Sylvain Lefebvre. “PCBend: Light up Your 3D Shapes with Foldable Circuit Boards.” <i>Transactions on Graphics</i>. Association for Computing Machinery, 2023. <a href=\"https://doi.org/10.1145/3592411\">https://doi.org/10.1145/3592411</a>.","ieee":"M. Freire, M. Bhargava, C. Schreck, P.-A. Hugron, B. Bickel, and S. Lefebvre, “PCBend: Light up your 3D shapes with foldable circuit boards,” <i>Transactions on Graphics</i>, vol. 42, no. 4. Association for Computing Machinery, 2023.","short":"M. Freire, M. Bhargava, C. Schreck, P.-A. Hugron, B. Bickel, S. Lefebvre, Transactions on Graphics 42 (2023).","ama":"Freire M, Bhargava M, Schreck C, Hugron P-A, Bickel B, Lefebvre S. PCBend: Light up your 3D shapes with foldable circuit boards. <i>Transactions on Graphics</i>. 2023;42(4). doi:<a href=\"https://doi.org/10.1145/3592411\">10.1145/3592411</a>"},"issue":"4","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","article_processing_charge":"No","doi":"10.1145/3592411","publisher":"Association for Computing Machinery","_id":"13049","date_updated":"2024-01-29T10:30:49Z","type":"journal_article","ddc":["006"],"quality_controlled":"1","isi":1,"year":"2023","external_id":{"isi":["001044671300108"]},"keyword":["PCB design and layout","Mesh geometry models"],"publication":"Transactions on Graphics","status":"public","project":[{"_id":"24F9549A-B435-11E9-9278-68D0E5697425","grant_number":"715767","name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","call_identifier":"H2020"}],"ec_funded":1,"acknowledgement":"We thank the reviewers for the valuable feedback. We also thank the Miba Machine Shop at ISTA, PCBWay, and PragoBoard for helping us with fabrication and assembly. This project was supported by the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation program (Grant Agreement No. 715767 – MATERIALIZABLE).","date_published":"2023-07-26T00:00:00Z","conference":{"name":"SIGGRAPH: Computer Graphics and Interactive Techniques Conference","start_date":"2023-08-06","end_date":"2023-08-10","location":"Los Angeles, CA, United States"}},{"abstract":[{"lang":"eng","text":"We study structural rigidity for assemblies with mechanical joints. Existing methods identify whether an assembly is structurally rigid by assuming parts are perfectly rigid. Yet, an assembly identified as rigid may not be that “rigid” in practice, and existing methods cannot quantify how rigid an assembly is. We address this limitation by developing a new measure, worst-case rigidity, to quantify the rigidity of an assembly as the largest possible deformation that the assembly undergoes for arbitrary external loads of fixed magnitude. Computing worst-case rigidity is non-trivial due to non-rigid parts and different joint types. We thus formulate a new computational approach by encoding parts and their connections into a stiffness matrix, in which parts are modeled as deformable objects and joints as soft constraints. Based on this, we formulate worst-case rigidity analysis as an optimization that seeks the worst-case deformation of an assembly for arbitrary external loads, and solve the optimization problem via an eigenanalysis. Furthermore, we present methods to optimize the geometry and topology of various assemblies to enhance their rigidity, as guided by our rigidity measure. In the end, we validate our method on a variety of assembly structures with physical experiments and demonstrate its effectiveness by designing and fabricating several structurally rigid assemblies."}],"intvolume":"        41","acknowledged_ssus":[{"_id":"M-Shop"}],"has_accepted_license":"1","publication_status":"published","publication_identifier":{"eissn":["1467-8659"],"issn":["0167-7055"]},"file_date_updated":"2022-03-27T17:34:11Z","title":"Worst-case rigidity analysis and optimization for assemblies with mechanical joints","oa_version":"Submitted Version","author":[{"last_name":"Liu","full_name":"Liu, Zhenyuan","id":"70f0d7cf-ae65-11ec-a14f-89dfc5505b19","first_name":"Zhenyuan","orcid":"0000-0001-9200-5690"},{"first_name":"Jingyu","last_name":"Hu","full_name":"Hu, Jingyu"},{"first_name":"Hao","last_name":"Xu","full_name":"Xu, Hao"},{"first_name":"Peng","last_name":"Song","full_name":"Song, Peng"},{"last_name":"Zhang","full_name":"Zhang, Ran","first_name":"Ran"},{"id":"49876194-F248-11E8-B48F-1D18A9856A87","full_name":"Bickel, Bernd","last_name":"Bickel","orcid":"0000-0001-6511-9385","first_name":"Bernd"},{"full_name":"Fu, Chi-Wing","last_name":"Fu","first_name":"Chi-Wing"}],"day":"01","scopus_import":"1","article_type":"original","date_created":"2022-03-27T17:34:17Z","volume":41,"language":[{"iso":"eng"}],"oa":1,"user_id":"4359f0d1-fa6c-11eb-b949-802e58b17ae8","issue":"2","citation":{"short":"Z. Liu, J. Hu, H. Xu, P. Song, R. Zhang, B. Bickel, C.-W. Fu, Computer Graphics Forum 41 (2022) 507–519.","ieee":"Z. Liu <i>et al.</i>, “Worst-case rigidity analysis and optimization for assemblies with mechanical joints,” <i>Computer Graphics Forum</i>, vol. 41, no. 2. Wiley, pp. 507–519, 2022.","ama":"Liu Z, Hu J, Xu H, et al. Worst-case rigidity analysis and optimization for assemblies with mechanical joints. <i>Computer Graphics Forum</i>. 2022;41(2):507-519. doi:<a href=\"https://doi.org/10.1111/cgf.14490\">10.1111/cgf.14490</a>","mla":"Liu, Zhenyuan, et al. “Worst-Case Rigidity Analysis and Optimization for Assemblies with Mechanical Joints.” <i>Computer Graphics Forum</i>, vol. 41, no. 2, Wiley, 2022, pp. 507–19, doi:<a href=\"https://doi.org/10.1111/cgf.14490\">10.1111/cgf.14490</a>.","apa":"Liu, Z., Hu, J., Xu, H., Song, P., Zhang, R., Bickel, B., &#38; Fu, C.-W. (2022). Worst-case rigidity analysis and optimization for assemblies with mechanical joints. <i>Computer Graphics Forum</i>. Wiley. <a href=\"https://doi.org/10.1111/cgf.14490\">https://doi.org/10.1111/cgf.14490</a>","chicago":"Liu, Zhenyuan, Jingyu Hu, Hao Xu, Peng Song, Ran Zhang, Bernd Bickel, and Chi-Wing Fu. “Worst-Case Rigidity Analysis and Optimization for Assemblies with Mechanical Joints.” <i>Computer Graphics Forum</i>. Wiley, 2022. <a href=\"https://doi.org/10.1111/cgf.14490\">https://doi.org/10.1111/cgf.14490</a>.","ista":"Liu Z, Hu J, Xu H, Song P, Zhang R, Bickel B, Fu C-W. 2022. Worst-case rigidity analysis and optimization for assemblies with mechanical joints. Computer Graphics Forum. 41(2), 507–519."},"month":"05","file":[{"date_created":"2022-03-27T17:34:11Z","file_size":19601689,"date_updated":"2022-03-27T17:34:11Z","creator":"bbickel","file_id":"10923","file_name":"paper.pdf","content_type":"application/pdf","access_level":"open_access","relation":"main_file","checksum":"b62188b07f5c000f1638c782ec92da41"}],"department":[{"_id":"BeBi"}],"ddc":["000"],"page":"507-519","quality_controlled":"1","publisher":"Wiley","doi":"10.1111/cgf.14490","article_processing_charge":"No","type":"journal_article","date_updated":"2023-08-03T06:17:13Z","_id":"10922","project":[{"_id":"24F9549A-B435-11E9-9278-68D0E5697425","name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","grant_number":"715767","call_identifier":"H2020"}],"status":"public","publication":"Computer Graphics Forum","date_published":"2022-05-01T00:00:00Z","acknowledgement":"This work was supported by the Research Grants Council of the Hong Kong Special Administrative Region, China [Project No.: CUHK 14201921] and the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement No 715767 – MATERIALIZABLE). We thank the anonymous reviewers for their insightful feedback; Christian Hafner for proofreading and discussions; Ziqi Wang,\r\nHaisen Zhao, and Martin Hafskjold Thoresen for the helpful discussions; and the Miba Machine Shop at IST Austria for 3D printing the BUNNY and BOOMERANG models.","ec_funded":1,"external_id":{"isi":["000802723900039"]},"year":"2022","isi":1},{"publication":"ACM Transactions on Graphics","status":"public","project":[{"grant_number":"M03319","name":"Perception-Aware Appearance Fabrication","_id":"eb901961-77a9-11ec-83b8-f5c883a62027"},{"_id":"24F9549A-B435-11E9-9278-68D0E5697425","name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","grant_number":"715767","call_identifier":"H2020"}],"date_published":"2022-06-01T00:00:00Z","acknowledgement":"This work is graciously supported by the following grant agencies: FWF Lise Meitner (Grant M 3319), SNSF (Grant 200502), ERC Starting Grant (MATERIALIZABLE-715767), NSF (Grant IIS-181507).\r\n","ec_funded":1,"related_material":{"link":[{"url":"https://ista.ac.at/en/news/machine-learning-3d-printing-fluids/","description":"News on ISTA website","relation":"press_release"}]},"external_id":{"arxiv":["2201.11819"]},"year":"2022","ddc":["000"],"quality_controlled":"1","publisher":"Association for Computing Machinery","article_processing_charge":"No","doi":"10.1145/3528223.3530144","type":"journal_article","_id":"11442","date_updated":"2023-05-31T12:38:21Z","language":[{"iso":"eng"}],"oa":1,"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","citation":{"ieee":"M. Piovarci <i>et al.</i>, “Closed-loop control of direct ink writing via reinforcement learning,” <i>ACM Transactions on Graphics</i>, vol. 41, no. 4. Association for Computing Machinery, 2022.","short":"M. Piovarci, M. Foshey, J. Xu, T. Erps, V. Babaei, P. Didyk, S. Rusinkiewicz, W. Matusik, B. Bickel, ACM Transactions on Graphics 41 (2022).","ama":"Piovarci M, Foshey M, Xu J, et al. Closed-loop control of direct ink writing via reinforcement learning. <i>ACM Transactions on Graphics</i>. 2022;41(4). doi:<a href=\"https://doi.org/10.1145/3528223.3530144\">10.1145/3528223.3530144</a>","apa":"Piovarci, M., Foshey, M., Xu, J., Erps, T., Babaei, V., Didyk, P., … Bickel, B. (2022). Closed-loop control of direct ink writing via reinforcement learning. <i>ACM Transactions on Graphics</i>. Association for Computing Machinery. <a href=\"https://doi.org/10.1145/3528223.3530144\">https://doi.org/10.1145/3528223.3530144</a>","mla":"Piovarci, Michael, et al. “Closed-Loop Control of Direct Ink Writing via Reinforcement Learning.” <i>ACM Transactions on Graphics</i>, vol. 41, no. 4, 112, Association for Computing Machinery, 2022, doi:<a href=\"https://doi.org/10.1145/3528223.3530144\">10.1145/3528223.3530144</a>.","chicago":"Piovarci, Michael, Michael Foshey, Jie Xu, Timothy Erps, Vahid Babaei, Piotr Didyk, Szymon Rusinkiewicz, Wojciech Matusik, and Bernd Bickel. “Closed-Loop Control of Direct Ink Writing via Reinforcement Learning.” <i>ACM Transactions on Graphics</i>. Association for Computing Machinery, 2022. <a href=\"https://doi.org/10.1145/3528223.3530144\">https://doi.org/10.1145/3528223.3530144</a>.","ista":"Piovarci M, Foshey M, Xu J, Erps T, Babaei V, Didyk P, Rusinkiewicz S, Matusik W, Bickel B. 2022. Closed-loop control of direct ink writing via reinforcement learning. ACM Transactions on Graphics. 41(4), 112."},"issue":"4","month":"06","arxiv":1,"article_number":"112","file":[{"relation":"main_file","checksum":"27f6fe41c6ff84d50445cc9b0176d45b","file_name":"2022_ACM_acceptedversion_Piovarci.pdf","success":1,"content_type":"application/pdf","access_level":"open_access","file_id":"11467","file_size":33994829,"date_created":"2022-06-28T08:32:58Z","creator":"dernst","date_updated":"2022-06-28T08:32:58Z"}],"department":[{"_id":"BeBi"}],"tmp":{"name":"Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)","legal_code_url":"https://creativecommons.org/licenses/by/4.0/legalcode","image":"/images/cc_by.png","short":"CC BY (4.0)"},"abstract":[{"text":"Enabling additive manufacturing to employ a wide range of novel, functional materials can be a major boost to this technology. However, making such materials printable requires painstaking trial-and-error by an expert operator,\r\nas they typically tend to exhibit peculiar rheological or hysteresis properties. Even in the case of successfully finding the process parameters, there is no guarantee of print-to-print consistency due to material differences between batches. These challenges make closed-loop feedback an attractive option where the process parameters are adjusted on-the-fly. There are several challenges for designing an efficient controller: the deposition parameters are complex and highly coupled, artifacts occur after long time horizons, simulating the deposition is computationally costly, and learning on hardware is intractable. In this work, we demonstrate the feasibility of learning a closed-loop control policy for additive manufacturing using reinforcement learning. We show that approximate, but efficient, numerical simulation is\r\nsufficient as long as it allows learning the behavioral patterns of deposition that translate to real-world experiences. In combination with reinforcement learning, our model can be used to discover control policies that outperform\r\nbaseline controllers. Furthermore, the recovered policies have a minimal sim-to-real gap. We showcase this by applying our control policy in-vivo on a single-layer, direct ink writing printer. ","lang":"eng"}],"intvolume":"        41","has_accepted_license":"1","file_date_updated":"2022-06-28T08:32:58Z","publication_identifier":{"eissn":["1557-7368"],"issn":["0730-0301"]},"publication_status":"published","oa_version":"Submitted Version","title":"Closed-loop control of direct ink writing via reinforcement learning","day":"01","author":[{"first_name":"Michael","id":"62E473F4-5C99-11EA-A40E-AF823DDC885E","full_name":"Piovarci, Michael","last_name":"Piovarci"},{"last_name":"Foshey","full_name":"Foshey, Michael","first_name":"Michael"},{"first_name":"Jie","full_name":"Xu, Jie","last_name":"Xu"},{"last_name":"Erps","full_name":"Erps, Timothy","first_name":"Timothy"},{"last_name":"Babaei","full_name":"Babaei, Vahid","first_name":"Vahid"},{"first_name":"Piotr","last_name":"Didyk","full_name":"Didyk, Piotr"},{"full_name":"Rusinkiewicz, Szymon","last_name":"Rusinkiewicz","first_name":"Szymon"},{"first_name":"Wojciech","full_name":"Matusik, Wojciech","last_name":"Matusik"},{"id":"49876194-F248-11E8-B48F-1D18A9856A87","full_name":"Bickel, Bernd","last_name":"Bickel","first_name":"Bernd","orcid":"0000-0001-6511-9385"}],"date_created":"2022-06-10T06:41:47Z","article_type":"original","volume":41},{"type":"journal_article","_id":"11735","date_updated":"2023-08-03T13:21:22Z","publisher":"Association for Computing Machinery","article_processing_charge":"No","doi":"10.1145/3528223.3530071","quality_controlled":"1","ddc":["000"],"related_material":{"link":[{"description":"News on ISTA website","url":"https://ista.ac.at/en/news/unlocking-interlocking-riddles/","relation":"press_release"}]},"external_id":{"isi":["000830989200018"]},"year":"2022","isi":1,"date_published":"2022-07-22T00:00:00Z","acknowledgement":"We thank the reviewers for the valuable comments, David Gontier for sharing the source code of the baseline design approach, Christian Hafner for proofreading the paper, Keenan Crane for the 3D model of Cow, and Thingiverse for the 3D models of Moai and Owl. This work was supported by the SUTD Start-up Research Grant (Number: SRG ISTD 2019 148), the Swiss National Science Foundation (NCCR Digital Fabrication Agreement #51NF40-141853), and\r\nthe European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (Grant Agreement No 715767 – MATERIALIZABLE).","ec_funded":1,"status":"public","publication":"ACM Transactions on Graphics","project":[{"grant_number":"715767","name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","call_identifier":"H2020","_id":"24F9549A-B435-11E9-9278-68D0E5697425"}],"date_created":"2022-08-07T22:01:57Z","article_type":"original","volume":41,"title":"Computational design of high-level interlocking puzzles","oa_version":"Submitted Version","scopus_import":"1","day":"22","author":[{"last_name":"Chen","full_name":"Chen, Rulin","first_name":"Rulin"},{"first_name":"Ziqi","full_name":"Wang, Ziqi","last_name":"Wang"},{"first_name":"Peng","full_name":"Song, Peng","last_name":"Song"},{"last_name":"Bickel","full_name":"Bickel, Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87","first_name":"Bernd","orcid":"0000-0001-6511-9385"}],"file_date_updated":"2022-08-28T07:56:19Z","publication_status":"published","publication_identifier":{"eissn":["1557-7368"],"issn":["0730-0301"]},"abstract":[{"text":"Interlocking puzzles are intriguing geometric games where the puzzle pieces are held together based on their geometric arrangement, preventing the puzzle from falling apart. High-level-of-difficulty, or simply high-level, interlocking puzzles are a subclass of interlocking puzzles that require multiple moves to take out the first subassembly from the puzzle. Solving a high-level interlocking puzzle is a challenging task since one has to explore many different configurations of the puzzle pieces until reaching a configuration where the first subassembly can be taken out. Designing a high-level interlocking puzzle with a user-specified level of difficulty is even harder since the puzzle pieces have to be interlocking in all the configurations before the first subassembly is taken out.\r\n\r\nIn this paper, we present a computational approach to design high-level interlocking puzzles. The core idea is to represent all possible configurations of an interlocking puzzle as well as transitions among these configurations using a rooted, undirected graph called a disassembly graph and leverage this graph to find a disassembly plan that requires a minimal number of moves to take out the first subassembly from the puzzle. At the design stage, our algorithm iteratively constructs the geometry of each puzzle piece to expand the disassembly graph incrementally, aiming to achieve a user-specified level of difficulty. We show that our approach allows efficient generation of high-level interlocking puzzles of various shape complexities, including new solutions not attainable by state-of-the-art approaches.","lang":"eng"}],"intvolume":"        41","has_accepted_license":"1","article_number":"150","file":[{"access_level":"open_access","content_type":"application/pdf","success":1,"file_name":"Chen-2022-High-LevelPuzzle_authorVersion.pdf","checksum":"0b51651be45b1b33f2072bd5d2686c69","relation":"main_file","date_updated":"2022-08-28T07:56:19Z","creator":"bbickel","file_size":16896871,"date_created":"2022-08-28T07:56:19Z","file_id":"11992"}],"department":[{"_id":"BeBi"}],"month":"07","user_id":"4359f0d1-fa6c-11eb-b949-802e58b17ae8","citation":{"chicago":"Chen, Rulin, Ziqi Wang, Peng Song, and Bernd Bickel. “Computational Design of High-Level Interlocking Puzzles.” <i>ACM Transactions on Graphics</i>. Association for Computing Machinery, 2022. <a href=\"https://doi.org/10.1145/3528223.3530071\">https://doi.org/10.1145/3528223.3530071</a>.","ista":"Chen R, Wang Z, Song P, Bickel B. 2022. Computational design of high-level interlocking puzzles. ACM Transactions on Graphics. 41(4), 150.","apa":"Chen, R., Wang, Z., Song, P., &#38; Bickel, B. (2022). Computational design of high-level interlocking puzzles. <i>ACM Transactions on Graphics</i>. Association for Computing Machinery. <a href=\"https://doi.org/10.1145/3528223.3530071\">https://doi.org/10.1145/3528223.3530071</a>","mla":"Chen, Rulin, et al. “Computational Design of High-Level Interlocking Puzzles.” <i>ACM Transactions on Graphics</i>, vol. 41, no. 4, 150, Association for Computing Machinery, 2022, doi:<a href=\"https://doi.org/10.1145/3528223.3530071\">10.1145/3528223.3530071</a>.","ama":"Chen R, Wang Z, Song P, Bickel B. Computational design of high-level interlocking puzzles. <i>ACM Transactions on Graphics</i>. 2022;41(4). doi:<a href=\"https://doi.org/10.1145/3528223.3530071\">10.1145/3528223.3530071</a>","ieee":"R. Chen, Z. Wang, P. Song, and B. Bickel, “Computational design of high-level interlocking puzzles,” <i>ACM Transactions on Graphics</i>, vol. 41, no. 4. Association for Computing Machinery, 2022.","short":"R. Chen, Z. Wang, P. Song, B. Bickel, ACM Transactions on Graphics 41 (2022)."},"issue":"4","language":[{"iso":"eng"}],"oa":1},{"main_file_link":[{"open_access":"1","url":"https://doi.org/10.1101/2022.03.16.484431"}],"publication_status":"submitted","abstract":[{"text":"Complex wiring between neurons underlies the information-processing network enabling all brain functions, including cognition and memory. For understanding how the network is structured, processes information, and changes over time, comprehensive visualization of the architecture of living brain tissue with its cellular and molecular components would open up major opportunities. However, electron microscopy (EM) provides nanometre-scale resolution required for full <jats:italic>in-silico</jats:italic> reconstruction<jats:sup>1–5</jats:sup>, yet is limited to fixed specimens and static representations. Light microscopy allows live observation, with super-resolution approaches<jats:sup>6–12</jats:sup> facilitating nanoscale visualization, but comprehensive 3D-reconstruction of living brain tissue has been hindered by tissue photo-burden, photobleaching, insufficient 3D-resolution, and inadequate signal-to-noise ratio (SNR). Here we demonstrate saturated reconstruction of living brain tissue. We developed an integrated imaging and analysis technology, adapting stimulated emission depletion (STED) microscopy<jats:sup>6,13</jats:sup> in extracellularly labelled tissue<jats:sup>14</jats:sup> for high SNR and near-isotropic resolution. Centrally, a two-stage deep-learning approach leveraged previously obtained information on sample structure to drastically reduce photo-burden and enable automated volumetric reconstruction down to single synapse level. Live reconstruction provides unbiased analysis of tissue architecture across time in relation to functional activity and targeted activation, and contextual understanding of molecular labelling. This adoptable technology will facilitate novel insights into the dynamic functional architecture of living brain tissue.","lang":"eng"}],"date_updated":"2024-03-25T23:30:11Z","_id":"11943","type":"preprint","date_created":"2022-08-23T11:07:59Z","author":[{"last_name":"Velicky","id":"39BDC62C-F248-11E8-B48F-1D18A9856A87","full_name":"Velicky, Philipp","first_name":"Philipp","orcid":"0000-0002-2340-7431"},{"orcid":"0000-0001-5665-0430","first_name":"Eder","last_name":"Miguel Villalba","id":"3FB91342-F248-11E8-B48F-1D18A9856A87","full_name":"Miguel Villalba, Eder"},{"full_name":"Michalska, Julia M","id":"443DB6DE-F248-11E8-B48F-1D18A9856A87","last_name":"Michalska","orcid":"0000-0003-3862-1235","first_name":"Julia M"},{"last_name":"Wei","full_name":"Wei, Donglai","first_name":"Donglai"},{"last_name":"Lin","full_name":"Lin, Zudi","first_name":"Zudi"},{"full_name":"Watson, Jake","id":"63836096-4690-11EA-BD4E-32803DDC885E","last_name":"Watson","first_name":"Jake","orcid":"0000-0002-8698-3823"},{"last_name":"Troidl","full_name":"Troidl, Jakob","first_name":"Jakob"},{"first_name":"Johanna","full_name":"Beyer, Johanna","last_name":"Beyer"},{"id":"43DF3136-F248-11E8-B48F-1D18A9856A87","full_name":"Ben Simon, Yoav","last_name":"Ben Simon","first_name":"Yoav"},{"orcid":"0000-0003-1216-9105","first_name":"Christoph M","last_name":"Sommer","full_name":"Sommer, Christoph M","id":"4DF26D8C-F248-11E8-B48F-1D18A9856A87"},{"first_name":"Wiebke","full_name":"Jahr, Wiebke","id":"425C1CE8-F248-11E8-B48F-1D18A9856A87","last_name":"Jahr"},{"last_name":"Cenameri","full_name":"Cenameri, Alban","id":"9ac8f577-2357-11eb-997a-e566c5550886","first_name":"Alban"},{"full_name":"Broichhagen, Johannes","last_name":"Broichhagen","first_name":"Johannes"},{"last_name":"Grant","full_name":"Grant, Seth G. N.","first_name":"Seth G. N."},{"orcid":"0000-0001-5001-4804","first_name":"Peter M","id":"353C1B58-F248-11E8-B48F-1D18A9856A87","full_name":"Jonas, Peter M","last_name":"Jonas"},{"id":"3E57A680-F248-11E8-B48F-1D18A9856A87","full_name":"Novarino, Gaia","last_name":"Novarino","orcid":"0000-0002-7673-7178","first_name":"Gaia"},{"last_name":"Pfister","full_name":"Pfister, Hanspeter","first_name":"Hanspeter"},{"first_name":"Bernd","orcid":"0000-0001-6511-9385","last_name":"Bickel","full_name":"Bickel, Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87"},{"orcid":"0000-0001-8559-3973","first_name":"Johann G","full_name":"Danzl, Johann G","id":"42EFD3B6-F248-11E8-B48F-1D18A9856A87","last_name":"Danzl"}],"doi":"10.1101/2022.03.16.484431","article_processing_charge":"No","day":"09","oa_version":"Preprint","publisher":"Cold Spring Harbor Laboratory","title":"Saturated reconstruction of living brain tissue","citation":{"short":"P. Velicky, E. Miguel Villalba, J.M. Michalska, D. Wei, Z. Lin, J. Watson, J. Troidl, J. Beyer, Y. Ben Simon, C.M. Sommer, W. Jahr, A. Cenameri, J. Broichhagen, S.G.N. Grant, P.M. Jonas, G. Novarino, H. Pfister, B. Bickel, J.G. Danzl, BioRxiv (n.d.).","ieee":"P. Velicky <i>et al.</i>, “Saturated reconstruction of living brain tissue,” <i>bioRxiv</i>. Cold Spring Harbor Laboratory.","ama":"Velicky P, Miguel Villalba E, Michalska JM, et al. Saturated reconstruction of living brain tissue. <i>bioRxiv</i>. doi:<a href=\"https://doi.org/10.1101/2022.03.16.484431\">10.1101/2022.03.16.484431</a>","mla":"Velicky, Philipp, et al. “Saturated Reconstruction of Living Brain Tissue.” <i>BioRxiv</i>, Cold Spring Harbor Laboratory, doi:<a href=\"https://doi.org/10.1101/2022.03.16.484431\">10.1101/2022.03.16.484431</a>.","apa":"Velicky, P., Miguel Villalba, E., Michalska, J. M., Wei, D., Lin, Z., Watson, J., … Danzl, J. G. (n.d.). Saturated reconstruction of living brain tissue. <i>bioRxiv</i>. Cold Spring Harbor Laboratory. <a href=\"https://doi.org/10.1101/2022.03.16.484431\">https://doi.org/10.1101/2022.03.16.484431</a>","chicago":"Velicky, Philipp, Eder Miguel Villalba, Julia M Michalska, Donglai Wei, Zudi Lin, Jake Watson, Jakob Troidl, et al. “Saturated Reconstruction of Living Brain Tissue.” <i>BioRxiv</i>. Cold Spring Harbor Laboratory, n.d. <a href=\"https://doi.org/10.1101/2022.03.16.484431\">https://doi.org/10.1101/2022.03.16.484431</a>.","ista":"Velicky P, Miguel Villalba E, Michalska JM, Wei D, Lin Z, Watson J, Troidl J, Beyer J, Ben Simon Y, Sommer CM, Jahr W, Cenameri A, Broichhagen J, Grant SGN, Jonas PM, Novarino G, Pfister H, Bickel B, Danzl JG. Saturated reconstruction of living brain tissue. bioRxiv, <a href=\"https://doi.org/10.1101/2022.03.16.484431\">10.1101/2022.03.16.484431</a>."},"date_published":"2022-05-09T00:00:00Z","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","oa":1,"status":"public","publication":"bioRxiv","language":[{"iso":"eng"}],"department":[{"_id":"PeJo"},{"_id":"GaNo"},{"_id":"BeBi"},{"_id":"JoDa"}],"year":"2022","month":"05","related_material":{"record":[{"relation":"dissertation_contains","status":"public","id":"12470"}]}},{"type":"journal_article","date_updated":"2023-08-03T13:21:55Z","_id":"11993","publisher":"Wiley","doi":"10.1111/cgf.14581","article_processing_charge":"No","quality_controlled":"1","ddc":["000"],"page":"435-452","keyword":["Computer Graphics and Computer-Aided Design"],"external_id":{"isi":["000842638900001"]},"isi":1,"year":"2022","date_published":"2022-09-01T00:00:00Z","publication":"Computer Graphics Forum","status":"public","article_type":"original","date_created":"2022-08-28T18:17:01Z","volume":41,"title":"State of the art in computational mould design","oa_version":"Submitted Version","author":[{"first_name":"Thomas","last_name":"Alderighi","full_name":"Alderighi, Thomas"},{"full_name":"Malomo, Luigi","last_name":"Malomo","first_name":"Luigi"},{"first_name":"Thomas","orcid":"0000-0002-1546-3265","last_name":"Auzinger","full_name":"Auzinger, Thomas","id":"4718F954-F248-11E8-B48F-1D18A9856A87"},{"full_name":"Bickel, Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87","last_name":"Bickel","orcid":"0000-0001-6511-9385","first_name":"Bernd"},{"last_name":"Cignoni","full_name":"Cignoni, Paulo","first_name":"Paulo"},{"last_name":"Pietroni","full_name":"Pietroni, Nico","first_name":"Nico"}],"scopus_import":"1","day":"01","publication_status":"published","publication_identifier":{"issn":["0167-7055"],"eissn":["1467-8659"]},"file_date_updated":"2022-08-28T18:18:08Z","intvolume":"        41","abstract":[{"text":"Moulding refers to a set of manufacturing techniques in which a mould, usually a cavity or a solid frame, is used to shape a liquid or pliable material into an object of the desired shape. The popularity of moulding comes from its effectiveness, scalability and versatility in terms of employed materials. Its relevance as a fabrication process is demonstrated by the extensive literature covering different aspects related to mould design, from material flow simulation to the automation of mould geometry design. In this state-of-the-art report, we provide an extensive review of the automatic methods for the design of moulds, focusing on contributions from a geometric perspective. We classify existing mould design methods based on their computational approach and the nature of their target moulding process. We summarize the relationships between computational approaches and moulding techniques, highlighting their strengths and limitations. Finally, we discuss potential future research directions.","lang":"eng"}],"has_accepted_license":"1","file":[{"description":"This is the pre-peer reviewed version of the following article: Alderighi, T., Malomo, L., Auzinger, T., Bickel, B., Cignoni, P. and Pietroni, N. (2022), State of the Art in Computational Mould Design. Computer Graphics Forum, which has been published in final form at https://doi.org/10.1111/cgf.14581. This article may be used for non-commercial purposes in accordance with Wiley Terms and Conditions for Use of Self-Archived Versions.","checksum":"c40cc8ceb7b7f0512172b883d712198e","relation":"main_file","content_type":"application/pdf","access_level":"open_access","file_name":"star_molding_preprint.pdf","title":"pre-peer reviewed version","file_id":"11994","creator":"bbickel","date_updated":"2022-08-28T18:18:08Z","file_size":32480850,"date_created":"2022-08-28T18:18:08Z"}],"department":[{"_id":"BeBi"}],"month":"09","user_id":"4359f0d1-fa6c-11eb-b949-802e58b17ae8","issue":"6","citation":{"mla":"Alderighi, Thomas, et al. “State of the Art in Computational Mould Design.” <i>Computer Graphics Forum</i>, vol. 41, no. 6, Wiley, 2022, pp. 435–52, doi:<a href=\"https://doi.org/10.1111/cgf.14581\">10.1111/cgf.14581</a>.","apa":"Alderighi, T., Malomo, L., Auzinger, T., Bickel, B., Cignoni, P., &#38; Pietroni, N. (2022). State of the art in computational mould design. <i>Computer Graphics Forum</i>. Wiley. <a href=\"https://doi.org/10.1111/cgf.14581\">https://doi.org/10.1111/cgf.14581</a>","ista":"Alderighi T, Malomo L, Auzinger T, Bickel B, Cignoni P, Pietroni N. 2022. State of the art in computational mould design. Computer Graphics Forum. 41(6), 435–452.","chicago":"Alderighi, Thomas, Luigi Malomo, Thomas Auzinger, Bernd Bickel, Paulo Cignoni, and Nico Pietroni. “State of the Art in Computational Mould Design.” <i>Computer Graphics Forum</i>. Wiley, 2022. <a href=\"https://doi.org/10.1111/cgf.14581\">https://doi.org/10.1111/cgf.14581</a>.","short":"T. Alderighi, L. Malomo, T. Auzinger, B. Bickel, P. Cignoni, N. Pietroni, Computer Graphics Forum 41 (2022) 435–452.","ieee":"T. Alderighi, L. Malomo, T. Auzinger, B. Bickel, P. Cignoni, and N. Pietroni, “State of the art in computational mould design,” <i>Computer Graphics Forum</i>, vol. 41, no. 6. Wiley, pp. 435–452, 2022.","ama":"Alderighi T, Malomo L, Auzinger T, Bickel B, Cignoni P, Pietroni N. State of the art in computational mould design. <i>Computer Graphics Forum</i>. 2022;41(6):435-452. doi:<a href=\"https://doi.org/10.1111/cgf.14581\">10.1111/cgf.14581</a>"},"language":[{"iso":"eng"}],"oa":1},{"citation":{"ieee":"P. Rao <i>et al.</i>, “VoRF: Volumetric Relightable Faces,” in <i>33rd British Machine Vision Conference</i>, London, United Kingdom, 2022.","short":"P. Rao, M. B R, G. Fox, T. Weyrich, B. Bickel, H.-P. Seidel, H. Pfister, W. Matusik, A. Tewari, C. Theobalt, M. Elgharib, in:, 33rd British Machine Vision Conference, British Machine Vision Association and Society for Pattern Recognition, 2022.","ama":"Rao P, B R M, Fox G, et al. VoRF: Volumetric Relightable Faces. In: <i>33rd British Machine Vision Conference</i>. British Machine Vision Association and Society for Pattern Recognition; 2022.","apa":"Rao, P., B R, M., Fox, G., Weyrich, T., Bickel, B., Seidel, H.-P., … Elgharib, M. (2022). VoRF: Volumetric Relightable Faces. In <i>33rd British Machine Vision Conference</i>. London, United Kingdom: British Machine Vision Association and Society for Pattern Recognition.","mla":"Rao, Pramod, et al. “VoRF: Volumetric Relightable Faces.” <i>33rd British Machine Vision Conference</i>, 708, British Machine Vision Association and Society for Pattern Recognition, 2022.","ista":"Rao P, B R M, Fox G, Weyrich T, Bickel B, Seidel H-P, Pfister H, Matusik W, Tewari A, Theobalt C, Elgharib M. 2022. VoRF: Volumetric Relightable Faces. 33rd British Machine Vision Conference. BMVC: British Machine Vision Conference, 708.","chicago":"Rao, Pramod, Mallikarjun B R, Gereon Fox, Tim Weyrich, Bernd Bickel, Hans-Peter Seidel, Hanspeter Pfister, et al. “VoRF: Volumetric Relightable Faces.” In <i>33rd British Machine Vision Conference</i>. British Machine Vision Association and Society for Pattern Recognition, 2022."},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","oa":1,"language":[{"iso":"eng"}],"department":[{"_id":"BeBi"}],"article_number":"708","file":[{"file_id":"12453","title":"VoRF: Volumetric Relightable Faces","file_size":5202710,"date_created":"2023-01-30T10:48:18Z","date_updated":"2023-01-30T10:48:18Z","creator":"bbickel","relation":"main_file","checksum":"b60b70bb48700aee709c85a69231821d","file_name":"vorf_main.pdf","access_level":"open_access","content_type":"application/pdf"},{"file_size":37953188,"date_created":"2023-01-30T10:48:29Z","creator":"bbickel","date_updated":"2023-01-30T10:48:29Z","title":"VoRF: Volumetric Relightable Faces – SUPPLEMENTAL MATERIAL –","file_id":"12454","file_name":"vorf_supp.pdf","content_type":"application/pdf","access_level":"open_access","relation":"supplementary_material","checksum":"ce5f4ce66eaaa1590ee5df989fca6f61"},{"checksum":"08aecca434b08fee75ee1efe87943718","relation":"supplementary_material","access_level":"open_access","content_type":"video/mp4","file_name":"video.mp4","file_id":"12455","date_updated":"2023-01-30T10:48:37Z","creator":"bbickel","date_created":"2023-01-30T10:48:37Z","file_size":57855492}],"month":"12","file_date_updated":"2023-01-30T10:48:37Z","publication_status":"published","has_accepted_license":"1","abstract":[{"text":"Portrait viewpoint and illumination editing is an important problem with several applications in VR/AR, movies, and photography. Comprehensive knowledge of geometry and illumination is critical for obtaining photorealistic results. Current methods are unable to explicitly model in 3D while handing both viewpoint and illumination editing from a single image. In this paper, we propose VoRF, a novel approach that can take even a single portrait image as input and relight human heads under novel illuminations that can be viewed from arbitrary viewpoints. VoRF represents a human head as a continuous volumetric field and learns a prior model of human heads using a coordinate-based MLP with separate latent spaces for identity and illumination. The prior model is learnt in an auto-decoder manner over a diverse class of head shapes and appearances, allowing VoRF to generalize to novel test identities from a single input image. Additionally, VoRF has a reflectance MLP that uses the intermediate features of the prior model for rendering One-Light-at-A-Time (OLAT) images under novel views. We synthesize novel illuminations by combining these OLAT images with target environment maps. Qualitative and quantitative evaluations demonstrate the effectiveness of VoRF for relighting and novel view synthesis even when applied to unseen subjects under uncontrolled illuminations.","lang":"eng"}],"date_created":"2023-01-30T10:47:06Z","scopus_import":"1","day":"01","author":[{"first_name":"Pramod","full_name":"Rao, Pramod","last_name":"Rao"},{"last_name":"B R","full_name":"B R, Mallikarjun","first_name":"Mallikarjun"},{"last_name":"Fox","full_name":"Fox, Gereon","first_name":"Gereon"},{"first_name":"Tim","last_name":"Weyrich","full_name":"Weyrich, Tim"},{"orcid":"0000-0001-6511-9385","first_name":"Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87","full_name":"Bickel, Bernd","last_name":"Bickel"},{"first_name":"Hans-Peter","last_name":"Seidel","full_name":"Seidel, Hans-Peter"},{"first_name":"Hanspeter","last_name":"Pfister","full_name":"Pfister, Hanspeter"},{"first_name":"Wojciech","full_name":"Matusik, Wojciech","last_name":"Matusik"},{"last_name":"Tewari","full_name":"Tewari, Ayush","first_name":"Ayush"},{"first_name":"Christian","last_name":"Theobalt","full_name":"Theobalt, Christian"},{"first_name":"Mohamed","last_name":"Elgharib","full_name":"Elgharib, Mohamed"}],"oa_version":"Published Version","title":"VoRF: Volumetric Relightable Faces","conference":{"location":"London, United Kingdom","name":"BMVC: British Machine Vision Conference","start_date":"2022-11-21","end_date":"2022-11-24"},"date_published":"2022-12-01T00:00:00Z","acknowledgement":"This work was supported by the ERC Consolidator Grant 4DReply (770784).","publication":"33rd British Machine Vision Conference","status":"public","year":"2022","main_file_link":[{"url":"https://bmvc2022.mpi-inf.mpg.de/708/","open_access":"1"}],"quality_controlled":"1","ddc":["000"],"_id":"12452","date_updated":"2023-10-31T08:40:55Z","type":"conference","article_processing_charge":"No","publisher":"British Machine Vision Association and Society for Pattern Recognition"},{"quality_controlled":"1","page":"7568-7588","ddc":["000"],"_id":"9241","date_updated":"2023-08-07T14:11:57Z","type":"journal_article","article_processing_charge":"No","doi":"10.1364/OE.406095","publisher":"The Optical Society","ec_funded":1,"date_published":"2021-03-01T00:00:00Z","acknowledgement":"H2020 Marie Skłodowska-Curie Actions (642841); European Research Council (715767); Grantová Agentura České Republiky (16-08111S, 16-18964S); Univerzita Karlova v Praze (SVV-2017-260452); Engineering and Physical Sciences Research Council (EP/K023578/1).\r\nWe are grateful to Stratasys Ltd. for access to the voxel-level print interface of the J750\r\nmachine.","status":"public","publication":"Optics Express","project":[{"_id":"2508E324-B435-11E9-9278-68D0E5697425","call_identifier":"H2020","grant_number":"642841","name":"Distributed 3D Object Design"},{"name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","grant_number":"715767","call_identifier":"H2020","_id":"24F9549A-B435-11E9-9278-68D0E5697425"}],"isi":1,"year":"2021","external_id":{"isi":["000624968100103"]},"file_date_updated":"2021-03-22T08:15:28Z","publication_status":"published","publication_identifier":{"eissn":["1094-4087"]},"has_accepted_license":"1","tmp":{"name":"Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)","legal_code_url":"https://creativecommons.org/licenses/by/4.0/legalcode","image":"/images/cc_by.png","short":"CC BY (4.0)"},"intvolume":"        29","abstract":[{"lang":"eng","text":"Volumetric light transport is a pervasive physical phenomenon, and therefore its accurate simulation is important for a broad array of disciplines. While suitable mathematical models for computing the transport are now available, obtaining the necessary material parameters needed to drive such simulations is a challenging task: direct measurements of these parameters from material samples are seldom possible. Building on the inverse scattering paradigm, we present a novel measurement approach which indirectly infers the transport parameters from extrinsic observations of multiple-scattered radiance. The novelty of the proposed approach lies in replacing structured illumination with a structured reflector bonded to the sample, and a robust fitting procedure that largely compensates for potential systematic errors in the calibration of the setup. We show the feasibility of our approach by validating simulations of complex 3D compositions of the measured materials against physical prints, using photo-polymer resins. As presented in this paper, our technique yields colorspace data suitable for accurate appearance reproduction in the area of 3D printing. Beyond that, and without fundamental changes to the basic measurement methodology, it could equally well be used to obtain spectral measurements that are useful for other application areas."}],"volume":29,"date_created":"2021-03-14T23:01:33Z","article_type":"original","scopus_import":"1","day":"01","author":[{"first_name":"Oskar","full_name":"Elek, Oskar","last_name":"Elek"},{"orcid":"0000-0002-3808-281X","first_name":"Ran","full_name":"Zhang, Ran","id":"4DDBCEB0-F248-11E8-B48F-1D18A9856A87","last_name":"Zhang"},{"full_name":"Sumin, Denis","last_name":"Sumin","first_name":"Denis"},{"last_name":"Myszkowski","full_name":"Myszkowski, Karol","first_name":"Karol"},{"last_name":"Bickel","id":"49876194-F248-11E8-B48F-1D18A9856A87","full_name":"Bickel, Bernd","first_name":"Bernd","orcid":"0000-0001-6511-9385"},{"full_name":"Wilkie, Alexander","last_name":"Wilkie","first_name":"Alexander"},{"first_name":"Jaroslav","last_name":"Křivánek","full_name":"Křivánek, Jaroslav"},{"first_name":"Tim","last_name":"Weyrich","full_name":"Weyrich, Tim"}],"title":"Robust and practical measurement of volume transport parameters in solid photo-polymer materials for 3D printing","oa_version":"Published Version","citation":{"chicago":"Elek, Oskar, Ran Zhang, Denis Sumin, Karol Myszkowski, Bernd Bickel, Alexander Wilkie, Jaroslav Křivánek, and Tim Weyrich. “Robust and Practical Measurement of Volume Transport Parameters in Solid Photo-Polymer Materials for 3D Printing.” <i>Optics Express</i>. The Optical Society, 2021. <a href=\"https://doi.org/10.1364/OE.406095\">https://doi.org/10.1364/OE.406095</a>.","ista":"Elek O, Zhang R, Sumin D, Myszkowski K, Bickel B, Wilkie A, Křivánek J, Weyrich T. 2021. Robust and practical measurement of volume transport parameters in solid photo-polymer materials for 3D printing. Optics Express. 29(5), 7568–7588.","mla":"Elek, Oskar, et al. “Robust and Practical Measurement of Volume Transport Parameters in Solid Photo-Polymer Materials for 3D Printing.” <i>Optics Express</i>, vol. 29, no. 5, The Optical Society, 2021, pp. 7568–88, doi:<a href=\"https://doi.org/10.1364/OE.406095\">10.1364/OE.406095</a>.","apa":"Elek, O., Zhang, R., Sumin, D., Myszkowski, K., Bickel, B., Wilkie, A., … Weyrich, T. (2021). Robust and practical measurement of volume transport parameters in solid photo-polymer materials for 3D printing. <i>Optics Express</i>. The Optical Society. <a href=\"https://doi.org/10.1364/OE.406095\">https://doi.org/10.1364/OE.406095</a>","ama":"Elek O, Zhang R, Sumin D, et al. Robust and practical measurement of volume transport parameters in solid photo-polymer materials for 3D printing. <i>Optics Express</i>. 2021;29(5):7568-7588. doi:<a href=\"https://doi.org/10.1364/OE.406095\">10.1364/OE.406095</a>","short":"O. Elek, R. Zhang, D. Sumin, K. Myszkowski, B. Bickel, A. Wilkie, J. Křivánek, T. Weyrich, Optics Express 29 (2021) 7568–7588.","ieee":"O. Elek <i>et al.</i>, “Robust and practical measurement of volume transport parameters in solid photo-polymer materials for 3D printing,” <i>Optics Express</i>, vol. 29, no. 5. The Optical Society, pp. 7568–7588, 2021."},"issue":"5","user_id":"4359f0d1-fa6c-11eb-b949-802e58b17ae8","oa":1,"language":[{"iso":"eng"}],"department":[{"_id":"BeBi"}],"file":[{"date_created":"2021-03-22T08:15:28Z","file_size":10873700,"date_updated":"2021-03-22T08:15:28Z","creator":"dernst","file_id":"9269","success":1,"file_name":"2021_OpticsExpress_Elek.pdf","access_level":"open_access","content_type":"application/pdf","relation":"main_file","checksum":"a9697ad83136c19ad87e46aa2db63cfd"}],"month":"03"},{"keyword":["multistability","mechanism","computational design","rigidity"],"external_id":{"isi":["000752079300003"]},"year":"2021","isi":1,"acknowledgement":"We would like to thank everyone who contributed to this paper, the authors of artworks for all the examples, including @macrovec-tor_official and Wikimedia for the FLAG semaphore, and @pikisuper-star for the FIGURINE. The photos of iconic poses in the teaser were supplied by (from left to right): Mike Hewitt/Olympics Day 8 - Athletics/Gettty Images, Oneinchpunch/Basketball player training on acourt in New york city/Shutterstock, and Andrew Redington/TigerWoods/Getty Images. We also want to express our gratitude to Christian Hafner for insightful discussions, the IST Austria machine shop SSU, all proof-readers, and anonymous reviewers. This project has received funding from the European Union’s Horizon 2020 research and innovation programme, under the Marie Skłodowska-Curie grant agreement No 642841 (DISTRO), and under the European Research Council grant agreement No 715767 (MATERIALIZABLE).","date_published":"2021-10-08T00:00:00Z","ec_funded":1,"status":"public","publication":"ACM Transactions on Graphics","project":[{"_id":"2508E324-B435-11E9-9278-68D0E5697425","call_identifier":"H2020","name":"Distributed 3D Object Design","grant_number":"642841"},{"_id":"24F9549A-B435-11E9-9278-68D0E5697425","call_identifier":"H2020","name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","grant_number":"715767"}],"type":"journal_article","_id":"9376","date_updated":"2023-08-08T13:31:38Z","publisher":"Association for Computing Machinery","article_processing_charge":"No","doi":"10.1145/3453477","quality_controlled":"1","ddc":["000"],"article_number":"186","file":[{"access_level":"open_access","content_type":"application/pdf","file_name":"Multistable-authorversion.pdf","checksum":"8564b3118457d4c8939a8ef2b1a2f16c","relation":"main_file","date_updated":"2021-05-08T17:36:59Z","creator":"bbickel","file_size":18926557,"date_created":"2021-05-08T17:36:59Z","file_id":"9377"},{"creator":"bbickel","date_updated":"2021-05-08T17:38:22Z","date_created":"2021-05-08T17:38:22Z","file_size":76542901,"file_id":"9378","content_type":"video/mp4","access_level":"open_access","file_name":"multistable-video.mp4","success":1,"checksum":"3b6e874e30bfa1bfc3ad3498710145a1","relation":"main_file"},{"file_name":"multistable-supplementary material.pdf","content_type":"application/pdf","access_level":"open_access","relation":"supplementary_material","description":"This document provides additional results and analyzes the robustness and limitations of our approach.","checksum":"20dc3bc42e1a912a5b0247c116772098","file_size":3367072,"date_created":"2021-12-17T08:13:51Z","creator":"bbickel","date_updated":"2021-12-17T08:13:51Z","title":"Supplementary Material for “Computational Design of Planar Multistable Compliant Structures”","file_id":"10562"}],"department":[{"_id":"BeBi"}],"month":"10","user_id":"4359f0d1-fa6c-11eb-b949-802e58b17ae8","citation":{"chicago":"Zhang, Ran, Thomas Auzinger, and Bernd Bickel. “Computational Design of Planar Multistable Compliant Structures.” <i>ACM Transactions on Graphics</i>. Association for Computing Machinery, 2021. <a href=\"https://doi.org/10.1145/3453477\">https://doi.org/10.1145/3453477</a>.","ista":"Zhang R, Auzinger T, Bickel B. 2021. Computational design of planar multistable compliant structures. ACM Transactions on Graphics. 40(5), 186.","apa":"Zhang, R., Auzinger, T., &#38; Bickel, B. (2021). Computational design of planar multistable compliant structures. <i>ACM Transactions on Graphics</i>. Association for Computing Machinery. <a href=\"https://doi.org/10.1145/3453477\">https://doi.org/10.1145/3453477</a>","mla":"Zhang, Ran, et al. “Computational Design of Planar Multistable Compliant Structures.” <i>ACM Transactions on Graphics</i>, vol. 40, no. 5, 186, Association for Computing Machinery, 2021, doi:<a href=\"https://doi.org/10.1145/3453477\">10.1145/3453477</a>.","ama":"Zhang R, Auzinger T, Bickel B. Computational design of planar multistable compliant structures. <i>ACM Transactions on Graphics</i>. 2021;40(5). doi:<a href=\"https://doi.org/10.1145/3453477\">10.1145/3453477</a>","ieee":"R. Zhang, T. Auzinger, and B. Bickel, “Computational design of planar multistable compliant structures,” <i>ACM Transactions on Graphics</i>, vol. 40, no. 5. Association for Computing Machinery, 2021.","short":"R. Zhang, T. Auzinger, B. Bickel, ACM Transactions on Graphics 40 (2021)."},"issue":"5","language":[{"iso":"eng"}],"oa":1,"date_created":"2021-05-08T17:37:08Z","article_type":"original","volume":40,"title":"Computational design of planar multistable compliant structures","oa_version":"Published Version","day":"08","author":[{"full_name":"Zhang, Ran","id":"4DDBCEB0-F248-11E8-B48F-1D18A9856A87","last_name":"Zhang","first_name":"Ran","orcid":"0000-0002-3808-281X"},{"last_name":"Auzinger","id":"4718F954-F248-11E8-B48F-1D18A9856A87","full_name":"Auzinger, Thomas","orcid":"0000-0002-1546-3265","first_name":"Thomas"},{"last_name":"Bickel","id":"49876194-F248-11E8-B48F-1D18A9856A87","full_name":"Bickel, Bernd","first_name":"Bernd","orcid":"0000-0001-6511-9385"}],"file_date_updated":"2021-12-17T08:13:51Z","publication_status":"published","publication_identifier":{"eissn":["1557-7368"],"issn":["0730-0301"]},"acknowledged_ssus":[{"_id":"M-Shop"}],"intvolume":"        40","tmp":{"name":"Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)","legal_code_url":"https://creativecommons.org/licenses/by/4.0/legalcode","image":"/images/cc_by.png","short":"CC BY (4.0)"},"abstract":[{"text":"This paper presents a method for designing planar multistable compliant structures. Given a sequence of desired stable states and the corresponding poses of the structure, we identify the topology and geometric realization of a mechanism—consisting of bars and joints—that is able to physically reproduce the desired multistable behavior. In order to solve this problem efficiently, we build on insights from minimally rigid graph theory to identify simple but effective topologies for the mechanism. We then optimize its geometric parameters, such as joint positions and bar lengths, to obtain correct transitions between the given poses. Simultaneously, we ensure adequate stability of each pose based on an effective approximate error metric related to the elastic energy Hessian of the bars in the mechanism. As demonstrated by our results, we obtain functional multistable mechanisms of manageable complexity that can be fabricated using 3D printing. Further, we evaluated the effectiveness of our method on a large number of examples in the simulation and fabricated several physical prototypes.","lang":"eng"}],"has_accepted_license":"1"},{"oa":1,"language":[{"iso":"eng"}],"citation":{"ieee":"X. Feng <i>et al.</i>, “Computational design of skinned Quad-Robots,” <i>IEEE Transactions on Visualization and Computer Graphics</i>, vol. 27, no. 6. IEEE, 2021.","short":"X. Feng, J. Liu, H. Wang, Y. Yang, H. Bao, B. Bickel, W. Xu, IEEE Transactions on Visualization and Computer Graphics 27 (2021).","ama":"Feng X, Liu J, Wang H, et al. Computational design of skinned Quad-Robots. <i>IEEE Transactions on Visualization and Computer Graphics</i>. 2021;27(6). doi:<a href=\"https://doi.org/10.1109/TVCG.2019.2957218\">10.1109/TVCG.2019.2957218</a>","apa":"Feng, X., Liu, J., Wang, H., Yang, Y., Bao, H., Bickel, B., &#38; Xu, W. (2021). Computational design of skinned Quad-Robots. <i>IEEE Transactions on Visualization and Computer Graphics</i>. IEEE. <a href=\"https://doi.org/10.1109/TVCG.2019.2957218\">https://doi.org/10.1109/TVCG.2019.2957218</a>","mla":"Feng, Xudong, et al. “Computational Design of Skinned Quad-Robots.” <i>IEEE Transactions on Visualization and Computer Graphics</i>, vol. 27, no. 6, 2881–2895, IEEE, 2021, doi:<a href=\"https://doi.org/10.1109/TVCG.2019.2957218\">10.1109/TVCG.2019.2957218</a>.","ista":"Feng X, Liu J, Wang H, Yang Y, Bao H, Bickel B, Xu W. 2021. Computational design of skinned Quad-Robots. IEEE Transactions on Visualization and Computer Graphics. 27(6), 2881–2895.","chicago":"Feng, Xudong, Jiafeng Liu, Huamin Wang, Yin Yang, Hujun Bao, Bernd Bickel, and Weiwei Xu. “Computational Design of Skinned Quad-Robots.” <i>IEEE Transactions on Visualization and Computer Graphics</i>. IEEE, 2021. <a href=\"https://doi.org/10.1109/TVCG.2019.2957218\">https://doi.org/10.1109/TVCG.2019.2957218</a>."},"issue":"6","user_id":"4359f0d1-fa6c-11eb-b949-802e58b17ae8","month":"06","department":[{"_id":"BeBi"}],"article_number":"2881-2895","file":[{"content_type":"application/pdf","access_level":"open_access","file_name":"2021_TVCG_Feng.pdf","success":1,"checksum":"a78e6ac94e33ade4ffaea66943d5f7dc","relation":"main_file","creator":"kschuh","date_updated":"2021-05-25T15:08:49Z","file_size":6183002,"date_created":"2021-05-25T15:08:49Z","file_id":"9427"}],"has_accepted_license":"1","intvolume":"        27","tmp":{"name":"Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)","legal_code_url":"https://creativecommons.org/licenses/by/4.0/legalcode","image":"/images/cc_by.png","short":"CC BY (4.0)"},"abstract":[{"text":"We present a computational design system that assists users to model, optimize, and fabricate quad-robots with soft skins. Our system addresses the challenging task of predicting their physical behavior by fully integrating the multibody dynamics of the mechanical skeleton and the elastic behavior of the soft skin. The developed motion control strategy uses an alternating optimization scheme to avoid expensive full space time-optimization, interleaving space-time optimization for the skeleton, and frame-by-frame optimization for the full dynamics. The output are motor torques to drive the robot to achieve a user prescribed motion trajectory. We also provide a collection of convenient engineering tools and empirical manufacturing guidance to support the fabrication of the designed quad-robot. We validate the feasibility of designs generated with our system through physics simulations and with a physically-fabricated prototype.","lang":"eng"}],"file_date_updated":"2021-05-25T15:08:49Z","publication_identifier":{"issn":["19410506"],"eissn":["10772626"]},"publication_status":"published","scopus_import":"1","day":"01","author":[{"first_name":"Xudong","full_name":"Feng, Xudong","last_name":"Feng"},{"first_name":"Jiafeng","full_name":"Liu, Jiafeng","last_name":"Liu"},{"last_name":"Wang","full_name":"Wang, Huamin","first_name":"Huamin"},{"last_name":"Yang","full_name":"Yang, Yin","first_name":"Yin"},{"first_name":"Hujun","full_name":"Bao, Hujun","last_name":"Bao"},{"first_name":"Bernd","orcid":"0000-0001-6511-9385","last_name":"Bickel","full_name":"Bickel, Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87"},{"first_name":"Weiwei","full_name":"Xu, Weiwei","last_name":"Xu"}],"title":"Computational design of skinned Quad-Robots","oa_version":"Published Version","volume":27,"date_created":"2021-05-23T22:01:42Z","status":"public","publication":"IEEE Transactions on Visualization and Computer Graphics","project":[{"_id":"24F9549A-B435-11E9-9278-68D0E5697425","call_identifier":"H2020","name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","grant_number":"715767"}],"ec_funded":1,"pmid":1,"date_published":"2021-06-01T00:00:00Z","acknowledgement":"The authors would like to thank anonymous reviewers for their constructive comments. Weiwei Xu is partially supported by Zhejiang Lab. Yin Yang is partially spported by NSF under Grant Nos. CHS 1845024 and 1717972. Weiwei Xu and Hujun Bao are supported by Fundamental Research Funds for the Central Universities. This project has received funding from the European Research Council (ERC) under the European Unions Horizon 2020 research and innovation programme (Grant agreement No 715767).","year":"2021","isi":1,"external_id":{"pmid":["31804937"],"isi":["000649620700009"]},"ddc":["000"],"quality_controlled":"1","article_processing_charge":"No","doi":"10.1109/TVCG.2019.2957218","publisher":"IEEE","_id":"9408","date_updated":"2023-08-08T13:45:46Z","type":"journal_article"},{"ec_funded":1,"acknowledgement":"We thank Sebastian Cucerca for processing and capturing the phys-cal printouts. This work was supported by the Charles University grant SVV-260588 and Czech Science Foundation grant 19-07626S. This project has received funding from the European Union’s Horizon 2020 research and innovation programme, under the Marie Skłodowska Curie grant agreements No 642841 (DISTRO) and No765911 (RealVision), and under the European Research Council grant agreement No 715767 (MATERIALIZABLE).","date_published":"2021-05-01T00:00:00Z","project":[{"_id":"2508E324-B435-11E9-9278-68D0E5697425","grant_number":"642841","name":"Distributed 3D Object Design","call_identifier":"H2020"},{"grant_number":"715767","name":"MATERIALIZABLE: Intelligent fabrication-oriented Computational Design and Modeling","call_identifier":"H2020","_id":"24F9549A-B435-11E9-9278-68D0E5697425"}],"publication":"Computer Graphics Forum","status":"public","year":"2021","isi":1,"external_id":{"isi":["000657959600017"]},"quality_controlled":"1","page":"205-219","ddc":["004"],"date_updated":"2023-08-14T08:01:50Z","_id":"9547","type":"journal_article","doi":"10.1111/cgf.142626","article_processing_charge":"No","publisher":"Wiley","issue":"2","citation":{"chicago":"Rittig, Tobias, Denis Sumin, Vahid Babaei, Piotr Didyk, Alexey Voloboy, Alexander Wilkie, Bernd Bickel, Karol Myszkowski, Tim Weyrich, and Jaroslav Křivánek. “Neural Acceleration of Scattering-Aware Color 3D Printing.” <i>Computer Graphics Forum</i>. Wiley, 2021. <a href=\"https://doi.org/10.1111/cgf.142626\">https://doi.org/10.1111/cgf.142626</a>.","ista":"Rittig T, Sumin D, Babaei V, Didyk P, Voloboy A, Wilkie A, Bickel B, Myszkowski K, Weyrich T, Křivánek J. 2021. Neural acceleration of scattering-aware color 3D printing. Computer Graphics Forum. 40(2), 205–219.","mla":"Rittig, Tobias, et al. “Neural Acceleration of Scattering-Aware Color 3D Printing.” <i>Computer Graphics Forum</i>, vol. 40, no. 2, Wiley, 2021, pp. 205–19, doi:<a href=\"https://doi.org/10.1111/cgf.142626\">10.1111/cgf.142626</a>.","apa":"Rittig, T., Sumin, D., Babaei, V., Didyk, P., Voloboy, A., Wilkie, A., … Křivánek, J. (2021). Neural acceleration of scattering-aware color 3D printing. <i>Computer Graphics Forum</i>. Wiley. <a href=\"https://doi.org/10.1111/cgf.142626\">https://doi.org/10.1111/cgf.142626</a>","ama":"Rittig T, Sumin D, Babaei V, et al. Neural acceleration of scattering-aware color 3D printing. <i>Computer Graphics Forum</i>. 2021;40(2):205-219. doi:<a href=\"https://doi.org/10.1111/cgf.142626\">10.1111/cgf.142626</a>","short":"T. Rittig, D. Sumin, V. Babaei, P. Didyk, A. Voloboy, A. Wilkie, B. Bickel, K. Myszkowski, T. Weyrich, J. Křivánek, Computer Graphics Forum 40 (2021) 205–219.","ieee":"T. Rittig <i>et al.</i>, “Neural acceleration of scattering-aware color 3D printing,” <i>Computer Graphics Forum</i>, vol. 40, no. 2. Wiley, pp. 205–219, 2021."},"user_id":"4359f0d1-fa6c-11eb-b949-802e58b17ae8","oa":1,"language":[{"iso":"eng"}],"department":[{"_id":"BeBi"}],"file":[{"relation":"main_file","checksum":"33271724215f54a75c39d2ed40f2c502","file_name":"ScatteringAwareColor3DPrinting_authorVersion.pdf","success":1,"content_type":"application/pdf","access_level":"open_access","file_id":"10120","file_size":26026501,"date_created":"2021-10-11T12:06:50Z","date_updated":"2021-10-11T12:06:50Z","creator":"bbickel"}],"month":"05","publication_identifier":{"issn":["0167-7055"],"eissn":["1467-8659"]},"publication_status":"published","file_date_updated":"2021-10-11T12:06:50Z","has_accepted_license":"1","abstract":[{"lang":"eng","text":"With the wider availability of full-color 3D printers, color-accurate 3D-print preparation has received increased attention. A key challenge lies in the inherent translucency of commonly used print materials that blurs out details of the color texture. Previous work tries to compensate for these scattering effects through strategic assignment of colored primary materials to printer voxels. To date, the highest-quality approach uses iterative optimization that relies on computationally expensive Monte Carlo light transport simulation to predict the surface appearance from subsurface scattering within a given print material distribution; that optimization, however, takes in the order of days on a single machine. In our work, we dramatically speed up the process by replacing the light transport simulation with a data-driven approach. Leveraging a deep neural network to predict the scattering within a highly heterogeneous medium, our method performs around two orders of magnitude faster than Monte Carlo rendering while yielding optimization results of similar quality level. The network is based on an established method from atmospheric cloud rendering, adapted to our domain and extended by a physically motivated weight sharing scheme that substantially reduces the network size. We analyze its performance in an end-to-end print preparation pipeline and compare quality and runtime to alternative approaches, and demonstrate its generalization to unseen geometry and material values. This for the first time enables full heterogenous material optimization for 3D-print preparation within time frames in the order of the actual printing time."}],"intvolume":"        40","volume":40,"article_type":"original","date_created":"2021-06-13T22:01:32Z","author":[{"last_name":"Rittig","full_name":"Rittig, Tobias","first_name":"Tobias"},{"last_name":"Sumin","full_name":"Sumin, Denis","first_name":"Denis"},{"first_name":"Vahid","full_name":"Babaei, Vahid","last_name":"Babaei"},{"last_name":"Didyk","full_name":"Didyk, Piotr","first_name":"Piotr"},{"first_name":"Alexey","full_name":"Voloboy, Alexey","last_name":"Voloboy"},{"first_name":"Alexander","last_name":"Wilkie","full_name":"Wilkie, Alexander"},{"last_name":"Bickel","full_name":"Bickel, Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87","orcid":"0000-0001-6511-9385","first_name":"Bernd"},{"first_name":"Karol","last_name":"Myszkowski","full_name":"Myszkowski, Karol"},{"first_name":"Tim","last_name":"Weyrich","full_name":"Weyrich, Tim"},{"full_name":"Křivánek, Jaroslav","last_name":"Křivánek","first_name":"Jaroslav"}],"scopus_import":"1","day":"01","title":"Neural acceleration of scattering-aware color 3D printing","oa_version":"Submitted Version"},{"article_processing_charge":"No","doi":"10.1145/3472749.3474798","publisher":"Association for Computing Machinery","_id":"10148","date_updated":"2021-10-19T19:29:06Z","type":"conference","page":"954-971","ddc":["000"],"quality_controlled":"1","year":"2021","publication":"34th Annual ACM Symposium","status":"public","project":[{"_id":"2508E324-B435-11E9-9278-68D0E5697425","call_identifier":"H2020","name":"Distributed 3D Object Design","grant_number":"642841"}],"ec_funded":1,"conference":{"start_date":"2021-10-10","end_date":"2021-10-14","name":"UIST: User Interface Software and Technology","location":"Virtual"},"date_published":"2021-10-10T00:00:00Z","acknowledgement":"Our gratitude goes out to Kamila Mushkina, Akhmajon Makhsadov, Jordan Espenshade, Bruno Fruchard, Roland Bennewitz, and Robert Drumm. This project has received funding from the EU’s Horizon 2020 research and innovation programme, under the Marie Skłodowska-Curie grant agreement No 642841 (DISTRO).","day":"10","author":[{"first_name":"Donald","full_name":"Degraen, Donald","last_name":"Degraen"},{"last_name":"Piovarci","full_name":"Piovarci, Michael","id":"62E473F4-5C99-11EA-A40E-AF823DDC885E","first_name":"Michael"},{"first_name":"Bernd","orcid":"0000-0001-6511-9385","full_name":"Bickel, Bernd","id":"49876194-F248-11E8-B48F-1D18A9856A87","last_name":"Bickel"},{"first_name":"Antonio","full_name":"Kruger, Antonio","last_name":"Kruger"}],"title":"Capturing tactile properties of real surfaces for haptic reproduction","oa_version":"Preprint","date_created":"2021-10-18T07:36:11Z","has_accepted_license":"1","abstract":[{"lang":"eng","text":"Tactile feedback of an object’s surface enables us to discern its material properties and affordances. This understanding is used in digital fabrication processes by creating objects with high-resolution surface variations to influence a user’s tactile perception. As the design of such surface haptics commonly relies on knowledge from real-life experiences, it is unclear how to adapt this information for digital design methods. In this work, we investigate replicating the haptics of real materials. Using an existing process for capturing an object’s microgeometry, we digitize and reproduce the stable surface information of a set of 15 fabric samples. In a psychophysical experiment, we evaluate the tactile qualities of our set of original samples and their replicas. From our results, we see that direct reproduction of surface variations is able to influence different psychophysical dimensions of the tactile perception of surface textures. While the fabrication process did not preserve all properties, our approach underlines that replication of surface microgeometries benefits fabrication methods in terms of haptic perception by covering a large range of tactile variations. Moreover, by changing the surface structure of a single fabricated material, its material perception can be influenced. We conclude by proposing strategies for capturing and reproducing digitized textures to better resemble the perceived haptics of the originals."}],"file_date_updated":"2021-10-18T07:36:03Z","publication_identifier":{"isbn":["978-1-4503-8635-7"]},"publication_status":"published","month":"10","department":[{"_id":"BeBi"}],"file":[{"access_level":"open_access","content_type":"application/pdf","file_name":"degraen-UIST2021_Texture_Appropriation_CR_preprint.pdf","checksum":"b0b26464df79b3a59e8ed82e4e19ab15","relation":"main_file","date_updated":"2021-10-18T07:36:03Z","creator":"bbickel","date_created":"2021-10-18T07:36:03Z","file_size":29796364,"file_id":"10149"}],"oa":1,"language":[{"iso":"eng"}],"citation":{"chicago":"Degraen, Donald, Michael Piovarci, Bernd Bickel, and Antonio Kruger. “Capturing Tactile Properties of Real Surfaces for Haptic Reproduction.” In <i>34th Annual ACM Symposium</i>, 954–71. Association for Computing Machinery, 2021. <a href=\"https://doi.org/10.1145/3472749.3474798\">https://doi.org/10.1145/3472749.3474798</a>.","ista":"Degraen D, Piovarci M, Bickel B, Kruger A. 2021. Capturing tactile properties of real surfaces for haptic reproduction. 34th Annual ACM Symposium. UIST: User Interface Software and Technology, 954–971.","apa":"Degraen, D., Piovarci, M., Bickel, B., &#38; Kruger, A. (2021). Capturing tactile properties of real surfaces for haptic reproduction. In <i>34th Annual ACM Symposium</i> (pp. 954–971). Virtual: Association for Computing Machinery. <a href=\"https://doi.org/10.1145/3472749.3474798\">https://doi.org/10.1145/3472749.3474798</a>","mla":"Degraen, Donald, et al. “Capturing Tactile Properties of Real Surfaces for Haptic Reproduction.” <i>34th Annual ACM Symposium</i>, Association for Computing Machinery, 2021, pp. 954–71, doi:<a href=\"https://doi.org/10.1145/3472749.3474798\">10.1145/3472749.3474798</a>.","ama":"Degraen D, Piovarci M, Bickel B, Kruger A. Capturing tactile properties of real surfaces for haptic reproduction. In: <i>34th Annual ACM Symposium</i>. Association for Computing Machinery; 2021:954-971. doi:<a href=\"https://doi.org/10.1145/3472749.3474798\">10.1145/3472749.3474798</a>","ieee":"D. Degraen, M. Piovarci, B. Bickel, and A. Kruger, “Capturing tactile properties of real surfaces for haptic reproduction,” in <i>34th Annual ACM Symposium</i>, Virtual, 2021, pp. 954–971.","short":"D. Degraen, M. Piovarci, B. Bickel, A. Kruger, in:, 34th Annual ACM Symposium, Association for Computing Machinery, 2021, pp. 954–971."},"user_id":"8b945eb4-e2f2-11eb-945a-df72226e66a9"}]
